[go: up one dir, main page]

WO2025179890A1 - Procédé d'interaction, dispositif électronique, support de stockage lisible et produit programme - Google Patents

Procédé d'interaction, dispositif électronique, support de stockage lisible et produit programme

Info

Publication number
WO2025179890A1
WO2025179890A1 PCT/CN2024/123731 CN2024123731W WO2025179890A1 WO 2025179890 A1 WO2025179890 A1 WO 2025179890A1 CN 2024123731 W CN2024123731 W CN 2024123731W WO 2025179890 A1 WO2025179890 A1 WO 2025179890A1
Authority
WO
WIPO (PCT)
Prior art keywords
conversation
user
reply
search
retrieval
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
PCT/CN2024/123731
Other languages
English (en)
Chinese (zh)
Other versions
WO2025179890A9 (fr
Inventor
汪芳山
康司辰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Publication of WO2025179890A1 publication Critical patent/WO2025179890A1/fr
Publication of WO2025179890A9 publication Critical patent/WO2025179890A9/fr
Pending legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models

Definitions

  • the present application relates to the field of computer technology, and in particular to an interaction method, an electronic device, a readable storage medium, and a program product.
  • artificial intelligence technology can be applied to scenarios where people exchange messages through electronic devices. For example, when people chat with each other through mobile applications, the application used by each user can help the user generate replies based on the messages sent by the other party, thereby assisting the user in replying to the other party.
  • AIGC artificial intelligence generated context
  • the embodiments of the present application provide an interactive method, an electronic device, a readable storage medium and a program product, which are used to solve the problem that in a person-to-person dialogue scenario, when a search is required, accurate responses cannot be provided to users in real time.
  • an embodiment of the present application provides an interaction method, comprising: displaying a first user conversation; if the first user conversation is a user conversation that requires retrieval, displaying a first reply related to a retrieval result corresponding to the first user conversation, wherein the retrieval result is retrieved based on the retrieval intention of the first user conversation; if the first user conversation is a user conversation that does not require retrieval, displaying a second reply to the first user conversation.
  • the electronic device will display the user conversation.
  • the electronic device can display replies in real time to help users send messages to each other.
  • users may touch upon topics that require database retrieval, such as movie viewings, event schedules, and so on.
  • a search is necessary. Therefore, different replies can be displayed depending on whether the user conversation requires retrieval. If a search is required, for example, if the current conversation involves knowledge quizzes, personal event schedules, user information, movie viewings, shopping, and so on, a first reply related to the search results can be displayed. This means that the first reply includes content related to the search results. If a search is not required, for example, if the two parties are simply chatting, a second reply will be displayed, which does not include content related to the search results.
  • the generated responses are more targeted.
  • the generated responses include content related to the search results, making them more accurate and improving the user experience with the dialogue system.
  • the method further includes: displaying the first reply or the second reply in the form of a card or a list, and the first reply or the second reply may include at least one data form of text, image, voice, video, and link.
  • the method further includes:
  • Acquire conversation scene data related to the first user conversation where the conversation scene data includes at least one of the conversation content of the first user conversation, user behavior information regarding the first user conversation, and physiological indicators of the user during the first user conversation; call a large language model, and the large language model determines, based on the conversation scene data, whether the first user conversation is a user conversation that needs to be retrieved; if the first user conversation is a user conversation that needs to be retrieved, the large language model identifies the retrieval intent of the first user conversation based on the conversation scene data.
  • the content of the first user's conversation may include the current and historical conversation content between the two parties; the user's behavioral information regarding the first user's conversation may include user behavior information such as editing, browsing, and page scrolling in the conversation system; and the user's physiological indicators during the first user conversation may include heart rate, facial expressions, etc.
  • the use of multi-dimensional conversation scene data can improve the accuracy of determining whether a search is necessary, so that when a search is required, the need for a search can be reflected in a timely manner, thereby improving the accuracy and timeliness of conversation processing.
  • the conversation scene data includes multi-dimensional data, the elements used to generate replies are richer. At this time, when a reply is generated based on multi-dimensional data, the generated reply is more accurate.
  • the first reply includes a fused retrieval result corresponding to the retrieval result.
  • the fused search results not only include the content corresponding to the search results, but also include the search replies corresponding to the search results that can be used to reply to the messages sent by the other party, and the search replies correspond to the search results, so that users can obtain more comprehensive, more relevant and more accurate replies.
  • the fused retrieval result is obtained in the following manner: the large language model generates the fused retrieval result based on the input retrieval result and the conversation scene data of the first user conversation.
  • retrieval results and the conversation scene data of the first user conversation are used as inputs of the large language model, so that the large model can be integrated with the retrieval results corresponding to the retrieval intent and the content analyzed in the conversation scene data, thereby providing users with high-quality responses.
  • the retrieval results are obtained in the following manner: according to the retrieval information corresponding to the retrieval intent, a search is performed in the resource database to obtain the retrieval results, wherein the retrieval information includes at least one of the keywords corresponding to the retrieval intent and the search sentences corresponding to the retrieval intent; the resource database includes a local resource database, or a cloud resource database, and the local resource database includes local data or local applications; the cloud resource database includes: applications or services deployed in the cloud.
  • the method further includes: when the first user conversation is a user conversation that needs to be retrieved, the large language model generates a second reply based on the conversation scene data.
  • the large language model has semantic understanding and reasoning capabilities, and the large language model can effectively generate a second response based on the dialogue scenario data.
  • the second reply includes a dialogue reply and a reply reason corresponding to the dialogue reply.
  • the generated reply may include the conversation reply and the reason for the reply.
  • the reason for the reply is the analysis of the generated conversation reply, and the reason for the reply or the conversation reply takes into account the emotional reactions of both parties in the conversation, thereby improving the accuracy of the reply information and avoiding the problem of both parties being disgusted due to inappropriate reply content.
  • the electronic device can use a pre-trained large language model, take the conversation scene data as input, obtain the conversation reply and the reason for the reply, and the output reply needs to take into account the emotional tendencies of both parties, so that the large model can generate conversation replies and reasons for the reply that meet the current user needs based on the conversation scene data, and the output content can take into account the intentions and emotional tendencies of both parties.
  • an embodiment of the present application provides an interaction method, applied to an electronic device, the method comprising: displaying a first user conversation; displaying a first reply to the first user conversation based on conversation scene data of the first user conversation, wherein the conversation scene data includes at least one of the user's behavioral information regarding the first user conversation and the user's physiological indicators of the first user conversation.
  • the content of the first user's conversation may include the current and historical conversation content between the two parties; the user's behavior information regarding the first user's conversation may include user behavior information such as editing, browsing, and page scrolling in the conversation system; and the user's physiological indicators during the first user conversation may include heart rate, facial expressions, etc. It is understood that the use of multi-dimensional conversation scenario data enriches the elements used to generate responses, thereby making the generated responses more accurate.
  • the first reply includes at least one of the following: a retrieval result corresponding to the conversation with the first user, a fused retrieval result related to the retrieval result, a conversation reply, and a reply reason corresponding to the conversation reply.
  • it also includes: calling a large language model, and the large language model determines whether the first user conversation is a user conversation that needs to be retrieved based on the conversation scene data; if the first user conversation is a user conversation that needs to be retrieved, displaying the retrieval results or the fused retrieval results; if the first user conversation is a user conversation that does not need to be retrieved, displaying the conversation reply and the reply reason corresponding to the conversation reply.
  • the need for retrieval is determined based on the conversation scene data, and the display is displayed when retrieval is needed and when retrieval is not needed. Display different content, making the displayed replies more targeted and improving user experience.
  • the fused retrieval result is obtained in the following manner: the large language model identifies the retrieval intent of the first user conversation based on the conversation scene data; the large language model searches the resource database based on the retrieval information corresponding to the retrieval intent to obtain the retrieval result, wherein the retrieval information includes at least one of the keywords corresponding to the retrieval intent and the search sentence corresponding to the retrieval intent; the large language model generates the fused retrieval result based on the input retrieval result and the conversation scene data of the first user conversation.
  • the dialogue response and the response reason corresponding to the dialogue response are obtained in the following manner: when the first user dialogue is a user dialogue that needs to be retrieved, the large language model generates a dialogue response and the response reason corresponding to the dialogue response based on the dialogue scene data.
  • the second aspect above also includes: displaying the first reply in the form of a card or a list, and the first reply may include at least one data form of text, image, voice, video, and link.
  • an embodiment of the present application provides an interaction method, applied to an electronic device, the method comprising: displaying a first user conversation; displaying a first reply corresponding to the first user conversation, the first reply comprising a conversation reply and a reply reason for the corresponding conversation reply.
  • the generated reply may include the conversation response and the reason for the reply.
  • the reason for the reply is an analysis of the generated conversation response, and the reason for the reply or the conversation response takes into account the emotional reactions of both parties in the conversation, thereby improving the accuracy of the reply information and avoiding the problem of inappropriate reply content causing resentment between both parties.
  • it also includes: obtaining conversation scene data related to the conversation with the first user, the conversation scene data including at least one of the conversation content of the first user conversation, the user's behavioral information regarding the conversation with the first user, and the user's physiological indicators of the conversation with the first user; calling the large language model, and the large language model generating a first reply based on the conversation scene data.
  • electronic devices can use pre-trained large language models, take conversation scene data as input, obtain conversation replies and reasons for replies, and the output replies need to consider the emotional tendencies of both parties, so that the large model can generate conversation replies and reasons for replies that meet the current user needs based on the conversation scene data, and the output content can take into account the intentions and emotional tendencies of both parties.
  • the method further includes: the large language model determining, based on the conversation scenario data, whether the first user conversation is a user conversation that needs to be retrieved; and generating the first reply if the first user conversation is a user conversation that does not need to be retrieved.
  • an embodiment of the present application provides an interaction method, applied to an electronic device, the method comprising: displaying a first user conversation, the first user conversation being a user conversation that needs to be retrieved; displaying a first reply related to a retrieval result corresponding to the first user conversation, the first reply comprising a fused retrieval result corresponding to the retrieval result, wherein the retrieval result is obtained by retrieval based on the retrieval intention of the first user conversation.
  • it also includes: obtaining conversation scene data related to the conversation with the first user, the conversation scene data including at least one of the conversation content of the first user conversation, the user's behavioral information regarding the conversation with the first user, and the user's physiological indicators of the conversation with the first user; calling the large language model, and the large language model generates the first reply based on the conversation scene data and the retrieval results.
  • the method further includes: the large language model determining, based on the conversation scenario data, whether the first user conversation is a user conversation that needs to be retrieved; and the large language model generating a first reply if the first user conversation is a user conversation that needs to be retrieved.
  • an embodiment of the present application provides an electronic device comprising: a memory for storing instructions, and one or more processors.
  • the processor executes any one of the interaction methods in the first aspect and its various implementations, the second aspect and its various implementations, the third aspect and its various implementations, or the fourth aspect and its various implementations.
  • an embodiment of the present application provides a computer-readable storage medium, on which instructions are stored, which execute on an electronic device any interaction method such as the first aspect and any interaction method in its various implementations of the first aspect, the second aspect and any interaction method in its various implementations of the second aspect, the third aspect and any interaction method in its various implementations of the third aspect, or the fourth aspect and any interaction method in its various implementations of the fourth aspect.
  • an embodiment of the present application provides a computer program product, the computer program product comprising: computer program code, when When the computer program code runs on a computer, it causes the computer to execute any interaction method of the first aspect and its various implementations, any interaction method of the second aspect and its various implementations, any interaction method of the third aspect and its various implementations, or any interaction method of the fourth aspect and its various implementations.
  • the beneficial effects of the fifth to seventh aspects can refer to the relevant beneficial effects of the first, second, third and fourth aspects, and will not be repeated here.
  • FIG1A shows a schematic diagram of a chat conversation scenario between people according to some embodiments of the present application
  • FIG1B shows a schematic diagram of a chat interface 001 according to some embodiments of the present application.
  • FIG2A shows a schematic diagram of a chat interface 002 according to some embodiments of the present application.
  • FIG2B shows a schematic diagram of a chat interface 003 according to some embodiments of the present application.
  • FIG3 shows a schematic diagram of an interaction method according to some embodiments of the present application.
  • FIG4A shows a schematic diagram of a system architecture 400 according to some embodiments of the present application.
  • FIG4B shows a schematic diagram of interaction between a terminal 100B and a cloud 200 according to some embodiments of the present application
  • FIG5A shows a schematic diagram of an interface 004 according to some embodiments of the present application.
  • FIG5B shows a schematic diagram of a search process according to some embodiments of the present application.
  • FIG6 shows a schematic diagram of modules of a terminal 100 and a cloud 200 according to some embodiments of the present application
  • FIG7 shows another module schematic diagram of the terminal 100 and the cloud 200 according to some embodiments of the present application.
  • FIG8 shows a schematic diagram of a device according to some embodiments of the present application.
  • FIG 1A illustrates a chat conversation scenario between people, according to some embodiments of the present application.
  • User KA and user KB communicate using the same chat application (e.g., instant messaging application 1 in the figure) on their respective terminals.
  • the chat application is a conversation system that allows both parties to engage in a chat conversation.
  • the conversation system on terminal 100A is referred to as "conversation system A0”
  • the conversation system on terminal 100B is referred to as “conversation system B0.”
  • user KA and user KB are communicating using conversation system A0 and conversation system B0, respectively.
  • the dialogue system can also recommend automatic reply content to users based on their chat records.
  • Figure 1B illustrates a chat interface 001 of terminal 100B based on dialogue system B0, during a chat between user KA and user KB, according to some embodiments of the present application.
  • the content sent by "User A-1" on interface 001 is the chat content sent by user KA via terminal 100A
  • the content sent by "User B-1" on interface 001 is the chat content sent by user KB via terminal 100B.
  • user KA sends a message to terminal 100B, as shown in the figure, corresponding to user A-1, via terminal 100A, "What are your plans after get off work today?"
  • User KB replies to user KA via terminal 100B, as shown in the figure, corresponding to user B-1, "I don't have anything to do tonight.”
  • User KA then sends a message to terminal 100B, as shown in the figure, corresponding to user A-1, "So?”
  • User KA replies to user A-1, "Then let's go see the newly released movie 1.”
  • the dialogue system will generate a simple reply “OK!” in the prompt card K11 as shown in FIG1B .
  • the conversation scenario shown in FIG1B includes content related to Movie 1, but the generated reply does not involve the movie content, and the generated reply is not accurate. Understandably, if the user needs to obtain accurate information about Movie 1, they need to search the application that allows purchasing movie tickets to obtain accurate information about Movie 1.
  • the present application proposes an interactive method, in which the conversation scene data of the two parties can be obtained first, and then it can be determined whether the conversation scene data needs to be retrieved. If retrieval is not required, for example, the two parties are just chatting, then a reply is generated directly based on the conversation scene data; if retrieval is required, for example, when the current conversation scene involves knowledge questions and answers, personal activity arrangements, user information, watching movies, shopping and other content, then a resource database, such as a local resource database or a cloud resource database, is searched accurately to obtain the required search results, thereby generating a reply based on the obtained search results.
  • a resource database such as a local resource database or a cloud resource database
  • the system determines whether a search is necessary and uses different methods to handle the situation, making the generated responses more targeted. Furthermore, when a search is required, specific information is retrieved from the resource database, and a response is generated based on the search results. This means that the response includes the search results, making the generated response more accurate and improving the user experience with the dialogue system. Furthermore, in scenarios where a search is not required, responses are generated directly based on the dialogue scenario data, saving time and computing resources.
  • a large model can be used to understand and analyze the conversation, thereby obtaining keywords or phrases related to watching movies that can be used for search, such as the keyword "movie 1.”
  • This keyword can then be searched across various resource databases to obtain accurate information.
  • a response related to the search results can then be generated based on the search results, making the generated response more accurate.
  • multi-dimensional data can be obtained as conversation scene data.
  • the conversation scene data can include not only the current and historical conversation content of the two parties, but also the user behavior information such as editing, browsing, page sliding, etc. of the current user in the conversation system, as well as the current user's physiological indicator data, such as heart rate, expression, etc., which are used to judge whether retrieval is needed, thereby improving the accuracy of judging whether retrieval is needed, so that when retrieval is needed, the need for retrieval can be reflected in a timely manner, improving the accuracy and timeliness of conversation processing.
  • the conversation scene data includes multi-dimensional data, the elements used to generate replies are richer. At this time, when generating replies based on multi-dimensional data, the generated replies are more accurate.
  • the generated response when a response is generated based on the conversation scenario data without the need for retrieval, the generated response may include the conversation response and the reason for the response.
  • the reason for the response is an analysis of the generated conversation response, and the reason for the response or the conversation response takes into account the emotional reactions of both parties to the conversation, thereby improving the accuracy of the response information and avoiding the problem of both parties being offended by inappropriate reply content.
  • the terminal device can use a pre-trained large model, take the conversation scenario data as input, obtain the conversation response and the reason for the response, and the output response needs to take into account the emotional tendencies of both parties, so that the large model can generate conversation responses and reasons for the response that meet the current user needs based on the conversation scenario data, and the output content can take into account the intentions and emotional tendencies of both parties.
  • FIG2A shows a schematic diagram of a chat interface 002 of a terminal 100B based on a dialogue system B1 during a chat between users KA and KB according to some embodiments of the present application.
  • Interface 002 of terminal 100B displays the content sent by user A-2 to user B-2: "What?" ", user KB can perform some reply operations in the input box K21 of the dialogue system B1 in response to the content sent by user KA. For example, user KB clicks and enters the text "I am going to the Internet cafe to play by myself” in the input box K21.
  • the dialogue system B1 When the dialogue system B1 detects that the time it takes for the user to click and enter text reaches the preset reply time, it can call the big model and use the text entered by the current user and the previous conversation content between the two parties as dialogue scene data, so that the big model generates a reply based on the dialogue scene data; it can also collect the psychological indicator data of the current user KB and the behavioral information of user KB, so that the dialogue scene data includes the psychological indicator data of the current user KB and the behavioral information of user KB for the big model to generate a reply.
  • the big model can analyze the intention of user KA to chat with the current user KB based on the dialogue scene data, which is actually the expectation to do something with user KB.
  • the big model then automatically generates a reply, which includes the dialogue reply and the reason for the reply.
  • the dialogue system B1 displays the generated reply in the form of a suggestion card K22.
  • the reply includes a box K22A corresponding to the reason for the reply and a box K22B corresponding to the conversational response.
  • box K22A contains the specific reason for the reply: "The other party asked about your evening plans and is looking forward to doing something with you. Your current response may upset the other party.”
  • box K22B contains the conversational response, "Then let's go play together!, generated by the large model displayed by dialogue system B1.
  • User KB can click the Send control K22C corresponding to the conversational response "Then let's go play together! to send the conversational response directly to user KA.
  • user KB can directly select and modify the conversational response "Then let's go play together! in box K22B, then send the modified content by clicking the Send control K22C.
  • user KB can also modify the original content in the input box based on the displayed dialogue reply "Then let's go play together!, for example, modifying it to the same content as the dialogue reply "Then let's go play together!, and then click the send control K23 corresponding to the content in the input box to send it.
  • the send control can also be omitted in the suggestion card K22.
  • the user can directly select the dialogue reply and reply reason in the suggestion card, and then the terminal 100B directly sends it to user KA.
  • the specific sending method is not specified here.
  • the specific reply reason displayed in the suggestion card K22 analyzes the intention of user KA to chat with the current user KB.
  • the original reply of user KB may cause user KA to be in a bad mood.
  • the psychological indicator data of user KB obtained, such as normal heart rate and calm mood, it also shows that user KB did not respond well.
  • User KA's dislike for the conversation can be addressed, allowing for a suitable response, "Let's go play together!”
  • the generated response reason and the conversational response fully consider the emotional tendencies of both parties, providing the current user with a more accurate, friendly, and appropriate response. This allows the user to communicate with the other party based on accurate and reasonable responses without causing dislike or discomfort.
  • various resource databases can be searched to obtain search results.
  • the search results obtained from the various resource databases can then be integrated with the content analyzed from the conversation scenario data to provide users with high-quality responses.
  • the macro model can integrate the search results obtained from the various resource databases with the response content generated by the macro model based on the conversation scenario data, so that the generated response includes the integrated search results corresponding to the search results.
  • the integrated search results not only include the content corresponding to the search results, but also include a search reply corresponding to the search results that can be used to reply to the message sent by the other party, and the search reply corresponds to the search results.
  • the search reply reflects the response content generated by the macro model for the conversation scenario data.
  • the generated and displayed integrated search results allow users to obtain more comprehensive, relevant, and accurate responses.
  • FIG2B illustrates a chat interface 003 of terminal 100B based on dialogue system B1 during a chat between users KA and KB.
  • the content of the conversation between "User A-2" and “B-2" on interface 003 is identical to the content already shared between “User A-1" and “B-1” in FIG1B .
  • user KB enters "Let's see where to go” in input box K31.
  • dialogue system B1 generates a response, as shown in suggestion card K32 in FIG2B .
  • the response displayed in suggestion card K32 includes a fused search result.
  • box K32A contains content corresponding to the search result: "Searching for ticket information for tonight's movie 1: XXX Cinema (XXX Store): ⁇ 49.9, Showtimes Tonight: 7:30 PM
  • box K32B contains the search response generated by the dialogue system, "Why don't we go to this place?"
  • the movie-watching content involved in the current conversation is strongly related to movie-related applications.
  • the terminal can retrieve valid information from the corresponding application and present the generated reply in an optimized manner. This allows the user KB to quickly obtain the information they need without having to search other resource databases.
  • the fused search results include content generated by the large model based on the conversation scenario data, eliminating the need for the user to enter a search reply related to the movie ticket search, thus improving the user experience.
  • the user can modify the content in box K32A or box K32B and then send it using the corresponding send controls K32C and K32D, respectively, or directly send the content in box K32A or box K32B using the send control.
  • the user can also modify the content in input box K31 based on the reply in suggestion card K32 and then send it using the send control K33 corresponding to input box K31.
  • the user can also directly select the content in suggestion card K32 and send it without using the send control.
  • the specific sending method is not specified here.
  • AIGC artificial intelligence generated context
  • ChatGPT ChatGPT
  • LaMDA LaMDA
  • PaLM PaLM
  • OPT-IML OPT-IML
  • the terminal can display the reply in the form of cards, lists or other forms, so that the user can intuitively obtain relevant information.
  • the dialogue system can be any application that can provide a chat function and can be deployed on any electronic device.
  • terminal 100A and terminal 100B can be any electronic device that runs a dialogue system, for example, including but not limited to mobile phones, tablet computers, vehicle-mounted equipment, augmented reality (AR)/virtual reality (VR) devices, ultra-mobile personal computers (UMPC), netbooks, personal digital assistants (PDA), servers, server clusters, etc., without limitation here.
  • augmented reality AR
  • VR virtual reality
  • UMPC ultra-mobile personal computers
  • PDA personal digital assistants
  • servers server clusters, etc., without limitation here.
  • Figure 3 provides a schematic diagram of an interactive method according to an embodiment of the present application.
  • a terminal device displays a user conversation and determines whether a search is required. Different responses are used depending on whether a search is required or not.
  • the following description is based on the processor in the terminal device as the execution subject. The specific steps are as follows:
  • a terminal device displays a user conversation and can monitor the conversation process in real time on the dialogue system.
  • conversation scene data related to the current user conversation is collected. For example, upon detecting the user's first action, conversation scene data is obtained.
  • a user conversation is any conversation content displayed on the terminal device, such as chat messages sent by two parties displayed on the terminal device.
  • the first operation can be, for example, a user clicking on an input box on the interactive interface and entering text, emoticons, or the like, or a user clicking and entering content in the input box for a preset reply time, or a user selecting a microphone for voice input, or a user selecting a reply control (not shown) configured on the interface. It is understood that the first operation can be any of a variety of operations, as long as they indicate that the user wants to chat with the other party, and these operations will not be detailed here.
  • Conversational context data may include, but is not limited to, current and historical conversation content between the two parties, user behavior information, and the current user's physiological indicator data.
  • Conversational content includes both sent and pre-sent content, and may include text, voice, video, images, and other information.
  • User behavior information includes the current user's actions within the conversation system, such as editing, browsing, and scrolling. For example, if the current chat content involves a previously discussed topic, the user can return to the previous chat content to view and edit it. In this case, user behavior information includes the user's scrolling actions, editing actions on the previous chat content, and the viewed chat content data.
  • the current user's physiological indicator data includes the current user's facial expressions, heart rate, and other information.
  • the terminal device can capture the user's facial expressions through a camera. These expressions may correspond to emotions such as anger, peace, happiness, and sadness.
  • heart rate the user's heart rate data can be obtained through internal devices capable of monitoring heart rate or other supporting equipment.
  • conversational context data may only include conversation content, and this is not a limitation. It is understandable that when using multi-dimensional dialogue scene data, the accuracy of the judgment results can be improved because the information of the dialogue scene data is more comprehensive, so that when retrieval is needed, it can be performed in real time, thereby improving the accuracy and timeliness of dialogue processing.
  • the 10 most recent sent contents of user A-2 and user B-2 in the current chat conversation and the pre-sent content entered by user KB in the input box K21 can be obtained; the user's expression captured by the camera, such as "peaceful", and the heart rate data of "75bpm" collected by the watch can also be obtained.
  • the 10 most recent sent contents of user A-2 and user B-2 in the current chat conversation, as well as the content entered by user KB in the input box K31 can be obtained; the user slides to the topic of "Movie 1" discussed the day before yesterday to view the behavioral information of the previous content; the user's expression captured by the camera, such as "peaceful", and the heart rate data of "75bpm" collected by the watch can also be obtained.
  • S302 Determine whether a search is required based on the conversation scene data.
  • the current user conversation may involve some content that needs to be retrieved, such as movies, food, restaurants, buildings, names of people, trademarks, place names, customs, travel services, transportation, gossip information, event arrangements, and specific content of various disciplines such as law, music, painting, medicine, etc., which are not listed here.
  • content such as movies, food, restaurants, buildings, names of people, trademarks, place names, customs, travel services, transportation, gossip information, event arrangements, and specific content of various disciplines such as law, music, painting, medicine, etc., which are not listed here.
  • a search is required.
  • no search is required.
  • the search intent can be accurately identified later to obtain the search results.
  • a search when a search is not required, there is no need to identify the search intent and subsequent search, and a reply is directly generated, thereby saving time and computing resources.
  • the terminal device invokes a trained large model to determine whether the current conversation scenario data requires retrieval.
  • the large model can determine whether the current user conversation is a user conversation requiring retrieval, i.e., whether a retrieval is necessary, by extracting keywords or key sentences corresponding to the content to be retrieved from the current conversation scenario data. If the large model determines that a retrieval is necessary, the process proceeds to S304. If the large model determines that a retrieval is not necessary, the process proceeds to S303.
  • the large model may be used to determine that no search is required, and the process proceeds to S303 .
  • the large model can be used to identify keywords related to movies and other content, thereby determining that a search is to be performed and entering S304.
  • the large model's ability to determine whether a search is necessary can be deployed on either the terminal device or the cloud device.
  • the terminal device can send the conversation scenario data to the cloud, which then specifically determines whether a search is necessary based on the conversation scenario data. That is, step S302 can also be executed on the cloud, with the cloud feeding back the large model's determination result to the terminal device.
  • the large model may include a discriminant model to determine whether retrieval is required.
  • the network model, convolutional network model and other algorithms are not required here.
  • any electronic device such as any terminal device or cloud device, can be used to pre-train the discriminant model in the large model.
  • the discriminant model can be trained as follows: Conversational scenario data related to various conversational scenarios is prepared in advance and used as training sample data.
  • the conversational scenario data used for training is referred to as "conversational scenario sample data.”
  • Each conversational scenario sample data is assigned a corresponding label.
  • Conversational scenario sample data that requires retrieval is labeled as "searchable,” while conversational scenario sample data that does not require retrieval is labeled as "non-searchable.”
  • the training sample data and corresponding labels are then fed into the initial model to be trained, and the model is trained to obtain the discriminant model.
  • the conversation scene sample data includes not only conversation scenes that do not need to be retrieved, such as conversation scenes for small talk such as asking the other party how old they are this year, their home address, what they are wearing, and praising each other, but also conversation scenes that need to be retrieved.
  • the conversation scene sample data that need to be retrieved can be generated by a pre-configured resource database to generate the required conversation scene sample data, so that the training sample data includes various types of content in multiple dimensions.
  • the resource database configured in the training process is only used for the training process. When the large model training is completed and actually used, the resource database for retrieval may include but is not limited to the aforementioned resource database, or may not include the aforementioned resource database, and this is not required here.
  • the pre-configured resource database for training the model can be: (1) Knowledge base: data containing common user questions and related answers, such as data with relevant answers such as the birth date of "Qin Shi Huang". (2) External data source: including external data sources such as online event calendars, social media platforms or user personal calendars that involve user security and other information. (3) Third-party application (APP) information: including registration information of third-party applications (APPs) and specific function information. It is understandable that third-party applications can provide various related functions. When it comes to functions, they also need to be retrieved, so they can be used as important data to determine whether to retrieve.
  • the conversation scene sample data used for training can include not only the conversation content, but also the user's behavior information and psychological indicator data.
  • sample conversation scene data can be generated based on the aforementioned data sources, which is widely representative.
  • step S303 is an optional step.
  • the terminal device generates a reply based on the conversation scenario data, and the generated reply may only include the conversation reply.
  • the terminal device generates a reply based on the conversation scenario data, and the generated reply may include both the conversation reply and the reply reason corresponding to the conversation reply.
  • the prompt card K22 shows the dialogue response and the reason for the response based on the current dialogue scene.
  • the terminal device can generate responses based on the conversation scenario data using a large model.
  • the large model can include a generative model for generating responses based on the conversation scenario data and other information.
  • the generative model can be a recurrent neural network, a transformer model, or other generative model, though this is not a requirement.
  • the large model's ability to generate responses can be deployed on the terminal device or on the cloud device.
  • the terminal device can receive the response content sent by the cloud and thus obtain the response.
  • any electronic device such as any terminal device or cloud device, can be used to pre-train the generative model in the large model.
  • the cloud device training the large model as an example, the cloud device obtains the generative model through the Prompt fine-tuning training method.
  • a large amount of conversation scene sample data obtained from the data source is input into the initial model for training, so that the model learns language capabilities, thereby enabling the model to complete the context, such as a pre-configured template with a fixed structure.
  • the parameters of a certain part of the model are fine-tuned so that the output results include reasonable conversation responses and reasons for the responses, and the output responses take into account the intentions and emotional tendencies of both parties.
  • a pre-set template structure is defined and filled in the corresponding position to generate human-readable responses and suggestions.
  • the template can be: "The other party asks about your evening arrangements (where "evening" indicates time), and is looking forward to doing something with you ("doing something with you” indicates the content of the action). Your current response may result in The other party is in a bad mood ("the other party is in a bad mood" means the consequence). Then fill in the key information into the template and get the corresponding reply.
  • conversation understanding technologies and semantic understanding algorithms can also be used to understand the user's questions, recommend responses that are consistent with the user's intent and take into account the other party's emotions, tone and other factors to provide a satisfactory answer.
  • S304 Identify the search intent and obtain search information corresponding to the search intent.
  • the terminal device can call the large model at this time, so that the large model can identify the search intent based on the current conversation scene data. Then, the terminal device generates search information that can be searched in various resource databases supported by the current terminal device based on the search intent, such as query parameters, so as to perform a search based on the search information.
  • search intent is an intention that can be used to accurately search various resource databases to obtain accurate information.
  • the search intent is a more precise expression of the intent when the search content is used for retrieval, which can involve various fields, including but not limited to music, law, writing, painting, learning, life and other aspects.
  • the search information can be a search sentence, keyword, etc. corresponding to the search intent.
  • the intent recognition capability of the large model can be deployed on the terminal device or on the cloud device.
  • the terminal device can identify the retrieval intent when a retrieval is required, and then generate retrieval information.
  • the cloud device can continue to identify the retrieval intent when a retrieval is required, and then generate retrieval information, that is, step S304 can be executed on the cloud.
  • the large model may also include an intent recognition model for identifying retrieval intent.
  • the intent recognition model may be a neural network model, a convolutional network model, or other algorithms, etc., which are not limited here.
  • the intent recognition model can understand and classify the user's intent in the conversation scenario and obtain a feature vector related to the intent. Using the classification algorithm and the probability model, the output feature vector is input into the classification algorithm or the probability model to determine the most likely retrieval intent. It is understandable that when a retrieval is required, the retrieval intent is further accurately identified through the large model, so that the required content can be accurately retrieved from various resource databases through the retrieval information corresponding to the retrieval intent.
  • the large model can identify the search intent involving movie 1, and then generate search information based on the identified search intent, for example, "Order 1 movie ticket for tonight's movie 1", or keywords such as "order”, “movie 1", “watch”, and "tonight”.
  • search information for example, "Order 1 movie ticket for tonight's movie 1", or keywords such as "order”, “movie 1", “watch”, and "tonight”.
  • the specific form of the search information is not required here.
  • the cloud device when training the intent recognition model, can adopt the Prompt fine-tuning training method to obtain the intent recognition model: by training the model with training sample data including various corpora, the model learns the general language representation, and then fine-tunes the pre-trained model according to different downstream tasks.
  • the intent recognition model is used to output the preliminary intent, and then the preliminary intent is fine-tuned to obtain the target retrieval intent, and then the model is fine-tuned according to the target retrieval intent.
  • the training sample data used to train the intent recognition model can be the dialogue scene sample data from the configured resource database used in the process of training the discriminant model. Please refer to the above step S302 for details, which will not be repeated here.
  • S305 Obtain search results based on the search information.
  • a search is performed in various resource databases to obtain search results.
  • the various resource databases may include a local resource database of the current terminal device and a cloud resource database connected to the current terminal device.
  • the data in the local resource database may be data stored in the local storage space, including device information, user preferences, contacts, text messages, call logs, photos, local videos, application data, etc.
  • the data in the cloud resource database may be third-party application or service data provided by the cloud, such as third-party services deployed in the cloud for positioning, ticket purchase, flight booking, reservations, etc. used by various applications.
  • the user can search the local data corresponding to the ticket purchasing application on the terminal device and the cloud service corresponding to the ticket purchasing application based on the search information "Order a ticket for tonight's Movie 1" and other information, thereby obtaining local search results and cloud search results.
  • the cloud search result obtained is: XXX Cinema (XXX Store): ⁇ 49.9, Tonight's Showtimes: 19:30
  • the search results are used as replies.
  • the local search results and the cloud search results obtained for the aforementioned movie 1 are combined and organized and displayed as a reply on the interface.
  • the large model fuses the search results into the reply content derived from the conversation scenario data to generate a reply, which in this case includes the fused search results.
  • the large model includes a generative model. The large model inputs the conversation scenario data and the search results into a trained generative model, which then fuses the search results to generate a fused search result.
  • the terminal device displays the fused search result generated by the large model to the user, providing a high-quality reply. It is understood that the method of fusion is not a requirement here.
  • the large model is integrated according to the generated content of the dialogue scene data and the retrieval results to obtain a reply.
  • the generated reply includes not only the fused retrieval results, but also the reasons for the reply corresponding to the fused retrieval results.
  • the reply generation capability of the large model can be deployed on the terminal device or on the cloud device.
  • the terminal device can directly display the reply generated by the large model.
  • the reply generation capability of the large model is deployed on the cloud device, the cloud device can generate the reply, that is, the above step S306 can be executed on the cloud. Then, the cloud device sends the reply to the terminal device, and the terminal device displays the received reply.
  • box K32A contains specific content corresponding to the search result: "Searching for ticket information for tonight's movie 1: XXX Cinema (XXX Store): ⁇ 49.9, Tonight's Showtimes: 19:30
  • box K32B contains the search response displayed on the terminal device: "Why don't we go to this place?" It's understandable that the movie-watching content involved in the current conversation is strongly related to movie-related applications.
  • the big model can retrieve valid information from the corresponding application, and the terminal device can present the response generated by the big model in a better way to prevent the user from receiving an accurate and detailed answer or solution.
  • the generative model used can be the same as the generative model used in the above step S303 when retrieval is not required.
  • the training samples used for training include conversation scene sample data involving retrieval intentions, and the obtained retrieval results and conversation scene samples are used as training samples input to the generative model for training, and then training is performed.
  • the training method can be the same as the training method of the generative model in the above step S303, or it can be different, which is not required here.
  • the terminal device displays the generated reply on the screen, such as displaying the generated reply in the form of a card or a list.
  • the generated reply may include at least one data form of text, image, voice, video, and link.
  • the user can edit or send the displayed reply. For example, the user can directly modify the displayed reply, and after the modification is completed, click the send control on the screen for sending the displayed reply to send the modified reply. For example, the user does not modify the displayed reply and can directly send the displayed reply through the send control for sending the displayed reply.
  • the user can also refer to the displayed reply and adaptively modify the input content in the original input box, and send the modified content in the input box through the send control for sending the content in the input box; or directly send the content in the input box through the send control for sending the content in the input box without modifying the input content in the original input box.
  • the user can also refer to the description of user KB sending a reply or the content in the input box in Figures 2A and 2B above, which will not be repeated here.
  • the terminal device obtains conversation scenario data, where the conversation scenario data includes the current and historical conversation content between the two parties, user behavior information, and the current user's physiological indicator data.
  • the terminal device can directly generate a reply using the large model, without having to perform operations such as determining whether a search is required and searching the resource database to obtain search results.
  • the conversation scenario data includes multidimensional data
  • the generated reply can be more accurate based on the multidimensional conversation scenario data.
  • the generated reply can be at least one of a search result, a fused search result, a conversation reply, and a corresponding reply reason.
  • executor of the interactive method shown in FIG3 may also be an application, for example, the above-mentioned dialogue system, which may call other applications to generate replies, or may directly generate replies in real time without calling other applications.
  • the function of generating responses in real time can be configured in the intelligent dialogue service (Agent) of the terminal device.
  • the intelligent dialogue service can operate in combination with the content of multiple rounds of dialogue.
  • the intelligent dialogue service can call the large model and display the response output by the large model.
  • the large model determines whether a search is necessary for the conversation content. If a search is required, the large model combines the search results obtained from local and cloud services with the responses generated by the large model based on the conversation scenario data to generate a fused search result.
  • the client then presents the response to the user in an appropriate manner and saves it locally, providing users with accurate, reasonable, and high-quality responses.
  • the dialogue system can also be directly configured with the function of generating a reply. In this case, there is no need to call the intelligent dialogue service, that is, the dialogue system can directly call the large model to generate a reply.
  • FIG. 4A shows a schematic diagram of a system architecture 400 according to some embodiments of the present application.
  • system architecture 400 is composed of two terminal devices (terminal 100A and terminal 100B) and a cloud device (cloud 200).
  • the terminal devices include but are not limited to mobile devices such as mobile phones and tablets that can be used for communication.
  • the cloud device is composed of at least one server and has large-scale model computing capabilities.
  • the terminal 100A includes a dialogue system 101A, an intelligent dialogue service 102A, and a local search 103A.
  • the dialogue system 101A is used to provide communication services to users.
  • the dialogue system 101A can be an application that can perform instant communication on the terminal 100A, such as a voice assistant, an intelligent assistant, and It is understandable that in other embodiments, the dialogue system 101A, the intelligent dialogue service 102A, and the local search 103A may be different applications, or may be combined into the same application, or any two of them may be combined into the same application, which is not required here.
  • Intelligent dialogue service 102A is used to generate responses based on the dialogue scenario data, invoking the large-scale model computing capabilities of cloud 200 or local search 103A. It is understood that the specific process by which intelligent dialogue service 102A generates responses based on the dialogue scenario data can be referred to steps S301-S305 shown in Figure 3 above and will not be elaborated here.
  • intelligent dialogue service 102A can be a system application on terminal 100A.
  • dialogue system 101A is configured to allow invocation of intelligent dialogue service 102A, dialogue system 101A can invoke intelligent dialogue service 102A upon detecting a user's operation in the input box, and generate a response in real time.
  • the local search 103A is used to search the local resource database according to the local search information to obtain the local search result.
  • terminal 100B includes a dialogue system 101B, an intelligent dialogue service 102B, and a local search engine 103B.
  • the functions of dialogue system 101B, intelligent dialogue service 102B, and local search engine 103B are essentially the same as those of dialogue system 101A, intelligent dialogue service 102A, and local search engine 103A, and are not described in detail here.
  • the cloud 200 includes a large model module 201 and a search engine 202 .
  • the large model module 201 is used to generate a reply based on the conversation scene data.
  • the large model module 201 may include a discriminant model, an intent recognition model, and a generation model.
  • the large model module 201 is used to obtain the conversation scene data from the terminal 100B or the terminal 100A, use the discriminant model, and determine whether a search is required based on the conversation scene data. If a search is required, the intent recognition model is used to identify the specific search intent in the conversation scene data, so as to obtain the search information based on the search intent, thereby enabling the search engine 202 or the local search 103B to perform the search. It is understandable that after the search information is generated, the search engine or third-party service can be called through the cloud service.
  • the cloud service can send the search information to the application programming interface (API) of the search engine 202 through a network protocol (such as HTTP, WebSocket, etc.), and the search engine 202 can also implement the call to the third-party service, or directly call the API of the third-party service to obtain relevant search results or service responses.
  • API application programming interface
  • the large model module 201 is also used to generate responses using the generated model. Specifically, when retrieval is not required, a response is generated directly based on the conversation scene data from the terminal 100A or the terminal 100B; when retrieval is required, a response is generated based on the conversation scene data and the retrieval results.
  • the specific process of the large model module 201 generating responses based on the conversation scene data can be referred to the description of the large model generating responses in FIG. 3 above, which will not be elaborated here.
  • the search engine 202 is used to obtain search results based on the search information sent by the large model module 201 and feed back the search results to the large model module 201. That is, the search engine 202 searches the cloud resource database based on the search information to obtain the search results.
  • system architecture 400 in FIG. 4A is merely a schematic diagram. In other embodiments, more or fewer modules may be included, and the modules may be combined or split, which will not be elaborated herein.
  • the following describes the specific process of the terminal 100B interacting with the cloud 200 to generate a reply during the conversation.
  • FIG4B shows a schematic diagram of the interaction process between the terminal 100B and the cloud 200 according to some embodiments provided by the embodiments of the present application.
  • Terminal 100B detects a first operation and obtains conversation scene data, including conversation content.
  • the conversation scene data may also include information such as user behavior or physiological indicators.
  • terminal 100B detects the first operation and obtains the dialogue scene data when allowed to operate by the user.
  • terminal 100B may obtain conversation scene data, including conversation content, user behavior, physiological indicators and other information, in the following manners.
  • terminal 100B can obtain the text, audio, image, and video information of the current conversation by calling the API of the chat application (such as conversation system B1).
  • the information may include the chat content between the user and the other party, the timestamp of the conversation text, the audio recording, the image content, and the video content. In this way, terminal 100B can obtain the basic content of the conversation for subsequent analysis and processing.
  • terminal 100B needs to obtain user behavior information. For example, when a user communicates with another party in the current dialogue system, the user's behavior information such as browsing and editing chat records can be obtained based on the current dialogue system. Terminal 100B can obtain screen recording permissions to record the user's browsing history and editing operations in the chat application to track the user's operation behavior, thereby obtaining the user's behavioral characteristics during the dialogue process, and thus more comprehensively understanding the current dialogue scene.
  • the terminal 100B can be connected to some other devices that collect user status, such as watches, bracelets, etc. At this time, the terminal 100B can interact with the application of the aforementioned collection device and obtain corresponding physiological indicator data, such as the user's expression, heartbeat, etc. In addition, the terminal device 100B can also obtain some sensor data based on built-in sensors, such as camera data, to obtain the user's current expression. This data can provide more information about the user's emotions and reactions, which helps to more accurately analyze and process conversation scenarios.
  • terminal 100B will only collect and use necessary information and comply with relevant privacy protection regulations.
  • terminal 100B sends the integrated conversation scene data to the cloud 200.
  • step S402 is an optional step.
  • S402 can also be performed by the cloud 200.
  • the terminal 100B may organize and integrate the collected conversation scene data according to a time sequence relationship, and send it to the cloud 200 for processing.
  • the conversation scene data includes information about all aspects of the conversation.
  • Terminal 100B can now fully understand the context of the conversation, enabling the subsequent large model in cloud 200 to generate more accurate and personalized responses and suggestions, thereby enhancing the intelligence of the conversation system and improving the user experience.
  • terminal 100B before sending the collected conversation scene data to the cloud 200, terminal 100B needs to convert the format of the collected conversation scene data and organize it into a transmittable data structure.
  • the specific data format will be determined based on actual needs and system design. Common data exchange formats such as JSON (javascript object notation) or Protocol Buffers can generally be used. It is understood that the actual data format can be adjusted based on specific needs and designs and is not limited here.
  • conversation scenario mentioned above includes but is not limited to (1) conversation content, which involves text, pictures, audio, and video information, (2) user behavior information, such as browsing chat records, editing information, etc., and (3) physiological indicator data, such as user expressions, heartbeat, and other physiological data, and these data will use different data formats and units depending on the specific device and application.
  • the sent data includes three messages between user B-2 and user A-2 as shown in Figure 2A, and a pre-reply from user B-2.
  • Conversation message sent content: ⁇ "timestamp”:2023-10-31T17:30:00;"sender”:"User A-2";"content”:"What are your plans after get off work today?" ⁇ "timestamp”:2023-10-31T17:30:10;”sender”:"User B-2";"content”:"I don't have any plans tonight" ⁇ "timestamp”:2023-10-31T17:30:20;”sender”:"User A-2";”content”:”So?” ⁇ ;
  • the data sent includes the timestamp of the conversation, the sender, and the content, all of which are recorded in the message list. Furthermore, it also includes the user's browsing history and user psychological indicators. This browsing history data is not specifically displayed in the current example.
  • the user's psychological indicators include a "neutral" expression and a heart rate of 80 beats per minute (bpm). This information can be further processed and analyzed to provide more personalized and accurate responses and suggestions.
  • the cloud 200 determines whether a search is required based on the conversation scenario data.
  • steps S404-S406 are optional steps.
  • the cloud 200 may determine whether a search is required through the large model. If the determination result output by the large model is an indicator indicating that a search is not required, the process proceeds to S404; otherwise, the process proceeds to S407.
  • cloud 200 may use natural language processing (NLP) technology to analyze the conversation. It may then discover that user A-2 is inquiring about user B-2's schedule, to which user B-2 responds that she has no schedule. Based on the content of this conversation segment, cloud 200 may analyze that no search is required, and therefore no search is necessary. Furthermore, because user KB did not explicitly reply with search-related content in the pre-sent message, no search results can be found in the corresponding resource database. For example, no targeted search for third-party application services is possible, and therefore no search is required. The process proceeds to S404 .
  • NLP natural language processing
  • cloud 200 will determine the need for a search based on the conversation scenario data.
  • the cloud 200 generates a response based on the conversation scenario data, including the conversation response and the reason for the response.
  • cloud 200 can generate accurate and coherent responses by combining the large model with the conversation context corresponding to the conversation scenario data.
  • the large model can use conversation understanding technology and semantic understanding algorithms to understand the user's questions in the conversation scenario data, recommend responses that meet the user's intent and take into account factors such as the other party's emotion and tone, and provide a satisfactory answer.
  • cloud 200 analyzes user A-2's question “So?” and user B-2's previous message, "I don't have anything to do tonight," to determine that user A-2 is asking user B-2 about their evening plans, hoping to do something together. Therefore, user B-2's pre-sent message suggests a rejection, potentially irritating the other person.
  • the generative model can generate a conversational response: "Then let's go play together! and a reason for the response: "The other person is asking about your evening plans because they're looking forward to doing something together. Your current response may irritate the other person.”
  • the cloud 200 sends a reply to the terminal 100B.
  • the cloud 200 can generate responses to the user's pre-sent content and existing chat content based on relevant algorithms or rules, and the generated responses take into account emotional judgment and factors that avoid causing resentment to the other party. After the response content is generated, it can be processed according to a certain format to obtain formatted transmission data and sent to terminal 100B.
  • the generated reply and other related information may be combined into a JSON object.
  • a JSON object may include the specific content of the reply and a timestamp.
  • Terminal 100B displays a reply.
  • terminal 100B displays the received reply, and the displayed reply includes the dialogue reply and the reason for the reply.
  • the terminal 100B parses the received JSON object and extracts the required fields, such as the conversation reply, timestamp, and reply reason in the JSON object, and then displays them on the interface to provide corresponding feedback to the user, thereby continuing the conversation.
  • the cloud 200 can efficiently transmit the reply and related information to the terminal 100B, so that the application or platform on the terminal 100B can flexibly process and present the response results.
  • the cloud 200 identifies the search intent.
  • the cloud 200 after determining that the search requirements are met, for example, the search requirements are to book a flight or watch a movie, the cloud 200 performs intent recognition based on the conversation scenario data through a large model to obtain a search intent that can be used for retrieval.
  • the cloud 200 generates search intentions related to watching a movie and booking air tickets, respectively, based on the conversation scenario data.
  • the cloud 200 generates search information according to the search intent.
  • the cloud 200 can generate search information, such as query parameters, to facilitate search as a search engine or local search based on the search intent. It is understood that since the data resource library available for search includes the terminal's data resource library and the cloud's data resource library, the generated search information includes local search information or cloud search information. For example, the local search information is the local query parameter, and the cloud search information is the cloud query parameter.
  • the format of the retrieved information can be JSON format or other formats, which is not required here.
  • the local search information may include keywords related to the search intent and search sentences related to the search intent.
  • the specific parameters involved include: query intent, user input (the user input is a search sentence generated by the cloud 200 that can be directly used for local search), the object of the query, and the context information required for the query, such as user identity, terminal device platform and application, etc.
  • the user input is a search phrase used to search in the cloud 200, for example, the search engine 202.
  • the cloud has third-party application services, such as third-party applications for travel or flight booking, searches can be performed on services deployed in the cloud in applications related to travel or flight booking.
  • step S409 and steps S410-S412 can be optional steps respectively.
  • the cloud 200 obtains a cloud search result based on the cloud search information.
  • the cloud 200 searches the cloud resource database based on the cloud search information to obtain search results. Specifically, the cloud 200 parses and processes the cloud search information using natural language processing (NLP) technology to generate a parsed query statement for use in the query to obtain cloud search results.
  • NLP natural language processing
  • the cloud search results include information such as a list of relevant search results, the total number of results, and the search engine's response time.
  • the cloud 200 sends local search information to the terminal 100B.
  • the cloud 200 may send local search information to the terminal 100B via a network protocol (such as HTTP, WebSocket, etc.).
  • a network protocol such as HTTP, WebSocket, etc.
  • terminal 100B obtains local search results based on local search information.
  • terminal 100B searches the local resource database based on the local search information to obtain local search results.
  • the local resource database may include local applications and local data.
  • Local applications are various software programs installed on a device and are typically stored in the device's local storage space.
  • Local data including data related to the device itself, such as device settings, user preferences, contacts, text messages, call logs, photos, videos, etc., is typically stored in the device's local storage space.
  • the terminal 100B uses a database query language (such as SQL, NoSQL query, etc.) or a local resource search algorithm based on the local retrieval information to search and match in the local resource database, and then obtains the local retrieval results.
  • a database query language such as SQL, NoSQL query, etc.
  • a local resource search algorithm based on the local retrieval information to search and match in the local resource database, and then obtains the local retrieval results.
  • the terminal 100B searches for information related to air tickets from Beijing to Sanya in the local application and the local data according to the local search information.
  • the terminal 100B sends the local search result to the cloud 200 .
  • the cloud 200 generates a reply based on the local and/or cloud search results and the conversation scene data, and the reply includes the fusion search results.
  • the cloud 200 generates a response based on the local search results and/or cloud search results, as well as the conversation scenario data using a large model.
  • the generated response includes a fused search result corresponding to the search requirements.
  • the fused search results may include: “The search system is searching for travel information and air ticket information for September 13th, including a link for a flight from 2:30 PM to 6:25 PM, priced at ⁇ 1,310, and a link for a flight from 6:30 AM to 10:20 AM, priced at ⁇ 1,680.” and “Great! I’m really looking forward to this trip!”
  • the cloud 200 parses the return body corresponding to the cloud search result and the information sent by the terminal 100B based on the return body obtained from the cloud search, for example, the return body from the cloud search engine and the information sent from the terminal 100B, constructs the input parameters of the large model based on the parsed content, and generates a reply based on the input parameters.
  • the return body corresponding to the cloud search results includes: the search result list, the total number of results and the response time of the search engine.
  • Parsing the return body corresponding to the cloud search results includes: parsing the return body corresponding to the cloud search results, extracting the search result list and other related information, such as the total number of results.
  • the input parameters for constructing a large model based on the parsed content include: constructing the input parameters of the large model based on the return body corresponding to the cloud search results, the input parameters include the search result list, the total number of results and other necessary fields, among which the specific content and form of the input parameters can be adjusted according to the specific large model and application scenario.
  • Generating a response based on input parameters includes: the large model generates a recommended response based on the constructed input parameters. It is understood that the large model can generate relevant recommended response content based on the search results, the total number of results, and other contextual information, combined with its own model weights and trained knowledge. It is understood that the recommended response content can be a specific answer to the user's query, suggestions, relevant resource links, or other information, customized according to the specific application scenario.
  • Flight Number The unique identifier of the flight
  • Departure City The name or code of the departure city
  • Departure Time The departure time
  • Arriv City The name or code of the arrival city
  • Airline The name or code of the operating airline
  • Price The price of the ticket.
  • Booking Interface Link A link to the flight booking app. This link directly opens the corresponding page in the app for booking tickets.
  • the booking interface link is a link provided to the client-side to display the relevant booking interface.
  • a template filling method can be used to fill in some information.
  • a pre-defined template structure can be defined and filled in the corresponding positions to generate human-readable replies and suggestions.
  • the generative model in the large model can be integrated based on the relevant search results (i.e. search results) obtained from the local machine (i.e. the terminal side) and the cloud as well as the original conversation scenario data, that is, the search results from multiple different sources can be integrated together, so that the generated replies can not only include and display the content corresponding to the precise needs to the user, but also provide users with high-quality reply content, for example, displaying recommendations related to the App.
  • search results i.e. search results
  • the local machine i.e. the terminal side
  • the cloud as well as the original conversation scenario data
  • the generated replies can not only include and display the content corresponding to the precise needs to the user, but also provide users with high-quality reply content, for example, displaying recommendations related to the App.
  • the cloud 200 sends a reply to the terminal 100B.
  • terminal 100B displays a reply.
  • the terminal 100B receives the reply and displays the reply. It is understandable that the terminal 100B can use the terminal side or cloud side resources and capabilities to provide the user with a comprehensive and accurate reply.
  • terminal 100A and terminal 100B are devices that display smart replies in real time.
  • the terminal device can display smart replies.
  • the conversation on terminal 100B is further elaborated below.
  • terminal device 100B can call smart conversation service 102B.
  • smart conversation service 102B can understand user KB's preference for travel destinations and then provide travel information about Dali, such as attraction introductions and food recommendations, based on local and cloud data resources, to provide user KB with reference.
  • Smart conversation service 102B can continue the conversation with user KA to learn more about itinerary plans, activity arrangements, orders, etc., to provide more personalized suggestions and services.
  • intelligent conversation service 102B can leverage the conversation context to infer the user's interest in departure time. Based on this information, intelligent conversation service 102B can invoke cloud services, such as a flight query interface, to provide the user with relevant information such as departure date, flight time, and price.
  • Figure 5A illustrates the data flow that intelligent conversation service 102B retrieves from the cloud when a search is required, and further details will not be provided here.
  • intelligent conversation service 102A recognizes the user's intent and further leverages conversation understanding technology and intent recognition algorithms to retrieve relevant travel information from the user's local device and the cloud. By integrating travel planning tools, flight search, and hotel reservations, intelligent conversation service 102A can provide user KA with travel recommendations, such as recommended itineraries, scenic spots, and destinations suitable for Hainan.
  • the interactive method proposed in the embodiments of this application can monitor conversations in real time, thereby avoiding the traditional practice of responding based on user input. Furthermore, by monitoring multiple inputs from the conversation partner in real time, including existing conversation content, pre-sent content, and factors such as the interlocutor's physiological indicators and behavior, comprehensive analysis and processing of multiple inputs, combined with previous conversation history information, can more comprehensively understand the needs and intentions of the conversation participants and perform comprehensive reasoning in conjunction with a large model. This enables intelligent conversation services to provide ideal responses to conversation participants, creating a more ideal and harmonious conversation experience.
  • FIG6 shows a hardware module diagram of a terminal 100 and a cloud 200 according to some embodiments of the present application.
  • the terminal 100 includes a first display module 501 , a first input module 502 , a first storage module 503 , a first calculation module 504 , and a first communication module 505 .
  • the cloud 200 includes a second storage module 601 , a second computing module 602 , and a second communication module 603 .
  • terminal 100 Specifically, for terminal 100:
  • the first display module 501 is used to output and display the interface.
  • the first display module 501 can display the responses generated by the intelligent dialogue service to the user in a visual form, which can be in the form of data such as text, images, videos, audio, and links. It can be understood that through the first display module 501, the user can intuitively view and understand the responses provided by the intelligent dialogue service.
  • the first input module 502 is used for user interaction with the dialogue system.
  • the first input module 502 can be a keyboard, touch screen, or voice recognition device. Through the first input module 502, users can ask questions, express needs, or provide relevant information to the intelligent dialogue service. Users can communicate with the dialogue system by inputting text on the keyboard, using touch screen taps or gestures, or using voice input.
  • the first storage module 503 is used to store various data and information required by the dialogue system, including parameters related to the intelligent dialogue service, local content, information about third-party applications on the local machine, local search results, etc. This data and information can be repeatedly called and referenced by the dialogue system or intelligent dialogue service for analysis, response generation, search, etc., thereby providing accurate and useful responses.
  • the first computing module 504 is used to process, analyze and calculate input data and information. It can be a central processing unit, including components such as a processor and memory, and is also responsible for processing tasks such as local search.
  • the first communication module 505 is used to connect and communicate with other devices or networks, and can support wireless or wired communication. Through the first communication module 505, the terminal 100 can connect to the Internet to access search capabilities, obtain the latest information and data, etc. The first communication module 505 can also transmit and interact with other devices or systems to achieve a wider range of functions and services.
  • the second storage module 601 is used to save and manage large amounts of data, models and other related information, which may include historical search data, user profiles, system profiles, model parameters, training data sets, etc. It can also be used for long-term storage and backup of data, as well as support the training, updating and deployment of large models.
  • the second computing module 602 is the core computing resource of the cloud 200 and is typically composed of a group of high-performance computers or server clusters equipped with powerful processors and graphics processing units. These computing resources are used to perform large-scale complex computing tasks, such as training machine learning models, executing and inferring deep learning algorithms, and large-scale data processing.
  • the second computing module 602 can provide higher computing power and parallelism, enabling the system to more efficiently perform advanced computing tasks such as intent recognition and recommended replies.
  • the second computing module 602 can use its powerful computing power and large-scale trained models to conduct more in-depth analysis and processing of user questions, generating more advanced, accurate, and personalized recommended replies.
  • the second communication module 603 is responsible for remote communication and data transmission with the terminal device.
  • the second communication module 603 connects to the terminal device via the internet to receive queries and requests from the terminal device in real time. Once a query or request is received, the communication device transmits it to the cloud computing device for processing and returns the processed results to the terminal device.
  • the communication device is also responsible for maintaining the connection with the terminal device to ensure stable and secure communication.
  • Figure 6 illustrates the data flow when no retrieval is required based on the conversation scenario data
  • Figure 7 illustrates the data flow when retrieval is required based on the conversation scenario data.
  • the first input module 502 receives the user's input and obtains input information, such as the user using a keyboard, A touch screen or voice recognition device interacts with terminal device 100B, and the first input module 502 receives input information such as text, voice, or video.
  • the first storage module 503 stores the input information and data collected by other collection modules, such as user behavior information and physiological indicators, to obtain current conversation scene data and send the current conversation scene data to the first calculation module 504 for processing.
  • the first calculation module 504 can perform preliminary processing on the conversation scene data to obtain processed conversation scene data, and then send the processed conversation scene data to the first communication module 505.
  • the first communication module 505 sends a request to the second communication module 603 in the cloud, including the conversation scene data.
  • the second communication module 603 sends the conversation scene data to the second calculation module 602. After determining that a search is not necessary, the second calculation module 602 directly generates a reply and sends the reply to the second communication module 603. The second communication module 603 then sends the reply to the first communication module 505. Under the control of the first calculation module 504, the first communication module 505 sends the reply to the first display module 501, which displays the reply.
  • the second computing module 602 when the second computing module 602 determines that a search is needed, it will generate a search intent. At this time, local search information and cloud search information can be generated, so that searches can be performed on the terminal and the cloud respectively.
  • the second computing module 602 obtains a second search result from the second storage module 601 based on the cloud search information.
  • the second computing module 602 sends the local search information to the first computing module 504 through the second communication module 603 and the first communication module 505.
  • the first computing module 504 searches from the first storage module based on the local search information to obtain a local search result. It can be understood that the results of the local search are mainly based on existing data and information, and can provide some quick and simple answers.
  • the second calculation module 602 integrates the local search results and the cloud search results to generate a reply.
  • the interaction method provided in the embodiment of the present application can select corresponding algorithms and technical tools according to specific needs and technical platforms during actual implementation.
  • privacy and data security should also be properly handled to protect the user's personal information.
  • more complex technical designs and improvements may be required.
  • the first computing module 504 on the terminal side can also be used to fuse the search results with the model response obtained by the large model based on the dialogue scenario data to obtain a response.
  • the specific implementation method is not required here.
  • FIG8 shows a structural diagram of an apparatus 800 according to some embodiments of the present application. It is understandable that the apparatus 800 can be a terminal device or a cloud device.
  • the apparatus 800 may include one or more processors 801, which may also be referred to as a processing unit, and may implement certain control functions.
  • the processor 801 may be a general-purpose processor or a dedicated processor, etc. For example, it may be a baseband processor or a central processing unit.
  • the baseband processor may be used to process communication protocols and communication data
  • the central processing unit may be used to control a communication device, such as a base station, a baseband chip, a terminal, a terminal chip, a DU or a CU, etc., to execute software programs and process data of the software programs.
  • the processor 801 may also store instructions and/or data 803, and the instructions and/or data 803 can be executed by the processor so that the device 800 executes the interaction method described in the above method embodiment.
  • processor 801 may include a transceiver unit for implementing receiving and transmitting functions.
  • the transceiver unit may be a transceiver circuit, an interface, or an interface circuit.
  • the transceiver circuit, interface, or interface circuit for implementing the receiving and transmitting functions may be separate or integrated.
  • the transceiver circuit, interface, or interface circuit may be used for reading and writing code/data, or may be used for transmitting or delivering signals.
  • the apparatus 800 may include a circuit that can implement the functions of the interactive method in the aforementioned method embodiment.
  • the device 800 may include one or more memories 802, on which instructions/data 804 may be stored.
  • the instructions may be executed on a processor, causing the device 800 to perform the method described in the above method embodiment.
  • the memory may also store data.
  • the processor may also store instructions and/or data.
  • the processor and memory may be provided separately or integrated. For example, the corresponding relationship described in the above method embodiment may be stored in the memory or in the processor.
  • the apparatus 800 may further include a transceiver 805 and/or an antenna 806.
  • the processor 801 may be referred to as a processing unit, and controls the apparatus 800.
  • the transceiver 805 may be referred to as a transceiver unit, a transceiver, a transceiver circuit, a transceiver device, an interface, an interface circuit, or a transceiver module, and is configured to implement transceiver functions.
  • the device 800 in the embodiment of the present application can be used to execute the interaction method described in Figures 3 and 4B in the embodiment of the present application.
  • the present application also provides a computer program product, which includes: computer program code, when the computer program code is run on a computer, enables the computer to implement the steps performed by the device 800 in any one of the above embodiments.
  • the present application also provides a computer-readable medium, which stores program code.
  • the program code runs on a computer, the computer implements the steps performed by the device 800 in any of the above embodiments.
  • the various embodiments disclosed in this application can be implemented in hardware, software, firmware, or a combination of these implementation methods.
  • the embodiments of the present application can be implemented as a computer program or program code executed on a programmable system, which includes at least one processor, a storage system (including volatile and non-volatile memory and/or storage elements), at least one input device, and at least one output device.
  • Program code can be applied to input instructions to perform the functions described herein and generate output information.
  • the output information can be applied to one or more output devices in a known manner.
  • a processing system includes any system having a processor such as, for example, a digital signal processor (DSP), a microcontroller, an application specific integrated circuit (ASIC), or a microprocessor.
  • DSP digital signal processor
  • ASIC application specific integrated circuit
  • Program code can be implemented with a high-level programming language or an object-oriented programming language to communicate with the processing system. Where necessary, program code can also be implemented in assembly language or machine language. In fact, the mechanism described in this application is not limited to the scope of any particular programming language. In either case, the language can be a compiled language or an interpreted language.
  • the disclosed embodiments may be implemented in hardware, firmware, software, or any combination thereof.
  • the disclosed embodiments may be implemented as instructions carried or stored on one or more temporary or non-temporary machine-readable (e.g., computer-readable) storage media, which may be read and executed by one or more processors.
  • the instructions may be distributed over a network or through other computer-readable media.
  • a machine-readable medium may include any mechanism for storing or transmitting information in a form readable by a machine (e.g., a computer), including but not limited to floppy disks, optical disks, optical discs, read-only memories (CD-ROMs), magneto-optical disks, read-only memories (ROMs), random access memories (RAMs), erasable programmable read-only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), magnetic or optical cards, flash memory, or a tangible machine-readable memory for transmitting information (e.g., carrier waves, infrared signals, digital signals, etc.) using the Internet in electrical, optical, acoustic, or other forms of propagation signals. Therefore, a machine-readable medium includes any type of machine-readable medium suitable for storing or transmitting electronic instructions or information in a form readable by a machine (e.g., a computer).
  • a logical unit/module can be a physical unit/module, or a part of a physical unit/module, or can be implemented as a combination of multiple physical units/modules.
  • the physical implementation of these logical units/modules themselves is not the most important.
  • the combination of functions implemented by these logical units/modules is the key to solving the technical problems raised by this application.
  • the above-mentioned device embodiments of this application do not introduce units/modules that are not closely related to solving the technical problems raised by this application. This does not mean that other units/modules do not exist in the above-mentioned device embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Human Computer Interaction (AREA)
  • Databases & Information Systems (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

La présente demande se rapporte au domaine technique des ordinateurs et en particulier à un procédé d'interaction, un dispositif électronique, un support de stockage lisible et un produit programme. Le procédé consiste à : acquérir d'abord des données de scène de conversation de deux parties de conversation ; puis déterminer si les données de scène de conversation doivent être récupérées ; si tel n'est pas le cas, générer directement une réponse sur la base des données de scène de conversation ; et si tel est le cas, effectuer une récupération précise à partir d'une base de données de ressources pour obtenir un résultat de récupération requis, ce qui permet de générer une réponse sur la base du résultat de récupération obtenu. On comprend que l'on détermine que la récupération doit être effectuée, des informations spécifiques sont récupérées à partir de la base de données de ressources lorsque la récupération doit être effectuée, et une réponse est générée sur la base du résultat de récupération, c'est-à-dire que le contenu de réponse comprend le contenu du résultat de récupération, de sorte que la réponse générée soit plus précise, ce qui permet d'améliorer l'expérience d'un utilisateur à l'aide du système de conversation.
PCT/CN2024/123731 2024-02-26 2024-10-09 Procédé d'interaction, dispositif électronique, support de stockage lisible et produit programme Pending WO2025179890A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202410211616.7A CN120578733A (zh) 2024-02-26 2024-02-26 一种交互方法、电子设备、可读存储介质和程序产品
CN202410211616.7 2024-02-26

Publications (2)

Publication Number Publication Date
WO2025179890A1 true WO2025179890A1 (fr) 2025-09-04
WO2025179890A9 WO2025179890A9 (fr) 2025-11-06

Family

ID=96858435

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2024/123731 Pending WO2025179890A1 (fr) 2024-02-26 2024-10-09 Procédé d'interaction, dispositif électronique, support de stockage lisible et produit programme

Country Status (2)

Country Link
CN (1) CN120578733A (fr)
WO (1) WO2025179890A1 (fr)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090327263A1 (en) * 2008-06-25 2009-12-31 Yahoo! Inc. Background contextual conversational search
CN110096516A (zh) * 2019-03-25 2019-08-06 北京邮电大学 自定义的数据库交互的对话生成方法及系统
CN110287297A (zh) * 2019-05-22 2019-09-27 深圳壹账通智能科技有限公司 对话答复方法、装置、计算机设备及计算机可读存储介质
CN113806503A (zh) * 2021-08-25 2021-12-17 北京库睿科技有限公司 一种对话融合方法和装置及设备
CN116049360A (zh) * 2022-11-29 2023-05-02 兴业银行股份有限公司 基于客户画像的智能语音对话场景话术干预方法及系统
CN116521841A (zh) * 2023-04-18 2023-08-01 百度在线网络技术(北京)有限公司 用于生成回复信息的方法、装置、设备及介质
CN116595150A (zh) * 2023-06-16 2023-08-15 上海微盟企业发展有限公司 一种对话推荐方法、装置、设备及存储介质

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090327263A1 (en) * 2008-06-25 2009-12-31 Yahoo! Inc. Background contextual conversational search
CN110096516A (zh) * 2019-03-25 2019-08-06 北京邮电大学 自定义的数据库交互的对话生成方法及系统
CN110287297A (zh) * 2019-05-22 2019-09-27 深圳壹账通智能科技有限公司 对话答复方法、装置、计算机设备及计算机可读存储介质
CN113806503A (zh) * 2021-08-25 2021-12-17 北京库睿科技有限公司 一种对话融合方法和装置及设备
CN116049360A (zh) * 2022-11-29 2023-05-02 兴业银行股份有限公司 基于客户画像的智能语音对话场景话术干预方法及系统
CN116521841A (zh) * 2023-04-18 2023-08-01 百度在线网络技术(北京)有限公司 用于生成回复信息的方法、装置、设备及介质
CN116595150A (zh) * 2023-06-16 2023-08-15 上海微盟企业发展有限公司 一种对话推荐方法、装置、设备及存储介质

Also Published As

Publication number Publication date
WO2025179890A9 (fr) 2025-11-06
CN120578733A (zh) 2025-09-02

Similar Documents

Publication Publication Date Title
US11303590B2 (en) Suggested responses based on message stickers
US11611565B2 (en) Systems and methods for providing an interactive media presentation
CN107846350B (zh) 一种语境感知网络聊天的方法、计算机可读介质和系统
CN111480159B (zh) 用于增强的企业列表的知识搜索引擎平台
EP3577610B1 (fr) Association de réunions à des projets à l'aide de mots-clés caractéristiques
CN111641514B (zh) 会议智能系统、用于会议智能的方法和存储介质
US20200005248A1 (en) Meeting preparation manager
US10078489B2 (en) Voice interface to a social networking service
KR101884881B1 (ko) 대화식 질문 및 응답
US8244589B2 (en) Personalized audio controlled shopping information service for a mobile device
US20190087707A1 (en) Artificial conversational entity methods and systems
US10824798B2 (en) Data collection for a new conversational dialogue system
JP2019508820A (ja) メッセージ交換スレッド用の自動提案
JP6968897B2 (ja) 登録されていないリソースによるオーディオベースのネットワークセッションの確立
US20180189017A1 (en) Synchronized, morphing user interface for multiple devices with dynamic interaction controls
CA3029720A1 (fr) Systemes et methodes d'assistant vocal de voisinage
US20250131359A1 (en) Augmenting chat-based workflows with large language models
US20180188896A1 (en) Real-time context generation and blended input framework for morphing user interface manipulation and navigation
CN114239604A (zh) 在线咨询处理方法、装置及计算机设备
CN119205133B (zh) 信息处理方法、设备、存储介质及程序产品
WO2025179890A1 (fr) Procédé d'interaction, dispositif électronique, support de stockage lisible et produit programme
CN100399329C (zh) 移动智能导航引擎系统
TWI897095B (zh) 自然語言訊息處理方法與系統
TWI882526B (zh) 智能對話導入方法與系統
CN121071198A (zh) 交互处理方法及电子设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 24926756

Country of ref document: EP

Kind code of ref document: A1