[go: up one dir, main page]

EP3828781A1 - Method and apparatus for generating question and answer dataset based on input paragraph - Google Patents

Method and apparatus for generating question and answer dataset based on input paragraph Download PDF

Info

Publication number
EP3828781A1
EP3828781A1 EP19212219.0A EP19212219A EP3828781A1 EP 3828781 A1 EP3828781 A1 EP 3828781A1 EP 19212219 A EP19212219 A EP 19212219A EP 3828781 A1 EP3828781 A1 EP 3828781A1
Authority
EP
European Patent Office
Prior art keywords
answer
question
generating
dataset
latent variable
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
EP19212219.0A
Other languages
German (de)
French (fr)
Inventor
Dong Hwan Kim
Woo Tae Jeong
Ah-rim SOHN
Gangho DO
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
42 Maru Inc
Original Assignee
42 Maru Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 42 Maru Inc filed Critical 42 Maru Inc
Priority to EP19212219.0A priority Critical patent/EP3828781A1/en
Priority to KR1020200119570A priority patent/KR102456994B1/en
Publication of EP3828781A1 publication Critical patent/EP3828781A1/en
Priority to KR1020220133034A priority patent/KR102655809B1/en
Priority to KR1020240045399A priority patent/KR20240049528A/en
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/3332Query translation
    • G06F16/3334Selection or weighting of terms from queries, including natural language queries
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation
    • G06N5/022Knowledge engineering; Knowledge acquisition
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/3331Query processing
    • G06F16/334Query execution
    • G06F16/3347Query execution using vector based model
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/951Indexing; Web crawling techniques
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/51Translation evaluation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/58Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • G06N3/0442Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • G06N3/0455Auto-encoder networks; Encoder-decoder networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0475Generative networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/094Adversarial learning

Definitions

  • the present invention discloses a method and an apparatus of generating various datasets by simultaneously generating and sampling a question-answer pair that may be generated in one passage.
  • Patent Literature 1 Korean Patent No. 10-1605430 (published on March 22, 2016 )
  • An object of the present invention is to provide a method and an apparatus of generating a question-answer generating model that simultaneously generates a question-answer dataset capable of being generated based on an input passage even though an answer candidate group is not provided unlike an existing question-answer (QA) model in which a user manually collects and learns a joint distribution of question-answer pairs, and a question-answer learning model capable of being automatically learned using the question-answer generating model.
  • QA question-answer
  • a method of generating a question-answer dataset based on a passage input may include: generating a hidden representation for each word in the input passage and storing the hidden representation in a memory; sampling a latent variable based on the hidden representation; calculating importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector; generating an answer span based on the latent variable and the weighted sum vector; and generating a question based on the answer span and an attention for the hidden representation in the memory.
  • the method may further include: storing a dataset of the generated question and an answer in the memory; and machine-learning a question-answer learning model using the stored data set of the question and the answer.
  • the method may further include: generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting a loss of the question-answer learning model and constraints; and performing adversarial machine learning using a dataset of a question and an answer generated using the generated latent variable.
  • the input passage may be generated by structuring infobox data extracted from data crawled from a website.
  • a paragraph standardized based on a data frequency in a table recognized through column recognition may be input as the passage.
  • the input passage may be data for which machine translation is completed, and the method may further include evaluating a quality of the machine translation by evaluating accuracy of an answer based on a dataset of the generated question and the answer.
  • an apparatus of generating a question-answer dataset may include: one or more processors; and one or more memories in which instructions, when executed by the one or more processors, causing the one or more processors to perform an operation are stored, wherein the operation performed by the one or more processors includes: an operation of generating a hidden representation for each word in an input passage and storing the hidden representation in a memory; an operation of sampling a latent variable based on the hidden representation; an operation calculating importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector; an operation of generating an answer span based on the latent variable and the weighted sum vector; and an operation of generating a question based on the answer span and an attention for the hidden representation in the memory.
  • the operation performed by the one or more processors may further include: an operation of storing a dataset of the generated question and an answer in the memory; and an operation of machine-learning a question-answer learning model using the stored data set of the question and the answer.
  • the operation performed by the one or more processors may further include: an operation of generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting a loss of the question-answer learning model and constraints; and an operation of performing adversarial machine learning using a dataset of a question and an answer generated using the generated latent variable.
  • a program stored in a medium in order to execute generation of a question-answer dataset based on a passage input in combination with a computer which is hardware is provided.
  • spatially relative terms “below”, “beneath”, “lower”, “above”, “upper”, and the like, may be used in order to easily describe correlations between one component and other components.
  • the spatially relative terms are to be understood as terms including different directions of the components at the time of use or at the time of operation in addition to directions illustrated in the drawings.
  • a component described as “below” or “beneath” another component may be placed “above” the other component. Therefore, an exemplary term “below” may encompass both of directions of above and below.
  • Components may be oriented in other directions as well, and thus, spatially relative terms may be interpreted according to orientations.
  • FIG. 1 is a conceptual diagram for schematically escribing a question-answer dataset generating model based on a passage input according to an exemplary embodiment.
  • a question-answer dataset generating model may include a memory 110, an encoder 120, an attention 130, an answer decoder 140, and a question decoder 150.
  • a passage 101 including several sentences is input to the encoder 120 as it is.
  • the passage 101 may include a phrase, a part, a paragraph, and the like.
  • the encoder 120 generates a hidden representation for each word in the input paragraph 101 and stores the hidden representation in the memory 110.
  • the encoder 120 may construct the hidden representation from an input through a neural network. Therefore, the hidden representation may mean a machine-readable data representation learned in a hidden layer of the neural network, and may be composed of a pair of keys and values.
  • the encoder 120 may store information about a meaning represented by a component (for example, a word) of a sentence as the hidden representation.
  • the encoder 120 performs sampling 102 based on the hidden representation to generate a latent variable 103, and the generated latent variable 103 is input to the answer decoder 140.
  • the answer decoder 140 calculates the importance of each component of the memory 110 through the attention 130 that uses the latent variable 103 as a query to generate a weighted sum vector of the memory 110. For example, each time the answer decoder 140 predicts an output word through the attention 130, the answer decoder 140 may refer to an entire input sentence in the encoder 120 once again. However, the answer decoder 140 does not refer to the entire input sentence in the same ratio, but may attentively refer to an input word portion associated with a word to be predicted at a corresponding time. Therefore, an attention function obtains similarities to 'Keys' obtained from all the hidden representations with respect to given 'Queries'. In addition, the obtained similarities are reflected in the respective 'values' mapped to the keys. In addition, a weighted sum vector may be obtained by adding all the values in which the similarities are reflected.
  • the answer decoder 140 predicts an answer span 104 based on the weighted sum vector.
  • the answer span may mean a start point and an end point where an answer is positioned in the input passage.
  • the question decoder 150 may receive the previously generated answer span 104, and perform calculation of the attention 130 of the memory 110 to finally generate the question 105.
  • a question-answer dataset composed of a pair of the generated question and an answer according to an exemplary embodiment may be stored in the memory 110, and the stored question-answer dataset may be used to machine-learn a question-answer learning model.
  • the dataset generating model automatically generates a dataset of a question and an answer from the input passage, and may thus sample several question-answer pairs with only one passage input, thereby constructing more diverse data as compared with an existing manual input model.
  • FIG. 2 is a flowchart for describing a method of generating a question-answer dataset based on a passage input according to an exemplary embodiment.
  • step S200 the method for generating a question-answer dataset according to an exemplary embodiment generates a hidden representation for each word in an input passage and stores the hidden representation in the memory.
  • the input passage may be generated by structuring infobox data extracted from data crawled from a website.
  • infobox data is not extracted from the crawled data
  • a paragraph standardized based on a data frequency in a table recognized through column recognition may be input as the passage.
  • a quality of the machine translation may be evaluated by evaluating accuracy of the answer based on the generated question-answer dataset.
  • a specific input passage will be described later with reference to FIGS. 3 to 4 .
  • step S210 the method of generating a question-answer dataset according to an exemplary embodiment samples a latent variable based on the hidden representation.
  • step S220 the method of generating a question-answer dataset according to an exemplary embodiment generates a weighted sum vector by calculating the importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query.
  • step S230 the method of generating a question-answer dataset according to an exemplary embodiment generates an answer span based on the latent variable and the weighted sum vector.
  • step S240 the method of generating a question-answer dataset according to an exemplary embodiment generates a question based on the answer span and the attention for the hidden representation in the memory.
  • a dataset of the generated question and an answer may be stored in the memory.
  • a question-answer model may be machine-learned using the stored dataset of the question and the answer.
  • an adversarial question for adversarial machine learning is also automatically generated from the input passage, such that that the question-answer model may be machine-learned to be closer to a real boundary.
  • x, y, c, and z correspond to a question, an answer, a paragraph, and a latent variable, respectively.
  • x,y,c ) is a function used for the encoder of the question-answer model
  • z,y,c ) are functions used for the answer decoder and the question decoder, respectively.
  • p ⁇ ( z ) is a marginal prior distribution of the latent variables
  • q ⁇ ( z ) is a marginal posterior distribution of the latent variables.
  • q ⁇ (z) may generate a distribution of latent variables generated from the input
  • p ⁇ ( z ) may generate an output close to the input using latent variables.
  • x n , y n , c n p ⁇ z ⁇ 1 N ⁇ n 1 N p ⁇ z
  • FIG. 3 is a view for describing an example of a method of generating a question-answer dataset in a paragraph according to an exemplary embodiment.
  • an answer 303 is first generated through sampling 302.
  • an answer 304 paired with the answer may be generated using the generated answer 303.
  • FIG. 4 is a view for describing a method of generating a question-answer dataset in a standardized infobox according to an exemplary embodiment.
  • the question-answer dataset generating model may obtain infobox data 400 through crawling from a website famous as an online knowledge encyclopedia website.
  • the infobox data 400 since the infobox data 400 is standardized, the infobox data 400 may be easily structured in a form of a corpus of questions and answers, and since these knowledge encyclopedia websites provide detailed classification data for fields or targets, an input passage may be generated in a preprocessing step so as to generate and machine-learn a question-answer by targeting a desired field.
  • FIG. 5 is a flowchart for describing a method of generating a question-answer dataset in a non-standardized page according to an exemplary embodiment.
  • step S500 the question-answer dataset construction model recognizes the title and the body of a web page from crawled data.
  • a table is recognized based on column recognition in the body. That is, the table is recognized in order to obtain data having a form similar to an infobox that may be obtained from the standardized site.
  • a paragraph is standardized by confirming whether or not there is a word satisfying a preset data frequency. For example, words that appear repeatedly, such as "country name” and "capital” may be confirmed, and the paragraph may be standardized based on these words.
  • the question-answer dataset generating model may receive the standardized paragraph and generate a question-answer dataset to construct a dataset of various questions and answers in a dataset provided in an information box provided in a standardized form or a dataset provided in a table form on an existing website.
  • the question-answer dataset generating model described above may extract a question-answer dataset from data for which learning and machine translation is completed, evaluate accuracy of an answer corresponding to the question, and utilize an evaluation result to verify a quality of learning or translation, beyond simply generating the question-answer dataset.
  • the question-answer dataset generating model can generate a correct question-answer dataset from the data for which the machine translation is completed, it may be evaluated that machine reading is well performed.
  • a question-answer dataset generating model discloses a method of generating a question-answer dataset that cannot be guessed by the current question-answer learning model through adversarial learning.
  • FIG. 6 is a view for describing a method of improving performance of a question-answer model through adversarial learning according to an exemplary embodiment.
  • the method of generating a question-answer dataset based on a passage input selects the latent variable through the encoder, and a question-answer data pair generated from the sampled latent variable needs to be a question-answer data pair whose correct answer cannot be guessed through a question-answer learning model.
  • the question-answer learning model calibrates a decision boundary by learning the question-answer data pair whose correct answer could not be guessed, and when such a process is repeated, the decision boundary of the question-answer learning model may become close to a true decision boundary, and performs of the question-answer learning model may also be improved.
  • a question-answer dataset for the adversarial learning may be determined by adjusting a loss of the question-answer learning model and constraints for positioning the latent variable in the data manifold.
  • the question-answer learning model in the adversarial manner described above may have an objective function of the following Equation 3: min ⁇ max ⁇ E x , y , c ⁇ p x y c L ⁇ x y c + E c u ⁇ p u c L ⁇ x u y u c u ⁇ D KL p ⁇ z
  • x, y, c, and z correspond to a question, an answer, a paragraph, and a latent variable, respectively.
  • c ) is a function used for a pre-processor of the question-answer model
  • c) is a function used for an agent
  • z,c ) is functions used for the answer decoder and the question decoder, respectively.
  • the encoder samples the latent variable ( z ), and a question and a correct answer generated from the sampling need to increase a loss ( ) of the question-answer model.
  • constraints are required in selecting z. That is, as z becomes distant from the previously learned distribution of z, an invalid question-answer pair is generated. Therefore, in order to prevent this problem, a value may be adjusted in such a manner that a penalty is given when a distance between the two distributions becomes distant (for example, use a KL divergence).
  • the question-answer learning model may generate a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting the loss and the constraints.
  • performance of the question-answer earning model may be improved by performing adversarial machine learning using a dataset of an adversarial question and answer generated using the generated latent variable.
  • FIG. 7 is a block diagram schematically illustrating an internal configuration of an apparatus of generating a question-answer dataset according to an exemplary embodiment.
  • an apparatus 100 of generating a question-answer dataset may include one or more memories 110 and a processor 190. Operations of the apparatus 100 of generating a question-answer dataset may be performed by executing a program stored in the memory 110 through the processor 190.
  • the processor 190 performs an operation of generating a hidden representation for each word in the input passage and storing the hidden representation in the memory 110.
  • the processor 190 samples a latent variable based on the hidden representation, and calculates the importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector.
  • the processor 190 may perform an operation of generating an answer span based on the latent variable and the weighted sum vector and generating a question based on the generated answer span and the attention for the hidden representation in the memory.
  • the processor 190 may further perform an operation of storing a dataset of the generated question and an answer in the memory 110 and a machine learning operation of machine-learning a question-answer learning model using the stored dataset of the question and the answer.
  • the processor 19 may further perform an operation of generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting the loss of the question-answer learning model and the constraints and an operation of performing adversarial machine learning using the question-answer dataset generated using the generated latent variable.
  • the apparatus 100 of generating a question-answer dataset described above may include one or more processors 190 and/or one or more memories 110.
  • each of the processors 190 may include an encoder 120, an answer decoder 140, and a question decoder 150, and may perform the operations described above with reference to FIGS 1 to 7 on each of the encoder 120, answer decoder 150, and question decoder 150.
  • the memory 110 may include a volatile and/or nonvolatile memory.
  • the one or more memories 110 may store instructions that, when executed by the one or more processors 190, cause the one or more processors 190 to perform an operation.
  • the program or the instruction which is software stored in the memory 110, may include an operating system for controlling resources of a server, an application, and/or middleware that provides various functions to the application so that the application may utilize the resources of the apparatus.
  • the one or more processors 190 may drive software (for example, programs, instructions) to control at least one component of the apparatus 100 connected to the processors 190.
  • the processor 190 may perform operations such as various operations, processing, data generation, and processing related to the present disclosure.
  • the processor 190 may load data or the like from the memory 110 or store the data or the like in the memory 110.
  • At least one of the components of the apparatus 100 may be omitted or another component may be added. Additionally or alternatively, some of the components may be implemented in an integrated form or be implemented as singular or plural entities.
  • the steps of the method or the algorithm described in connection with an exemplary embodiment of the present invention may be implemented directly by hardware, by a software module executed by hardware, or by a combination of hardware and software.
  • the software module may reside on a random access memory (RAM), a read only memory (ROM), an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM), a flash memory, a hard disk, a removable disk, a compact disk (CD)-ROM, or may reside on any type of computer extractable recording medium well-known in the art.
  • a dataset of a question and an answer is automatically generated from an input passage, and several question-answer pairs may thus be sampled with only one passage input, thereby constructing more diverse data as compared with an existing model.
  • a dataset of various questions and answers may be constructed in a dataset provided in an information box provided in a standardized form or a dataset provided in a table form on an existing website.
  • the method of generating a question-answer dataset extracts a question-answer dataset from data for which learning or machine translation is completed, thus it may be utilized to verify a quality of the learning or the translation by evaluating accuracy of an answer corresponding to a question.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Human Computer Interaction (AREA)
  • Medical Informatics (AREA)
  • Machine Translation (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Electrically Operated Instructional Devices (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)

Abstract

A method of generating a question-answer dataset based on a passage input may include: generating a hidden representation for each word in the input passage and storing the hidden representation in a memory; sampling a latent variable based on the hidden representation; calculating importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector; generating an answer span based on the latent variable and the weighted sum vector; and generating a question based on the answer span and an attention for the hidden representation in the memory.

Description

    BACKGROUND OF THE INVENTION Field of the Invention
  • The present invention discloses a method and an apparatus of generating various datasets by simultaneously generating and sampling a question-answer pair that may be generated in one passage.
  • Description of the Related Art
  • Due to the recent development of technology, various question-answer systems using deep learning technology have been developed. Recently, a system that gives an answer appropriate for a question using a question-answer (QA) model when a user gives a question in a natural language to the system has been demonstrated. However, since such a deep learning-based QA model requires a large amount of learning data, it could not but automatically give only answers for some specific fields or topics. Furthermore, there was a problem that it takes a significant amount of time and cost to manually construct such learning data.
  • Citation List Patent Literature
  • Patent Literature 1: Korean Patent No. 10-1605430 (published on March 22, 2016 )
  • SUMMARY OF THE INVENTION
  • An object of the present invention is to provide a method and an apparatus of generating a question-answer generating model that simultaneously generates a question-answer dataset capable of being generated based on an input passage even though an answer candidate group is not provided unlike an existing question-answer (QA) model in which a user manually collects and learns a joint distribution of question-answer pairs, and a question-answer learning model capable of being automatically learned using the question-answer generating model.
  • According to an aspect of the present invention, a method of generating a question-answer dataset based on a passage input may include: generating a hidden representation for each word in the input passage and storing the hidden representation in a memory; sampling a latent variable based on the hidden representation; calculating importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector; generating an answer span based on the latent variable and the weighted sum vector; and generating a question based on the answer span and an attention for the hidden representation in the memory.
  • The method may further include: storing a dataset of the generated question and an answer in the memory; and machine-learning a question-answer learning model using the stored data set of the question and the answer.
  • The method may further include: generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting a loss of the question-answer learning model and constraints; and performing adversarial machine learning using a dataset of a question and an answer generated using the generated latent variable.
  • The input passage may be generated by structuring infobox data extracted from data crawled from a website.
  • In a case where the infobox data is not extracted from the crawled data, a paragraph standardized based on a data frequency in a table recognized through column recognition may be input as the passage.
  • The input passage may be data for which machine translation is completed, and the method may further include evaluating a quality of the machine translation by evaluating accuracy of an answer based on a dataset of the generated question and the answer.
  • According to another aspect of the present invention, an apparatus of generating a question-answer dataset may include: one or more processors; and one or more memories in which instructions, when executed by the one or more processors, causing the one or more processors to perform an operation are stored, wherein the operation performed by the one or more processors includes: an operation of generating a hidden representation for each word in an input passage and storing the hidden representation in a memory; an operation of sampling a latent variable based on the hidden representation; an operation calculating importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector; an operation of generating an answer span based on the latent variable and the weighted sum vector; and an operation of generating a question based on the answer span and an attention for the hidden representation in the memory.
  • The operation performed by the one or more processors may further include: an operation of storing a dataset of the generated question and an answer in the memory; and an operation of machine-learning a question-answer learning model using the stored data set of the question and the answer.
  • The operation performed by the one or more processors may further include: an operation of generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting a loss of the question-answer learning model and constraints; and an operation of performing adversarial machine learning using a dataset of a question and an answer generated using the generated latent variable.
  • According to another exemplary embodiment, a program stored in a medium in order to execute generation of a question-answer dataset based on a passage input in combination with a computer which is hardware is provided.
  • Detailed contents of other exemplary embodiments of the present invention are described in a detailed description and are illustrated in the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
    • FIG, 1 is a conceptual diagram for schematically describing a question-answer dataset generating model based on a passage input according to an exemplary embodiment.
    • FIG. 2 is a flowchart for describing a method of generating a question-answer dataset based on a passage input according to an exemplary embodiment.
    • FIG. 3 is a view for describing an example of a method of generating a question-answer dataset in a paragraph according to an exemplary embodiment.
    • FIG. 4 is a view for describing a method of generating a question-answer dataset in a standardized infobox according to an exemplary embodiment.
    • FIG. 5 is a flowchart for describing a method of generating a question-answer dataset in a non-standardized page according to an exemplary embodiment.
    • FIG. 6 is a view for describing a method of improving performance of a question-answer model through adversarial learning according to an exemplary embodiment.
    • FIG. 7 is a block diagram schematically illustrating an internal configuration of an apparatus of generating a question-answer dataset according to an exemplary embodiment.
    DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • Various advantages and features of the present invention and methods accomplishing them will become apparent from the following description of exemplary embodiments with reference to the accompanying drawings. However, the present invention is not limited to exemplary embodiments to be described below, but may be implemented in various different forms, these exemplary embodiments will be provided only in order to make the present invention complete and allow those skilled in the art to completely recognize the scope of the present invention, and the present invention will be defined by the scope of the claims.
  • Terms used in the present disclosure are for describing exemplary embodiments rather than limiting the present invention. In the present disclosure, a singular form includes a plural form unless explicitly stated otherwise. Terms "comprise" and/or "comprising" used in the present disclosure do not exclude the existence or addition of one or more other components other than the mentioned components. Throughout the present disclosure, the same components will be denoted by the same reference numerals, and a term "and/or" includes each and all combinations of one or more of the mentioned components. Terms "first", "second" and the like are used to describe various components, but these components are not limited by these terms. These terms are used only in order to distinguish one component from another component. Therefore, a first component mentioned below may be a second component within the technical spirit of the present invention.
  • Unless defined otherwise, all terms (including technical and scientific terms) used in the present disclosure have the same meaning as meanings commonly understood by those skilled in the art to which the present invention pertains. In addition, terms defined in generally used dictionaries are not ideally or excessively interpreted unless they are specifically defined clearly.
  • Spatially relative terms "below", "beneath", "lower", "above", "upper", and the like, may be used in order to easily describe correlations between one component and other components. The spatially relative terms are to be understood as terms including different directions of the components at the time of use or at the time of operation in addition to directions illustrated in the drawings. For example, in a case of overturning component illustrated in the drawings, a component described as "below" or "beneath" another component may be placed "above" the other component. Therefore, an exemplary term "below" may encompass both of directions of above and below. Components may be oriented in other directions as well, and thus, spatially relative terms may be interpreted according to orientations.
  • Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the accompanying drawings.
  • FIG. 1 is a conceptual diagram for schematically escribing a question-answer dataset generating model based on a passage input according to an exemplary embodiment.
  • Referring to FIG. 1, a question-answer dataset generating model according to an exemplary embodiment may include a memory 110, an encoder 120, an attention 130, an answer decoder 140, and a question decoder 150.
  • According to the method of generating a question-answer dataset according to an exemplary embodiment, a passage 101 including several sentences is input to the encoder 120 as it is. Here, the passage 101 may include a phrase, a part, a paragraph, and the like.
  • The encoder 120 generates a hidden representation for each word in the input paragraph 101 and stores the hidden representation in the memory 110. Here, the encoder 120 may construct the hidden representation from an input through a neural network. Therefore, the hidden representation may mean a machine-readable data representation learned in a hidden layer of the neural network, and may be composed of a pair of keys and values. For example, the encoder 120 may store information about a meaning represented by a component (for example, a word) of a sentence as the hidden representation.
  • Next, the encoder 120 performs sampling 102 based on the hidden representation to generate a latent variable 103, and the generated latent variable 103 is input to the answer decoder 140.
  • The answer decoder 140 calculates the importance of each component of the memory 110 through the attention 130 that uses the latent variable 103 as a query to generate a weighted sum vector of the memory 110. For example, each time the answer decoder 140 predicts an output word through the attention 130, the answer decoder 140 may refer to an entire input sentence in the encoder 120 once again. However, the answer decoder 140 does not refer to the entire input sentence in the same ratio, but may attentively refer to an input word portion associated with a word to be predicted at a corresponding time. Therefore, an attention function obtains similarities to 'Keys' obtained from all the hidden representations with respect to given 'Queries'. In addition, the obtained similarities are reflected in the respective 'values' mapped to the keys. In addition, a weighted sum vector may be obtained by adding all the values in which the similarities are reflected.
  • The answer decoder 140 predicts an answer span 104 based on the weighted sum vector. Here, the answer span may mean a start point and an end point where an answer is positioned in the input passage.
  • Next, the question decoder 150 may receive the previously generated answer span 104, and perform calculation of the attention 130 of the memory 110 to finally generate the question 105.
  • A question-answer dataset composed of a pair of the generated question and an answer according to an exemplary embodiment may be stored in the memory 110, and the stored question-answer dataset may be used to machine-learn a question-answer learning model.
  • Therefore, the dataset generating model according to an exemplary embodiment automatically generates a dataset of a question and an answer from the input passage, and may thus sample several question-answer pairs with only one passage input, thereby constructing more diverse data as compared with an existing manual input model.
  • Hereinafter, a method of generating a question-answer dataset based on a passage input according to an exemplary embodiment will be described in more detail.
  • FIG. 2 is a flowchart for describing a method of generating a question-answer dataset based on a passage input according to an exemplary embodiment.
  • In step S200, the method for generating a question-answer dataset according to an exemplary embodiment generates a hidden representation for each word in an input passage and stores the hidden representation in the memory. Here, the input passage may be generated by structuring infobox data extracted from data crawled from a website. Alternatively, in a case where the infobox data is not extracted from the crawled data, a paragraph standardized based on a data frequency in a table recognized through column recognition may be input as the passage. Furthermore, if the input passage is data for which machine translation is completed, a quality of the machine translation may be evaluated by evaluating accuracy of the answer based on the generated question-answer dataset. A specific input passage will be described later with reference to FIGS. 3 to 4.
  • In step S210, the method of generating a question-answer dataset according to an exemplary embodiment samples a latent variable based on the hidden representation.
  • In step S220, the method of generating a question-answer dataset according to an exemplary embodiment generates a weighted sum vector by calculating the importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query.
  • In step S230, the method of generating a question-answer dataset according to an exemplary embodiment generates an answer span based on the latent variable and the weighted sum vector.
  • In step S240, the method of generating a question-answer dataset according to an exemplary embodiment generates a question based on the answer span and the attention for the hidden representation in the memory. Here, a dataset of the generated question and an answer may be stored in the memory. Furthermore, a question-answer model may be machine-learned using the stored dataset of the question and the answer. In addition, an adversarial question for adversarial machine learning is also automatically generated from the input passage, such that that the question-answer model may be machine-learned to be closer to a real boundary.
  • More specifically, the method for generating a question-answer dataset described above may be implemented through an objective function shown in the following Equation 1: L θ , ϕ , ψ x y c = E x , y , c p x y c E z q ϕ z | x , y , c logp θ y | z , c logp θ x | z , y , c + D KL q ϕ z p ψ z
    Figure imgb0001
    Here x, y, c, and z correspond to a question, an answer, a paragraph, and a latent variable, respectively. qφ ( z|x,y,c ) is a function used for the encoder of the question-answer model, and pθ ( y|z,c ),pθ ( x|z,y,c ) are functions used for the answer decoder and the question decoder, respectively. pψ (z) is a marginal prior distribution of the latent variables, and qφ (z) is a marginal posterior distribution of the latent variables. For example, qθ (z) may generate a distribution of latent variables generated from the input, and pψ ( z ) may generate an output close to the input using latent variables. Meanwhile, since it is very difficult to calculate a Kullback-Leibler (KL) divergence for optimization (for reducing a difference between the two distributions described above), it may be approximated as represented by the following Equation 2: q ϕ z 1 N n = 1 N q ϕ z | x n , y n , c n p ψ z 1 N n = 1 N p ψ z | c n
    Figure imgb0002
  • Meanwhile, how the passage input is performed and how a question-answer dataset construction model according to an exemplary embodiment may be utilized will be described with reference to FIGS. 3 and 4.
  • FIG. 3 is a view for describing an example of a method of generating a question-answer dataset in a paragraph according to an exemplary embodiment.
  • Referring to FIG. 3, when a paragraph 301 is input, an answer 303 is first generated through sampling 302. In addition, an answer 304 paired with the answer may be generated using the generated answer 303.
  • For example, in a case where there is a sentence "Yi Sun-shin (April 28, 1545 to December 16, 1598 (November 19 by the lunar calendar)) is a military officer of the mid-Joseon Dynasty" in an input passage, an answer "April 28, 1545" may be first generated through sampling. In addition, various types of questions such as "When was Admiral Yi Sun-shin's birthday?" Or "When was Yi Sun Shin born?" may be generated based on the answer generated according to the method for generating a question-answer dataset according to an exemplary embodiment. Therefore, in a case of using the question-answer generating model, various datasets for machine-learning the question-answer model may be constructed.
  • FIG. 4 is a view for describing a method of generating a question-answer dataset in a standardized infobox according to an exemplary embodiment.
  • Referring to FIG. 4, the question-answer dataset generating model according to an exemplary embodiment may obtain infobox data 400 through crawling from a website famous as an online knowledge encyclopedia website.
  • Particularly, since the infobox data 400 is standardized, the infobox data 400 may be easily structured in a form of a corpus of questions and answers, and since these knowledge encyclopedia websites provide detailed classification data for fields or targets, an input passage may be generated in a preprocessing step so as to generate and machine-learn a question-answer by targeting a desired field.
  • However, since all websites do not provide the standardized infobox data 400 as illustrated in FIG. 4, crawled data is generally non-standardized. Therefore, in this case, how to create a passage to be input will be described.
  • FIG. 5 is a flowchart for describing a method of generating a question-answer dataset in a non-standardized page according to an exemplary embodiment.
  • Referring to FIG. 5, in step S500, the question-answer dataset construction model recognizes the title and the body of a web page from crawled data. In addition, in step S510, a table is recognized based on column recognition in the body. That is, the table is recognized in order to obtain data having a form similar to an infobox that may be obtained from the standardized site.
  • Then, in step S520, a paragraph is standardized by confirming whether or not there is a word satisfying a preset data frequency. For example, words that appear repeatedly, such as "country name" and "capital" may be confirmed, and the paragraph may be standardized based on these words.
  • In step S530, the question-answer dataset generating model may receive the standardized paragraph and generate a question-answer dataset to construct a dataset of various questions and answers in a dataset provided in an information box provided in a standardized form or a dataset provided in a table form on an existing website.
  • The question-answer dataset generating model described above according to another exemplary embodiment may extract a question-answer dataset from data for which learning and machine translation is completed, evaluate accuracy of an answer corresponding to the question, and utilize an evaluation result to verify a quality of learning or translation, beyond simply generating the question-answer dataset.
  • For example, if the question-answer dataset generating model according to an exemplary embodiment can generate a correct question-answer dataset from the data for which the machine translation is completed, it may be evaluated that machine reading is well performed.
  • Meanwhile, in a case of a recent question-answer learning model, performance is already significantly high because a preliminary training is performed using a language model objective or the like in a large amount of corpus. Therefore, even though more data are simply used for learning, performance of the question-answer model are not dramatically improved. Particularly, since most of question-answer pairs automatically generated through the question-answer dataset generating model according to an exemplary embodiment may already be guessed by the question-answer learning model, they may not be very helpful. Therefore, a method for improving performance of machine-learning is required.
  • Therefore, a question-answer dataset generating model according to another exemplary embodiment discloses a method of generating a question-answer dataset that cannot be guessed by the current question-answer learning model through adversarial learning.
  • FIG. 6 is a view for describing a method of improving performance of a question-answer model through adversarial learning according to an exemplary embodiment.
  • Referring to FIG. 6, in order to perform adversarial learning, data present in a data manifold while being outside a decision boundary of the current question-answer model need to be sampled. Specifically, the method of generating a question-answer dataset based on a passage input according to an exemplary embodiment selects the latent variable through the encoder, and a question-answer data pair generated from the sampled latent variable needs to be a question-answer data pair whose correct answer cannot be guessed through a question-answer learning model. In this case, the question-answer learning model calibrates a decision boundary by learning the question-answer data pair whose correct answer could not be guessed, and when such a process is repeated, the decision boundary of the question-answer learning model may become close to a true decision boundary, and performs of the question-answer learning model may also be improved.
  • Meanwhile, referring to FIG. 6, a question-answer dataset for the adversarial learning may be determined by adjusting a loss of the question-answer learning model and constraints for positioning the latent variable in the data manifold.
  • For example, the question-answer learning model in the adversarial manner described above may have an objective function of the following Equation 3: min ω max π E x , y , c p x y c L ω x y c + E c u p u c L ω x u y u c u D KL p π z | c u p ψ z | c u s . t . z p π z | c u , x u , y u p θ x , y | z , c u
    Figure imgb0003
    Here x, y, c, and z correspond to a question, an answer, a paragraph, and a latent variable, respectively. qπ (z|c) is a function used for a pre-processor of the question-answer model, pπ (z|c) is a function used for an agent, and pθ ( y|z,c ),pθ ( x|z,y,c ) are functions used for the answer decoder and the question decoder, respectively.
  • Meanwhile, the encoder samples the latent variable (z), and a question and a correct answer generated from the sampling need to increase a loss (
    Figure imgb0004
    ) of the question-answer model. However, since there may be no relationship between the question and the correct answer generated as such, constraints are required in selecting z. That is, as z becomes distant from the previously learned distribution of z, an invalid question-answer pair is generated. Therefore, in order to prevent this problem, a value may be adjusted in such a manner that a penalty is given when a distance between the two distributions becomes distant (for example, use a KL divergence).
  • Therefore, the question-answer learning model according to an exemplary embodiment may generate a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting the loss and the constraints.
  • In addition, performance of the question-answer earning model may be improved by performing adversarial machine learning using a dataset of an adversarial question and answer generated using the generated latent variable.
  • FIG. 7 is a block diagram schematically illustrating an internal configuration of an apparatus of generating a question-answer dataset according to an exemplary embodiment.
  • Referring to FIG. 7, an apparatus 100 of generating a question-answer dataset according to an exemplary embodiment may include one or more memories 110 and a processor 190. Operations of the apparatus 100 of generating a question-answer dataset may be performed by executing a program stored in the memory 110 through the processor 190.
  • The processor 190 according to an exemplary embodiment performs an operation of generating a hidden representation for each word in the input passage and storing the hidden representation in the memory 110.
  • In addition, the processor 190 samples a latent variable based on the hidden representation, and calculates the importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector.
  • In addition, the processor 190 may perform an operation of generating an answer span based on the latent variable and the weighted sum vector and generating a question based on the generated answer span and the attention for the hidden representation in the memory.
  • In addition, the processor 190 may further perform an operation of storing a dataset of the generated question and an answer in the memory 110 and a machine learning operation of machine-learning a question-answer learning model using the stored dataset of the question and the answer.
  • In addition, the processor 19 may further perform an operation of generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting the loss of the question-answer learning model and the constraints and an operation of performing adversarial machine learning using the question-answer dataset generated using the generated latent variable.
  • Meanwhile, the apparatus 100 of generating a question-answer dataset described above may include one or more processors 190 and/or one or more memories 110. In addition, each of the processors 190 may include an encoder 120, an answer decoder 140, and a question decoder 150, and may perform the operations described above with reference to FIGS 1 to 7 on each of the encoder 120, answer decoder 150, and question decoder 150.
  • In addition, the memory 110 may include a volatile and/or nonvolatile memory. The one or more memories 110 may store instructions that, when executed by the one or more processors 190, cause the one or more processors 190 to perform an operation. In the present disclosure, the program or the instruction, which is software stored in the memory 110, may include an operating system for controlling resources of a server, an application, and/or middleware that provides various functions to the application so that the application may utilize the resources of the apparatus.
  • The one or more processors 190 may drive software (for example, programs, instructions) to control at least one component of the apparatus 100 connected to the processors 190. In addition, the processor 190 may perform operations such as various operations, processing, data generation, and processing related to the present disclosure. In addition, the processor 190 may load data or the like from the memory 110 or store the data or the like in the memory 110.
  • In an exemplary embodiment, at least one of the components of the apparatus 100 may be omitted or another component may be added. Additionally or alternatively, some of the components may be implemented in an integrated form or be implemented as singular or plural entities.
  • Meanwhile, the steps of the method or the algorithm described in connection with an exemplary embodiment of the present invention may be implemented directly by hardware, by a software module executed by hardware, or by a combination of hardware and software. The software module may reside on a random access memory (RAM), a read only memory (ROM), an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM), a flash memory, a hard disk, a removable disk, a compact disk (CD)-ROM, or may reside on any type of computer extractable recording medium well-known in the art.
  • According to an exemplary embodiment, a dataset of a question and an answer is automatically generated from an input passage, and several question-answer pairs may thus be sampled with only one passage input, thereby constructing more diverse data as compared with an existing model.
  • In addition, a dataset of various questions and answers may be constructed in a dataset provided in an information box provided in a standardized form or a dataset provided in a table form on an existing website.
  • Furthermore, since the method of generating a question-answer dataset according to an exemplary embodiment extracts a question-answer dataset from data for which learning or machine translation is completed, thus it may be utilized to verify a quality of the learning or the translation by evaluating accuracy of an answer corresponding to a question.
  • Effects of the present invention are not limited to the effects described above, and other effects that are not mentioned may be obviously understood by those skilled in the art from the following description.
  • Although the exemplary embodiments of the present invention have been described with reference to the accompanying drawings, those skilled in the art will understand that various modifications and alterations may be made without departing from the spirit or essential feature of the present invention. Therefore, it is to be understood that the exemplary embodiments described hereinabove are illustrative rather than being restrictive in all aspects.

Claims (10)

  1. A method of generating a question-answer dataset based on a passage input, comprising:
    generating a hidden representation for each word in the input passage and storing the hidden representation in a memory;
    sampling a latent variable based on the hidden representation;
    calculating importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector;
    generating an answer span based on the latent variable and the weighted sum vector; and
    generating a question based on the answer span and an attention for the hidden representation in the memory.
  2. The method of claim 1, further comprising:
    storing a dataset of the generated question and an answer in the memory; and
    machine-learning a question-answer learning model using the stored data set of the question and the answer.
  3. The method of claim 2, wherein:
    generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting a loss of the question-answer learning model and constraints; and
    performing adversarial machine learning using a dataset of a question and an answer generated using the generated latent variable.
  4. The method of claim 1, wherein:
    the input passage is generated by structuring infobox data extracted from data crawled from a website.
  5. The method of claim 4, wherein
    in a case where the infobox data is not extracted from the crawled data, a paragraph standardized based on a data frequency in a table recognized through column recognition is input as the passage.
  6. The method of claim 1, wherein:
    the input passage is data for which machine translation is completed,
    the method further comprising evaluating a quality of the machine translation by evaluating accuracy of an answer based on a dataset of the generated question and the answer.
  7. An apparatus of generating a question-answer dataset, comprising:
    one or more processors; and
    one or more memories in which instructions, when executed by the one or more processors, causing the one or more processors to perform an operation are stored,
    wherein the operation performed by the one or more processors includes:
    an operation of generating a hidden representation for each word in an input passage and storing the hidden representation in a memory;
    an operation of sampling a latent variable based on the hidden representation;
    an operation calculating importance of the hidden representation stored in the memory through an attention that uses the sampled latent variable as a query to generate a weighted sum vector;
    an operation of generating an answer span based on the latent variable and the weighted sum vector; and
    an operation of generating a question based on the answer span and an attention for the hidden representation in the memory.
  8. The apparatus of generating a question-answer dataset of claim 7, wherein the operation performed by the one or more processors further includes:
    an operation of storing a dataset of the generated question and an answer in the memory; and
    an operation of machine-learning a question-answer learning model using the stored data set of the question and the answer.
  9. The apparatus of generating a question-answer dataset of claim 8, wherein the operation performed by the one or more processors further includes:
    an operation of generating a latent variable which is valid and in which the question-answer learning model cannot guess a correct answer by adjusting a loss of the question-answer learning model and constraints; and
    an operation of performing adversarial machine learning using a dataset of a question and an answer generated using the generated latent variable.
  10. A program for generating a question-answer dataset based on a passage input, stored in a medium in order to execute the method of any one of claims 1 to 6 in combination with a computer which is hardware.
EP19212219.0A 2019-11-28 2019-11-28 Method and apparatus for generating question and answer dataset based on input paragraph Ceased EP3828781A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP19212219.0A EP3828781A1 (en) 2019-11-28 2019-11-28 Method and apparatus for generating question and answer dataset based on input paragraph
KR1020200119570A KR102456994B1 (en) 2019-11-28 2020-09-17 Method and apparatus for generating question and answer dataset based on input paragraph
KR1020220133034A KR102655809B1 (en) 2019-11-28 2022-10-17 Method and apparatus for generating question and answer dataset based on input paragraph
KR1020240045399A KR20240049528A (en) 2019-11-28 2024-04-03 Method and apparatus for generating question and answer dataset based on input paragraph

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
EP19212219.0A EP3828781A1 (en) 2019-11-28 2019-11-28 Method and apparatus for generating question and answer dataset based on input paragraph

Publications (1)

Publication Number Publication Date
EP3828781A1 true EP3828781A1 (en) 2021-06-02

Family

ID=68731904

Family Applications (1)

Application Number Title Priority Date Filing Date
EP19212219.0A Ceased EP3828781A1 (en) 2019-11-28 2019-11-28 Method and apparatus for generating question and answer dataset based on input paragraph

Country Status (2)

Country Link
EP (1) EP3828781A1 (en)
KR (3) KR102456994B1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3828778A1 (en) * 2019-11-29 2021-06-02 42 Maru Inc. Method and apparatus for generating q & a model by using adversarial learning
EP3846088A1 (en) * 2019-12-30 2021-07-07 42 Maru Inc. Method and apparatus of generating question-answer learning model through reinforcement learning
CN119621906A (en) * 2024-11-28 2025-03-14 国家电网有限公司大数据中心 A large-model explainable question answering method and system based on knowledge graph fusion

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20230056930A (en) 2021-10-21 2023-04-28 삼성에스디에스 주식회사 Apparatus and method for generating multihop question answering
KR102790072B1 (en) * 2021-11-29 2025-04-01 고려대학교 산학협력단 Apparatus of generating query and method thereof
KR102771427B1 (en) * 2021-12-29 2025-02-24 (주)노스스타컨설팅 System and Method of Machine Reading Comprehension for Animal disease information search based on paragraph exploration and paragraph selection

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101605430B1 (en) 2014-02-14 2016-03-22 주식회사 플런티코리아 System and method for constructing a questionnaire database, search system and method using the same

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102033435B1 (en) * 2014-11-19 2019-11-11 한국전자통신연구원 System and Method for Question and answer of Natural Language and Paraphrase Module
KR101686068B1 (en) * 2015-02-24 2016-12-14 한국과학기술원 Method and system for answer extraction using conceptual graph matching
KR101982345B1 (en) * 2017-07-21 2019-05-24 주식회사 마인드셋 Apparatus and method for composing music using artificial intelligence
KR102608469B1 (en) * 2017-12-22 2023-12-01 삼성전자주식회사 Method and apparatus for generating natural language

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101605430B1 (en) 2014-02-14 2016-03-22 주식회사 플런티코리아 System and method for constructing a questionnaire database, search system and method using the same

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
VASWANI ASHISH ET AL: "Attention Is All You Need", 6 December 2017 (2017-12-06), XP055812037, Retrieved from the Internet <URL:https://arxiv.org/pdf/1706.03762.pdf> [retrieved on 20210609] *
VISHWAJEET KUMAR ET AL: "A Framework for Automatic Question Generation from Text using Deep Reinforcement Learning", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 15 August 2018 (2018-08-15), XP081454729 *
VISHWAJEET KUMAR ET AL: "Automating Reading Comprehension by Generating Question and Answer Pairs", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 7 March 2018 (2018-03-07), XP080863241 *
VISHWAJEET KUMAR ET AL: "Putting the Horse Before the Cart:A Generator-Evaluator Framework for Question Generation from Text", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 15 August 2018 (2018-08-15), XP081491542 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3828778A1 (en) * 2019-11-29 2021-06-02 42 Maru Inc. Method and apparatus for generating q & a model by using adversarial learning
EP3846088A1 (en) * 2019-12-30 2021-07-07 42 Maru Inc. Method and apparatus of generating question-answer learning model through reinforcement learning
CN119621906A (en) * 2024-11-28 2025-03-14 国家电网有限公司大数据中心 A large-model explainable question answering method and system based on knowledge graph fusion

Also Published As

Publication number Publication date
KR20210066704A (en) 2021-06-07
KR102655809B1 (en) 2024-04-08
KR102456994B1 (en) 2022-10-21
KR20220143804A (en) 2022-10-25
KR20240049528A (en) 2024-04-16

Similar Documents

Publication Publication Date Title
EP3828781A1 (en) Method and apparatus for generating question and answer dataset based on input paragraph
US12373671B2 (en) Method and apparatus for generating Q and A model by using adversarial learning
US20210201192A1 (en) Method and apparatus of generating question-answer learning model through reinforcement learning
US11656851B2 (en) Long-range modeling of source code files by syntax hierarchy
CN119493849B (en) Question-answering task processing model training method, device, equipment and storage medium
CN104933158A (en) Training method and device and reasoning method and device for mathematical problem solving model
CN119377372B (en) Answer acquisition method, computer program product, device and storage medium
CN119129604A (en) Medical teaching plan generation method, device, equipment, storage medium and program product
CN109522397A (en) Information processing method and device based on semanteme parsing
CN118095272A (en) Text recognition method and device, electronic equipment and storage medium
EP3846088A1 (en) Method and apparatus of generating question-answer learning model through reinforcement learning
CN120450035A (en) A method and system for generating mathematical problems based on adaptive multi-level retrieval enhancement
CN112699675B (en) Text processing method, device, equipment and computer readable storage medium
He et al. [Retracted] Application of Grammar Error Detection Method for English Composition Based on Machine Learning
Avramidis Sentence-level ranking with quality estimation
KR102355591B1 (en) Method and apparatus of generating question-answer learning model through reinforcement learning
CN113010655A (en) Answer and interference item generation method and device for reading and understanding of machine
CN117171332A (en) Intelligent question-answering method and system based on AI
CN111949769B (en) Method and device for enhancing robustness of reading and understanding system
Kaur et al. Punjabi grammar checker: Automatic correction of punjabi language sentences using transformer
Junaidi et al. DEVELOPMENT AND OPTIMIZATION OF SIMPLE-O: AN AUTOMATED ESSAY SCORING SYSTEM FOR JAPANESE LANGUAGE BASED ON BERT, BILSTM, AND BIGRU
Taviss Asm2Seq: Explainable Assembly Code Functional Summary Generation
Liang et al. GeoGRPO: Investigating the Stepwise-GRPO Enhancement in RLHF Framework
Zhou et al. Dynamic knowledge correction via abductive for domain question answering
CN119150882A (en) Text response method, text response device, computer equipment and storage medium

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN PUBLISHED

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20211202

RBV Designated contracting states (corrected)

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20231030

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN REFUSED

18R Application refused

Effective date: 20241118