US12277870B2 - Interface to natural language generator for generation of knowledge assessment items - Google Patents
Interface to natural language generator for generation of knowledge assessment items Download PDFInfo
- Publication number
- US12277870B2 US12277870B2 US17/809,217 US202217809217A US12277870B2 US 12277870 B2 US12277870 B2 US 12277870B2 US 202217809217 A US202217809217 A US 202217809217A US 12277870 B2 US12277870 B2 US 12277870B2
- Authority
- US
- United States
- Prior art keywords
- assessment items
- items
- model
- knowledge assessment
- nlg
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B7/00—Electrically-operated teaching apparatus or devices working with questions and answers
- G09B7/06—Electrically-operated teaching apparatus or devices working with questions and answers of the multiple-choice answer-type, i.e. where a given question is provided with a series of answers and a choice has to be made from the answers
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/55—Rule-based translation
- G06F40/56—Natural language generation
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/02—Electrically-operated educational appliances with visual presentation of the material to be studied, e.g. using film strip
Definitions
- Natural Language Generators may be used to generate a wide variety of content.
- NLGs Natural Language Generators
- assessment items e.g., test questions
- generating test questions and other items configured to measure knowledge in specific subject areas or content domains may be difficult.
- An item generation interface may generate knowledge assessment items directed a subject area based on a set of model items collectively directed to the subject area.
- the item generation interface may group the set of model assessment items into a plurality of similar item groups using numeric features corresponding to the model assessment items. Similar item groups may include model assessment items covering conceptually similar concepts within the subject area.
- a conditioning input may be generated for each of the item groups based on the numeric features corresponding to the model assessment items in the item group. Responsive to providing the conditioning inputs to a transformer-based natural language generation model, the item generation interface may receive raw assessment items from the transformer-based natural language generation model. Knowledge assessment items may be identified and/or developed from the raw assessment items.
- FIG. 1 illustrates an example system including a NLG interface providing for generation of knowledge assessment items using a NLG accessible over a network in accordance with some embodiments of the disclosure.
- FIG. 3 illustrates a diagram of example components of a NLG interface in accordance with some embodiments of the disclosure.
- FIG. 5 is a schematic diagram of an example computer system which may be used to implement various embodiments in the examples described herein.
- FIG. 7 illustrates an example process for generating conditioning input for a NLG, in accordance with various embodiments of the disclosure.
- FIG. 8 illustrates an example process for performing items verification of items generated by a NLG, in accordance with various embodiments of the disclosure.
- Producers of standardized tests, licensing exams, and other assessment tools may face difficulties in generating the large volume of assessment items needed to ensure test security, particularly when tests may be given to many individuals at different locations and/or at different times (e.g., asynchronous distribution). For example, for online tests taken asynchronously, reuse of testing items may incentivize earlier test takers to provide testing items to later test takers, such that the later tests may not accurately assess the knowledge of individual test takers. Further, other entities, such as those providing preparation for such exams, may face difficulties in generating practice testing items which mirror actual testing items both in subject matter coverage and structure.
- testing items that closely mirror, but are not identical to, already released testing items provides test takers with more robust preparation, as the test takers do not encounter the same questions multiple times in practice situations.
- generating large volumes of testing items to fulfill such requirements may utilize subject matter experts specially trained in authoring testing items, which may naturally limit the number of testing items available. Because of the increasing need to rely on these testing items to do important knowledge and skill measurements in more areas, meeting the need for good test items is increasingly a challenge. Ineffective test items may lead to ineffective measurements.
- Machine Learning and Artificial Intelligence models may provide useful tools for generating high volumes of knowledge assessment items (e.g., standardized test questions) for practice and learning applications, delivery of varied tests to asynchronous test takers, and other applications.
- knowledge assessment items e.g., standardized test questions
- models pre-trained for specific tasks and/or subject matters may be time consuming to develop, and may require large volumes of data to provide useful output.
- subject areas which change over time the burden of generating new models to ensure new or different concepts are represented in generated items may outweigh benefits of using language generation models to generate knowledge assessment items. For example, professional licensure examinations may be updated on a monthly, quarterly, or yearly basis to reflect updated regulations, best practices, and other developments.
- Pre-trained models may be useful to reduce or remove barriers related to training models for specific tasks.
- Such pre-trained Natural Language Generation (NLG) models may be utilized to generate various linguistic structures. These NLG models may use unsupervised language modeling, meaning that the NLG models may perform various tasks and may generate language in a variety of content domains. For example, NLG models may mimic conditioning input in both form and content to generate literary passages on given topics, computer code, summarize text, answer questions, etc.
- NLG models may mimic conditioning input in both form and content to generate literary passages on given topics, computer code, summarize text, answer questions, etc.
- these models are trained on large corpuses of data, it may be difficult to generate meaningful content in very specific content domains. Accordingly, using pre-trained NLG models to generate large volumes of factually accurate, semantically correct, and appropriately difficult items useful in assessing knowledge in various focused subject areas is difficult.
- Generating meaningful, technically correct, and semantically correct output is important when generating knowledge assessment items, as technically incorrect question stems, confusing answer choices, or other issues may render the items ineffective to measure knowledge of test takers in the subject area.
- providing a large volume of model items to a natural language generation model in a randomized order may cause the natural language generation model to generate knowledge assessment items that are ineffective to measure knowledge of test takers across the discrete concepts of the subject area. This is because while natural language generation models using next token prediction may generally process individual inputs (e.g., a single model item) bi-directionally, the inputs may have reduced weight based on their order in the input.
- model items provided to the natural language generation model may be given substantially more weight than other inputs, such that the generated model item may more closely mirror one model item, regardless of whether the model item is representative of the set of model items as a whole.
- the natural language generation model predicts next tokens with lower probabilities, which may result in nonsensical or technically incorrect output. It may be similarly difficult to generate assessment items of multiple cognitive types (e.g., recall and application questions) and assessment items spanning several difficulty levels.
- assessments of multiple cognitive types e.g., recall and application questions
- assessment items spanning several difficulty levels.
- a large volume of assessment items covering all concepts to be tested with varying cognitive types and difficulty levels is often needed.
- a user may provide the natural language generation model with a conditioning input consisting of model items covering several concepts and/or concept areas, such as electronic circuits, magnetic fields, electrostatics, and electromagnetism.
- a conditioning input provides a model item directed to electronic circuits in a position given more weight by the NLG
- generated items are more likely to cover electronic circuits than other concepts where model items directed to the concept are provided elsewhere in the conditioning input.
- the model items may be conceptually far apart (e.g., include different wording and test different concepts)
- the natural language generation model is more likely to make predictions that are nonsensical or technically incorrect. Accordingly, generating knowledge assessment items in this manner using a natural language generation model may still involve subject matter experts for review and substantial modifications to transform items generated by the natural language generation model into items usable to assess knowledge of test takers in a subject area.
- a NLG interface as disclosed herein may assess a set of model knowledge assessment items directed to a subject area to generate conditioning inputs to a NLG that are more likely to result in generation of useful knowledge assessment items by the NLG.
- Useful knowledge assessment items may, individually, be technically accurate, coherent, include appropriate answer choices for multiple choice questions (e.g., keys and distractors), and be at an appropriate level for a given assessment (e.g., at a grade appropriate reading level).
- the NLG interface may provide a large volume of knowledge assessment items that are unique when compared to one another and to model items, span each concept to be tested in a given subject matter area, include items of multiple cognitive types (e.g., recall and application items), and include items spanning an acceptable difficulty range.
- the NLG interface may provide large volumes of content useful in, for example, generating unique assessments for individual test takers, providing large volumes of practice content, etc.
- the NLG interface may compensate for the architecture of the NLG by, for example, identifying semantically and conceptually similar items in the set of model knowledge assessment items and creating conditioning inputs for groups of similar items. In this manner, the model items provided in a conditioning input are conceptually closer together, or more similar, such that the NLG may more accurately predict next tokens in a generated item. Items may be grouped based on several factors including, in some examples, content coverage, difficulty, cognitive type, simultaneously. Accordingly, the NLG interface may produce sets of generated items spanning concepts, difficulty distributions, cognitive type, and other metrics as desired by users.
- the NLG interface may order inputs corresponding to model items forming a conditioning input such that model items more representative of the group of model items are more heavily weighted by the NLG in comparison to model items that may be conceptual outliers.
- the NLG interface may provide conditioning input to the NLG for each of the conceptual groups identified by the interface such that items generated by the NLG conceptually cover the breadth of the subject area covered by the original set of model items provided by the user to the interface. Accordingly, the NLG interface may generate large volumes of technically and lexically accurate assessment items spanning provided concepts, difficulty, and cognitive types, such that the generated items are useful in assessing the knowledge of target test-takers in a given content area.
- the NLG interface also provides significantly more items than could realistically be provided by human content writers with subject matter expertise, improving assessment security and allowing assessment creators to keep pace with demands for varied assessments covering defined subject areas with similar degrees of difficulty such that results between test takers assessed with different assessments are comparable even where the assessments are not item by item identical.
- the NLG interface may also provide an interface to a user for editing and refining items generated by the NLG responsive to the conditioning inputs created by the interface.
- the NLG interface may provide raw items generated by the NLG to a user device via a user interface at the user device.
- the user interface may allow the user to change aspects of a raw item and provide the edits to the interface to generate the remainder of the item based on the edits.
- the NLG interface may update the conditioning input originally provided to the NLG to generate the raw item and provide the updated conditioning input to the NLG to re-generate the remainder of the item. Accordingly, items generated by the NLG may be updated or refined via the interface instead of being wholly discarded, saving additional resources in generating large volumes of knowledge assessment items.
- the user device 106 and/or additional user devices may be implemented using any number of computing devices including, but not limited to, a computer, a laptop, tablet, mobile phone, smart phone, wearable device (e.g., AR/VR headset, smart watch, smart glasses, or the like), smart speaker, vehicle (e.g., automobile), or appliance.
- the user device 106 may include one or more processors, such as a central processing unit (CPU) and/or graphics processing unit (GPU).
- the user devices may generally perform operations by executing executable instructions (e.g., software) using the processor(s).
- the network 108 may be implemented using one or more of various systems and protocols for communications between computing devices.
- the network 108 or various portions of the network 108 may be implemented using the Internet, a local area network (LAN), a wide area network (WAN), and/or other networks.
- data may be communicated according to protocols and/or standards including near field communication (NFC), Bluetooth, cellular connections, and the like.
- NFC near field communication
- Various components of the system 100 may communicate using different network protocols or communications protocols based on location.
- the NLG interface 110 and the NLG 112 may be hosted within a cloud computing environment and may communicate with each other using communication and/or network protocols used by the cloud computing environment.
- Examples described herein may include storage devices, such as one or more databases, storing data such as user data 102 and/or interface data 104 .
- Such storage devices may be databases, servers, or other repositories of data accessible via the Internet or other network 108 .
- User data 102 and/or interface data 104 may be stored across more than one physical device and may, in some implementations, include storage components and devices belonging to multiple entities, users, or third parties.
- User data 102 may include, for example, sets of model knowledge assessment items covering various subject areas, generated knowledge assessment items, content maps corresponding to sets of model knowledge assessment items, writing guidelines corresponding to sets of model knowledge assessment items, metadata regarding model knowledge items in the sets of model knowledge assessment items, and additional user settings.
- Such user data 102 may be used by the NLG interface 110 to produce conditioning inputs meeting user specifications or to ensure that items generated by the NLG 112 meet user specifications.
- user data 102 may include user-provided metadata such as item topic categorization, key and distractor labels, item cognitive type (e.g., recall, application, etc.), and difficulty metrics (e.g., p-value and point biserial) for each item in a set of model assessment items.
- Such metadata may be used by the NLG interface 110 to provide a variety of assessment items.
- the NLG interface 110 may, in some embodiments, use difficulty metrics to ensure that generated items are of a similar difficulty level or span a desired range of difficulty levels.
- cognitive type metadata may be used by the NLG interface 110 to provide a user with generated items covering several cognitive types for use in assessments.
- a content map or test plan may be stored as user data 102 and may provide information on concepts represented by a set of model knowledge assessment items as well as how such concepts are interrelated.
- a content map formatted as a tree may provide a hierarchical contextual representation of concepts covered by a set of model knowledge assessment items.
- the content map may provide information about interrelated concepts. For example, concepts that have a common parent concept may be closely related, while concepts that share no ancestors in the content map may be conceptually unrelated.
- the NLG interface 110 may use such information when, for example, forming clusters of similar items within a set of model knowledge items to ensure the concepts intended to be covered by a set of model knowledge assessment items are included in the generated items.
- user data 102 may also include metadata for each model item in the set of model knowledge assessment items indicating where the model items fit within a conceptual map (e.g., which concepts are tested by the item).
- Interface data 104 may include instructions used by the NLG interface 110 , previously used conditioning inputs, user feedback correlated with previously used conditioning inputs, and the like. Interface data 104 may also include various parameters used by the NLG interface 110 , such as a number of clusters of similar items to use in the clustering step, specifications of the interface of the NLG 112 , etc. In some implementations, interface data 104 may include “default settings” for various parameters of the NLG interface 110 which may be updated by users in some situations. For example, a user may update the number of clusters of similar items to better fit a particular set of model assessment items. The user may also update, for example, which version or size of the NLG 112 is used, in some embodiments.
- Components of the system 100 shown in FIG. 1 are exemplary and may vary in some embodiments.
- the NLG interface 110 may be distributed across multiple computing elements, such that components of the NLG interface 110 communicate with one another through the network 108 .
- computing resources dedicated to the NLG interface 110 may vary over time based on various factors such as usage of the NLG interface 110 .
- FIG. 2 illustrates a diagram of an example NLG interface 110 in accordance with some embodiments of the disclosure.
- the NLG interface 110 includes several logical sub-components, including a text encoder 116 , clustering 118 , input generation 120 , and item verification 122 .
- the NLG interface 110 may generally receive model assessment items 114 (as shown in FIG. 3 ) and encode the model assessment items 114 using the text encoder 116 to generate numeric representations of the model assessment items 114 .
- Clustering 118 may cluster, classify, or otherwise group the numerical representations of the model items 114 .
- Input generation 120 may format input to the NLG 112 to provide the NLG 112 with instructions to generate assessment items for the groups or clusters generated by clustering 118 .
- Item verification 122 may perform checks of assessment items generated by the NLG 112 and items passing such checks may be provided to a user or communicated to a location (e.g., storage) as generated assessment items.
- the NLG interface 110 may access various data via network 108 , such as interface data 104 , to perform its various functions.
- the user device 106 may be a computing device associated with an end user or other user generating knowledge assessment items.
- the NLG interface 110 may communicate with the user device 106 over the network 108 to provide a user interface 126 to the NLG interface 110 .
- the user interface 126 may allow a user to provide requests to the NLG interface 110 and to view and interact with items generated by the NLG 112 based on conditioning input generated by the NLG interface 110 .
- the user interface 126 may be implemented as a React, Javascript-based interface for interaction with the NLG interface 110 .
- the user interface 126 may provide instructions to the NLG interface 110 to generate knowledge assessment items based on a set of model knowledge assessment items.
- the instructions may include, in various embodiments, location information for the set of model knowledge assessment items, content maps including concepts covered by the set of model knowledge assessment items, specifications for the generated items, metadata pertaining to the model knowledge assessment items, and other information used by the NLG interface 110 and/or the NLG 112 .
- the user interface 126 may also display generated items and allow for a user to edit generated items or provide additional information about the generated items. For example, the user interface 126 may display a question stem and answer choices for a generated item and allow the user to edit the question stem and regenerate answer choices based on the edited question stem, choose keys and distractors from the answer choices, rate or tag the generated item, and the like.
- the NLG interface 110 may include or utilize one or more hosts or combinations of compute resources, which may be located, for example, at one or more servers, cloud computing platforms, computing clusters, and the like. Generally, the NLG interface 110 is implemented by compute resources including hardware for memory 111 and a processor 109 . For example, the NLG interface 110 may utilize or include one or more processors, such as a CPU, GPU, and/or programmable or configurable logic. In some embodiments, various components of the NLG interface 110 may be distributed across various computing resources, such that the components of the NLG interface 110 communicate with one another through the network 108 or using other communications protocols.
- the memory 111 may include instructions implementing a text encoder 116 when executed by the processor 109 .
- the memory 111 may include instructions for full functionality of a text encoder 116 while, in other embodiments, the memory 111 may include instructions for communicating with a pre-configured text encoder to encode model knowledge assessment items.
- the text encoder 116 may also, in some implementations, augment text encodings generated by a pre-configured text encoder to include additional specifications for test items, such as admissible or target lexical metric ranges such as type-token ratio, Flesch Reading Ease, and Coleman-Liau Index.
- the instructions implementing clustering 118 may use other methods to group model knowledge assessment items.
- instructions for clustering 118 may implement a classifier trained to place items into predetermined groups.
- clustering 118 may use fully unsupervised learning or may be configured to take into account predetermined class labels.
- the items may be placed into subgroups based on semantic and lexical similarities identified based on the numeric representations of the model items, which generally results in more meaningful and accurate groupings than those that may be developed based on reviewing only text of the model knowledge assessment items.
- clustering 118 may group items based on all information included in, for example, an encoding vector.
- the memory 111 may include instructions implementing input generation 120 when executed by the processor 109 .
- the instructions for input generation 120 generally create conditioning input to transmit to a NLG 112 using text of the model assessment items and information generated by clustering 118 including, in some examples, groupings of model knowledge assessment items, centroids for the groupings or clusters identified through clustering 118 , and numeric representations of the model knowledge assessment items.
- Input generation 120 may generally format conditioning input for each grouping of the model items and provide each conditioning input to the NLG 112 in turn to generate knowledge assessment items responsive to the conditioning input.
- input generation 120 may also incorporate feedback from other components of the NLG interface 110 to generate or update conditioning inputs to the NLG 112 .
- a user interface may provide the generated items to the user device and the user interface may provide a mechanism for users to provide feedback on generated items.
- input generation 120 may include instructions to reorder inputs responsive to continued negative feedback about items generated using a particular conditioning input.
- input generation 120 may receive user edits to text of generated items and may augment the conditioning input with the edited portions of a generated item to regenerate the item.
- a conditioning input may include input corresponding to each model knowledge assessment item in a group or cluster of model knowledge assessment items. Such input may be formatted based on a specific API or interface of the NLG 112 but may generally include a series of characters representing the text and formatting of a model knowledge assessment items. Such inputs are generally evaluated by the NLG 112 as examples, where the NLG 112 evaluates the inputs for patterns, similarities, and other shared characteristics of the inputs before generating a new knowledge assessment item mirroring the inputs. This approach may be referred to as “few-shot learning” and is described in greater detail with respect to FIG. 4 herein.
- Input generation 120 may include instructions for generating the individual inputs corresponding to the model items and ordering the individual inputs to form a conditioning input. For example, input generation 120 may parse tokens forming each of the model items within a cluster and then concatenate the tokens to form a conditioning input of input tokens. Input generation 120 may also format generated conditioning input according to the interface provided by the NLG 112 and communicate the conditioning input to the NLG 112 . For example, input generation 120 may provide a terminating character or token where used by the NLG 112 and/or may provide a token limit for generated items with the conditioning input. In some examples, input generation 120 may provide a description of the example task or other information used by the NLG 112 .
- input generation 120 may order inputs corresponding to the model items based on a distance between the numerical representations of the model items and the centroid or prototype of the cluster in the multidimensional space mapped by the clustering 118 instructions. Accordingly, items more representative of the group or cluster as a whole are given more weight by the NLG 112 by virtue of the architecture of the NLG 112 . For example, in some embodiments, items positioned later in the conditioning input are more heavily weighted when choosing output tokens that items positioned earlier in the conditioning input. Accordingly, items closer to the centroid or prototype of the cluster may be positioned at the end of the conditioning input.
- the instructions for input generation 120 may provide other portions of the conditioning input, such as seeds used by the NLG 112 to generate knowledge assessment items. Further, the instructions for input generation 120 may format conditioning input based on specifications and/or interfaces of the NLG 112 . For example, where the interface of the NLG 112 expects the conditioning input to be terminated by a particular token (e.g., a null or specific terminating character), the instructions for input generation 120 may add the particular token to the conditioning input before providing the conditioning input to the NLG 112 .
- a particular token e.g., a null or specific terminating character
- the memory 111 may include instructions for implementing item verification 122 when executed by the processor 109 .
- the instructions for item verification 122 may include instructions for performing checks of items generated by the NLG 112 .
- item verification 122 may review text of a generated item to ensure that the item does not include undesirable elements, such as repeated answer choices, nonsensical phrases, repeated words, and the like.
- Item verification 122 may discard items that include such undesirable elements and may, in some embodiments, flag undesirable elements and redirect them, providing the conditioning input to the NLG 112 a second time to generate a new item.
- item verification 122 may also analyze additional aspects of generated items to compare to user specifications.
- item verification 122 may analyze lexical complexity, word count, question stem length, reading level, or other specifications to ensure that generated items match the user specifications before providing the item to a user device or storing the item with other generated items.
- User specifications may include, for example type-token ratio, Flesch Reading Ease, Coleman-Liau Index, and the like. Performing item verification 122 concerning target user specifications may be especially important, for example, when generated items are intended for a particular grade level, reading ability, or other metric which may be underrepresented in text corpuses used to train NLGs 112 .
- Item verification 122 may further do a duplication or similarity check to ensure that the generated item is not either a duplicate of a model item or similar enough to a model item that it may be less useful to, for example, an entity providing test preparation looking for a variety of testing items.
- a duplication or similarity check may compare the tokens of the generated item to tokens of the model items and may discard a generated item when more than a threshold number of sequential tokens in the generated item match a sequence of items in a model item.
- a duplication or similarity check may also be performed over archive items (e.g., publicly available items or items previously generated for a user using the NLG interface 110 ) to ensure that the generated item represents new content. Accordingly, a user can continually request new items in a content area and without being provided with the same or very similar items, as duplicate items could be difficult to manage when, for example, providing randomized sets of items for assessments as an individual assessment could include the same item more than once.
- the NLG interface 110 may include or interact with additional software or hardware components for various functions.
- an authorization service may be utilized to connect user devices to the NLG interface 110 .
- Users may sign into the NLG interface 110 and may be directed to an authorization service.
- the user may access data and use the user interface 126 to interact with the NLG interface 110 .
- the NLG interface 110 may be accessible by means other than the user interface 126 .
- various users may connect to the NLG interface 110 through, for example, command line instructions, and may provide commands to the NLG interface 110 through scripts, text commands, or means other than a user interface 126 .
- Additional services and/or applications may be utilized by the NLG interface 110 to perform functions such as querying databases, communicating between components, managing compute resources delegated to the NLG interface 110 , and other functions.
- the NLG interface 110 may be implemented using various computing systems.
- an example computing system 200 may be used for implementing various embodiments in the examples described herein.
- processor 109 and memory 111 may be located at one or several computing systems 200 .
- user device 106 is also implemented by a computing system 200 .
- This disclosure contemplates any suitable number of computing systems 200 .
- a computing system 200 may be a server, a desktop computing system, a mainframe, a mesh of computing systems, a laptop or notebook computing system, a tablet computing system, an embedded computer system, a system-on-chip, a single-board computing system, or a combination of two or more of these.
- the computing system 200 may include one or more computing systems; be unitary or distributed; span multiple locations; span multiple locations; span multiple machines; span multiple data centers; or reside in a cloud, which may include one or more cloud components in one or more networks.
- Computing system 200 includes a bus 210 (e.g., an address bus and a data bus) or other communication mechanism for communicating information, which interconnects subsystems and devices, such as processor 208 , memory 202 (e.g., RAM), static storage 204 (e.g., ROM), dynamic storage 206 (e.g., magnetic or optical), communications interface 216 (e.g., modem, Ethernet card, a network interface controller (NIC) or network adapter for communicating with an Ethernet or other wire-based network, a wireless NIC (WNIC) or wireless adapter for communicating with a wireless network, such as a WI-FI network), input/output (I/O) interface 220 (e.g., keyboard, keypad, mouse, microphone).
- the computing system 200 may include one or more of any such components.
- processor 208 includes hardware for executing instructions, such as those making up a computer program.
- the processor 208 circuitry includes circuitry for performing various processing functions, such as executing specific software for perform specific calculations or tasks.
- I/O interface 220 includes hardware, software, or both, providing one or more interfaces for communication between computing system 200 and one or more I/O devices. Computing system 200 may include one or more of these I/O devices, where appropriate. One or more of these I/O devices may enable communication between a person and computing system 200 .
- communications interface 216 includes hardware, software, or both providing one or more interfaces for communication (such as, for example, packet-based communication) between computing system 200 and one or more other computer systems or one or more networks.
- One or more memory buses (which may each include an address bus and a data bus) may couple processor 208 to memory 202 .
- Bus 210 may include one or more memory buses, as described below.
- one or more memory management units (MMUs) reside between processor 208 and memory 202 and facilitate accesses to memory 202 requested by processor 208 .
- bus 210 includes hardware, software, or both coupling components of computing system 200 to each other.
- computing system 200 performs specific operations by processor 208 executing one or more sequences of one or more instructions contained in memory 202 .
- instructions for the text encoder 116 , clustering 118 , input generation 120 , and item verification 122 may be contained in memory 202 and may be executed by the processor 208 .
- Such instructions may be read into memory 202 from another computer readable/usable medium, such as static storage 204 or dynamic storage 206 .
- static storage 204 or dynamic storage 206 such as static storage 204 or dynamic storage 206 .
- hard-wired circuitry may be used in place of or in combination with software instructions.
- particular embodiments are not limited to any specific combination of hardware circuitry and/or software.
- the term “logic” shall mean any combination of software or hardware that is used to implement all or part of particular embodiments disclosed herein.
- Non-volatile media includes, for example, optical or magnetic disks, such as static storage 504 or dynamic storage 206 .
- Volatile media includes dynamic memory, such as memory 202 .
- Computing system 200 may transmit and receive messages, data, and instructions, including program, e.g., application code, through communications link 218 and communications interface 216 .
- Received program code may be executed by processor 208 as it is received, and/or stored in static storage 204 or dynamic storage 206 , or other storage for later execution.
- a database 214 may be used to store data accessible by the computing system 200 by way of data interface 212 .
- user data 102 and/or interface data 104 may each be stored using a database 214 .
- centroids or prototypes may be formatted as numerical representations (e.g., feature vectors) but may not encode an actual representative knowledge assessment item. Rather, the centroids may provide a theoretical center point of the cluster from which to calculate distance for ordering inputs by the input generation 120 .
- the centroids may be calculated after mapping the model items to the multidimensional space, such as through a k-means clustering algorithm. Centroids may also be estimated based on, for example, a content map of the model items.
- the numeric encodings of model items 114 , numeric representations of centroids, and cluster information may be provided to input generation 120 for generation of conditioning input to the NLG 112 .
- FIG. 4 A illustrates example processing of a conditioning input by a NLG 112 , in accordance with some embodiments of the disclosure.
- the NLG interface 110 may generate conditioning input 144 to the NLG 112 .
- input generation 120 may generate the conditioning input using numeric representations of model items, and cluster or group information including centroids or prototype items and membership of model items in a specific cluster.
- FIGS. 4 A- 4 C focus on item generation for a cluster 132 including model items 132 a , 132 b , 132 d , 132 e , and 132 f and a centroid 132 c.
- the conditioning input 144 is composed of individual input tokens (shown as Input_ 1 , Input_ 2 , Input_ 3 , Input_ 4 , and Input_ 5 in FIG. 4 A ) which are individually considered by the NLG 112 .
- Input tokens may be words, special characters (spaces, newline, end character, etc.), numerals, or other elements individually considered by the NLG 112 .
- the conditioning input 144 is generally created by concatenating input tokens making up each of the model items in the cluster or group corresponding to the conditioning input. For example, in FIG.
- model items may be formed from more tokens, and the number of tokens in a model item or in the conditioning input 144 as a whole may be limited by the NLG 112 .
- the NLG 112 may generally include logic and components for pattern identification 146 and item generation 148 .
- pattern identification 146 receives the inputs making up the conditioning input 144 sequentially. For each input, pattern identification 146 generates at least one contextual embedding for the input (e.g., a contextual embedding vector), which represents the original token, including the token's position within the input. Pattern identification 146 may retain or store contextual embeddings for previously processed inputs in the conditioning input 144 and may determine a contextual embedding for an input token using the stored contextual embeddings for previously processed inputs.
- a contextual embedding vector e.g., a contextual embedding vector
- pattern identification 146 may determine contextual embeddings for each token in a conditioning input as a linear combination of other tokens in the conditioning input. Accordingly, contextual embeddings for later-occurring tokens may be based on a richer set of data.
- the contextual embedding for the last occurring input e.g., Input_ 5 in FIG. 4 A
- Item generation 148 may generally sample from the probability distribution produced by pattern identification 146 to produce a next token (e.g., Output_ 1 , Output_ 2 , and Output_ 3 in FIG. 4 A ).
- An item is generally composed of each of the output tokens chosen by item generation 148 and may be terminated, in some examples, when the output obtained from the probability distribution is an end or termination token.
- the output tokens generated by item generation 148 may be provided to pattern identification 146 .
- Pattern identification 146 may then process the output token by creating a contextual embedding of the token in view of the stored contextual embeddings for the input tokens to generate a probability distribution for the next output token.
- Item generation 148 may then sample from this updated probability distribution to generate the next output token. Item generation 148 may generate the Item from the outputs once a terminating output token is produced. This item may be referred to herein as a raw item or generated item, indicating that the item is composed of output from the NLG 112 before additional editing, content checks, and the like.
- Pattern identification 146 and item generation 148 may be implemented by any architecture processing tokens in the above described manner to complete next token prediction language generation.
- the NLG 112 may be implemented by a transformer language generation model.
- pattern identification 146 may be implemented by any number of successive transformer blocks, where each transformer block processes the input tokens sequentially, producing and storing contextual embeddings for each processed token.
- Transformer blocks may produce a contextual embedding for a new token by comparing the new token to stored contextual embeddings for each already-processed token using multi-head attention.
- the transformer block may store the contextual embedding for the new token and pass the contextual embedding for the new token to the next transformer block in sequence within pattern identification 146 .
- the next transformer block may then use the received contextual embedding and stored contextual embeddings to produce an additional contextual embedding for the new element.
- the transformer blocks continue in sequence, storing and producing contextual embeddings for a new input token. As each transformer block builds upon the contextual embedding produced by the previous transformer block, each successive transformer block is able to produce a contextual embedding capturing an increasingly rich and sophisticated representation of the original input element.
- transformer blocks may be implemented using decoders including feedforward neural networks, each with weights determined by the multi-head attention calculated by the transformer block during token processing.
- Other types of transformer architecture may also be used to implement the NLG 112 , such an encoder-only or encoder/decoder architecture.
- NLG 112 Due to the sequential processing of tokens by the NLG 112 described above, inputs appearing later in the conditioning input may be considered more heavily when generating the probability distribution, such that the ordering of model items within the conditioning input 144 affects the output tokens generated by the NLG 112 . Accordingly, an item generated by the NLG 112 may more closely resemble items provided later in the conditioning input 144 . Where items provided later in the conditioning input 144 are not representative of the cluster of model items as a whole, generated items may likewise be less representative of the cluster of model items as a whole, and may less successfully capture concepts represented by the cluster.
- FIG. 4 B shows a theoretical item mapped relative to items 132 a , 132 b , 132 d , 132 e , and 132 f and centroid 132 c of the cluster 132 .
- the mapping shown in FIG. 4 B may be, for example, a mapping of a numeric representation of the item relative to the numeric representations of the items of the cluster shown, for example, in FIG. 3 .
- the conditional input provided to the NLG 112 to generate the item of FIG. 4 B is generated by ordering the items in the cluster 132 arbitrarily, with the tokens making up item 132 a appearing last in the conditioning input 144 . As shown in FIG. 4 B , the generated item may then be similar to item 132 a .
- FIG. 4 B shows a theoretical item mapped relative to items 132 a , 132 b , 132 d , 132 e , and 132 f and centroid 132 c of the cluster 132 .
- FIG. 4 C shows another theoretical item mapped relative to the items and centroid of the cluster 132 .
- the items of the cluster 132 are ordered based on a distance between the numeric representation of the item and the centroid 132 c of the cluster.
- items closer to the centroid 132 c may be ordered last in the conditioning input 144 so that the generated item is more likely to be representative of the cluster 132 as a whole.
- the conditioning input 144 may be generated such that the tokens making up item 132 f appear last in the conditional input 144 .
- each of the items may be ordered within the conditioning input 144 in this way, such that the item furthest from the centroid 132 c is considered first, with the remaining items in the cluster 132 appearing in the conditioning input 144 in descending order of distance to the centroid 132 c.
- FIGS. 4 A, 4 B, and 4 C are described with respect to a decoder transformer based natural language model where the last input in a conditioning input is given a higher weight
- other transformer based natural language models may be used, and the inputs ordered differently, without departing from the scope of this disclosure.
- the NLG 112 may be implemented using an architecture where the first item in the conditioning input 144 is given a higher weight and the items may be ordered such that the item closest to the centroid (e.g., most representative of the cluster) is considered first to generate an item representative of the cluster.
- FIG. 6 illustrates an example process 300 for generating conditioning input for a NLG, in accordance with various embodiments of the disclosure.
- the process 300 may be executed by various components of the NLG interface 110 .
- the NLG interface 110 receives a set of model knowledge assessment items.
- the NLG interface 110 may receive the set of model knowledge assessment items directly via a user request (e.g., the user may upload or otherwise provide the model items with a request to the NLG interface 110 ).
- the NLG interface 110 may retrieve model knowledge assessment items from a remote location based on information provided with the user request. For example, a user request may include a request to generate items for practice of a standardized test, without providing model items.
- the NLG interface 110 may retrieve model items from a public repository of published practice items for the standardized test responsive to the user request.
- the knowledge assessment items may be provided in various formats including plain text, xml, json, csv, or other formats. Accordingly, the NLG interface 110 may include instructions for parsing the model knowledge assessment items from various formats.
- the model knowledge assessment items are provided to the text encoder 116 once the items are received by the NLG interface 110 .
- the NLG interface 110 generates numeric representations of the model knowledge assessment items.
- the text encoder 116 may generate numeric representations of the model knowledge assessment items, such as feature vectors.
- the text items may be provided to an encoder (e.g., a USE or BERT encoder) which may generate feature vectors encoding the model items, including lexical features, syntactic features, formatting, and other features.
- the text encoder 116 within the NLG interface 110 may perform the initial encoding function.
- the text encoder 116 may augment the generated feature vectors to include, for example, specifications for knowledge assessment items, metadata relating to the items (e.g., difficulty scores, categorizations, item type, etc.), concepts covered by the items within a content domain and other information.
- the NLG interface 110 generates clusters of model knowledge assessment items based on the numeric representations of the model knowledge assessment items.
- a clustering component 118 of the NLG interface 110 may group the model knowledge assessment items using the numeric representations of the model knowledge assessment items using, in various examples, clustering algorithms predicated on a set number of clusters, classifiers trained using similar items and metadata pairings, and other methods.
- clustering 118 may dynamically determine a number of clusters using the specific data. Accordingly, clustering 118 may be accomplished using various methods without departing from the scope of the present disclosure.
- clustering 118 may calculate centroids or theoretical centers for each cluster of model knowledge assessment items for use in generating conditioning input for each of the clusters of model knowledge assessment items.
- FIG. 7 illustrates an example process 400 for generating conditioning input for a NLG, in accordance with various embodiments of the disclosure.
- the process 400 may be a specific implementation of the process 300 . Like the process 300 , the process 400 may be performed by various components of the NLG interface 110 .
- the NLG interface 110 receives a set of model knowledge assessment items.
- the NLG interface 110 generates multi-dimensional feature vectors corresponding to each of the model knowledge assessment items. Blocks 402 and 404 may be implemented using processes and elements of the NLG interface 110 discussed with respect to blocks 302 and 304 of process 300 .
- the NLG interface 110 augments the multi-dimensional feature vectors by adding specifications for generated knowledge assessment items to the multi-dimensional feature vectors.
- added specifications may include desired length of the items, lexical indicators of the generated items, and the like.
- block 406 may add calculations of the specifications for the model items to the multi-dimensional feature vectors. For example, where a desired word count for generated items is provided as a user specification, the text encoder 116 may count words in each of the model items and update the multi-dimensional feature vectors with the word counts.
- FIG. 8 illustrates an example process 500 for performing item verification of items generated by a NLG 112 , in accordance with various embodiments of the disclosure.
- the process 500 may be performed by various components of the NLG interface 110 .
- the NLG interface 110 receives a raw knowledge assessment item from a NLG model 112 .
- a raw knowledge assessment item may be an item obtained directly from the NLG model 112 without further editing or analysis by the NLG interface 110 .
- the raw knowledge assessment item may include, in various embodiments, a question stem, answer choices, diagrams, and/or passages or other material provided in conjunction with the item.
- the raw knowledge assessment item may be provided in text format including formatting.
- Raw knowledge assessment items may be provided to item verification 122 components of the NLG interface 110 in various implementations.
- the NLG interface 110 analyzes one or more features of the raw knowledge assessment item.
- item verification 122 may perform high level checks, such as scanning the raw item for repeated words, repeated answer choices, or other easily identifiable issues. Item verification 122 may also calculate or determine other features such as word count, number of answer choices, etc. Further, in some embodiments, item verification 122 may be further configured to check more complex features such as content coverage, coherency, reading ease, etc.
- the NLG interface 110 determines whether the one or more features of the raw knowledge assessment item match specifications for knowledge assessment items provided by a user.
- the specifications may, in various implementations, be provided by the user through the user interface 126 . In other implementations, specifications may be standardized across test authoring entities and may be retrieved from other locations, such as user data 102 .
- the NLG interface 110 executes block 508 and saves the raw knowledge assessment item as a generated knowledge assessment item. Where the one or more features of the raw knowledge assessment item do not match the specifications for knowledge assessment items, the NLG interface 110 executes block 510 and discards or re-generates the raw knowledge assessment item. Where item verification 122 can easily identify and correct unmet specifications (e.g., repeated answer choices or repeated words), the item may be re-generated by providing an edited version of the item to input generation 120 , which may edit the conditioning input used to generate the item to include the edits to the item. The updated conditioning input may then be provided to the NLG 112 to regenerate the item.
- unmet specifications e.g., repeated answer choices or repeated words
- the NLG interface 110 receives an edit to the knowledge assessment item from the user.
- the edit may be made via the user interface 126 an may include, in various embodiments, deleting or adding text to the question stem, deleting or adding text to answer choices, deletion and regeneration of one or more answer choices, deletion and regeneration of the question stem, and the like.
- the edit may be received by input generation 120 components of the NLG interface 110 for regeneration of the item based on the edit or edits.
- the NLG interface 110 generates, using the NLG 112 model, at least a portion of the knowledge assessment item based on the updated conditioning input.
- the portion of the item generated may be generated by providing the updated conditioning input to the NLG 112 .
- the re-generated item may be analyzed by item verification 122 of the NLG interface 110 before being returned to the user. In other embodiments, the regenerated item may be returned directly to the user interface 126 .
- the NLG interface 110 may track or otherwise record edits made to generated items (e.g., with interface data 104 ) in order to refine methods used by the NLG interface 110 .
- input generation 120 may update the conditioning input to reduce influence of the item (e.g., by reordering the inputs of the conditioning input such that the input corresponding to the item is considered after other items by the NLG 112 ). Accordingly, user interaction with the NLG interface 110 may provide additional feedback for generating large volumes of knowledge assessment items using a NLG 112 .
- the NLG interface 110 may direct various NLG 112 models to provide large volumes of usable knowledge assessment items.
- the NLG interface 110 may work with the architecture of the NLG 112 and the process of few-shot learning to create large volumes of knowledge assessment items covering various subject areas and may provide technically accurate items.
- the NLG interface 110 may include additional features for dynamic editing and regeneration of knowledge assessment items and may learn and refine inputs over time to converge on conditioning inputs which generate higher quality items. Accordingly, the NLG interface 110 works in conjunction with the NLG 112 to improve the functioning of the NLG 112 for generation of knowledge assessment items.
- the technology described herein may be implemented as logical operations and/or modules in one or more systems.
- the logical operations may be implemented as a sequence of processor-implemented steps directed by software programs executing in one or more computer systems and as interconnected machine or circuit modules within one or more computer systems, or as a combination of both.
- the descriptions of various component modules may be provided in terms of operations executed or effected by the modules.
- the resulting implementation is a matter of choice, dependent on the performance requirements of the underlying system implementing the described technology.
- the logical operations making up the embodiments of the technology described herein are referred to variously as operations, steps, objects, or modules.
- logical operations may be performed in any order, unless explicitly claimed otherwise or a specific order is inherently necessitated by the claim language.
- articles of manufacture are provided as computer program products that cause the instantiation of operations on a computer system to implement the procedural operations.
- One implementation of a computer program product provides a non-transitory computer program storage medium readable by a computer system and encoding a computer program. It should further be understood that the described technology may be employed in special purpose devices independent of a personal computer.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
Claims (18)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/809,217 US12277870B2 (en) | 2021-06-28 | 2022-06-27 | Interface to natural language generator for generation of knowledge assessment items |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202163215794P | 2021-06-28 | 2021-06-28 | |
| US17/809,217 US12277870B2 (en) | 2021-06-28 | 2022-06-27 | Interface to natural language generator for generation of knowledge assessment items |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20220415203A1 US20220415203A1 (en) | 2022-12-29 |
| US12277870B2 true US12277870B2 (en) | 2025-04-15 |
Family
ID=84541152
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/809,217 Active 2043-06-16 US12277870B2 (en) | 2021-06-28 | 2022-06-27 | Interface to natural language generator for generation of knowledge assessment items |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US12277870B2 (en) |
| WO (1) | WO2023278980A1 (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20240378397A1 (en) * | 2023-05-09 | 2024-11-14 | Jim Liu | Dynamic content generation method |
Families Citing this family (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US12277870B2 (en) * | 2021-06-28 | 2025-04-15 | Prometric Llc | Interface to natural language generator for generation of knowledge assessment items |
| US20240282214A1 (en) * | 2022-02-25 | 2024-08-22 | MGHL Partners Inc. | System for categorizing items arising from automated item generation |
| US12423523B2 (en) * | 2022-12-14 | 2025-09-23 | International Business Machines Corporation | Generating semantic triplets from unstructured text using named entities |
| CN115861824B (en) * | 2023-02-23 | 2023-06-06 | 汕头大学 | Remote Sensing Image Recognition Method Based on Improved Transformer |
| CN117290429B (en) * | 2023-11-24 | 2024-02-20 | 山东焦易网数字科技股份有限公司 | Method for calling data system interface through natural language |
| US20250217600A1 (en) * | 2023-12-29 | 2025-07-03 | Shopify Inc. | System and method for generating review summaries based on classified topics |
| CN119337114B (en) * | 2024-12-18 | 2025-04-01 | 北京火山引擎科技有限公司 | Method, apparatus, medium and product for language model evaluation |
| CN119558412A (en) * | 2025-02-05 | 2025-03-04 | 山东浪潮科学研究院有限公司 | A large language model evaluation system and method |
Citations (40)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20030216913A1 (en) * | 2002-05-14 | 2003-11-20 | Microsoft Corporation | Natural input recognition tool |
| US8517738B2 (en) | 2008-01-31 | 2013-08-27 | Educational Testing Service | Reading level assessment method, system, and computer program product for high-stakes testing applications |
| US8838659B2 (en) | 2007-10-04 | 2014-09-16 | Amazon Technologies, Inc. | Enhanced knowledge repository |
| US9082309B1 (en) | 2013-03-15 | 2015-07-14 | Querium Corporation | Dynamic question template system and architecture |
| US9098492B2 (en) | 2005-08-01 | 2015-08-04 | Amazon Technologies, Inc. | Knowledge repository |
| US20150324459A1 (en) * | 2014-05-09 | 2015-11-12 | Chegg, Inc. | Method and apparatus to build a common classification system across multiple content entities |
| US20160162456A1 (en) * | 2014-12-09 | 2016-06-09 | Idibon, Inc. | Methods for generating natural language processing systems |
| US20170169829A1 (en) * | 2015-12-11 | 2017-06-15 | Microsoft Technology Licensing, Llc | Language and domain independent model based approach for on-screen item selection |
| US20170177660A1 (en) * | 2015-12-16 | 2017-06-22 | Adobe Systems Incorporated | Natural language embellishment generation and summarization for question-answering systems |
| US20170371861A1 (en) * | 2016-06-24 | 2017-12-28 | Mind Lakes, Llc | Architecture and processes for computer learning and understanding |
| US9858828B1 (en) | 2013-03-15 | 2018-01-02 | Querium Corporation | Expert systems and methods for dynamic assessment and assessment authoring |
| US20180101598A1 (en) * | 2016-10-10 | 2018-04-12 | International Business Machines Corporation | Personalized Approach to Handling Hypotheticals in Text |
| US20180336183A1 (en) * | 2017-05-22 | 2018-11-22 | International Business Machines Corporation | Deep Embedding for Natural Language Content Based on Semantic Dependencies |
| US10249207B2 (en) | 2016-01-19 | 2019-04-02 | TheBeamer, LLC | Educational teaching system and method utilizing interactive avatars with learning manager and authoring manager functions |
| US20190272325A1 (en) * | 2014-11-25 | 2019-09-05 | Truthful Speaking, Inc. | Written word refinement system & method for truthful transformation of spoken and written communications |
| US10504268B1 (en) | 2017-04-18 | 2019-12-10 | Educational Testing Service | Systems and methods for generating facial expressions in a user interface |
| US10504376B2 (en) | 2006-10-25 | 2019-12-10 | Reliant Exams As | System and method for improving the quality of computer generated exams |
| US20190378429A1 (en) * | 2018-06-07 | 2019-12-12 | SelectQ Corporation | Intelligent and Contextual System for Knowledge Progression and Quiz Management |
| US10664862B1 (en) * | 2011-06-21 | 2020-05-26 | Contextlogic, Inc. | Topic inference based contextual content |
| US20200193317A1 (en) * | 2017-05-19 | 2020-06-18 | Riiid Inc. | Method, device and computer program for estimating test score |
| US20200218988A1 (en) * | 2019-01-08 | 2020-07-09 | International Business Machines Corporation | Generating free text representing semantic relationships between linked entities in a knowledge graph |
| US20200334331A1 (en) * | 2019-04-17 | 2020-10-22 | International Business Machines Corporation | Machine learned sentence span inclusion judgments |
| US20200342054A1 (en) * | 2019-04-29 | 2020-10-29 | International Business Machines Corporation | Traversal-based sentence span judgements |
| US20200356729A1 (en) * | 2017-12-15 | 2020-11-12 | Microsoft Technology Licensing, Llc | Generation of text from structured data |
| US20200372106A1 (en) * | 2019-05-24 | 2020-11-26 | International Business Machines Corporation | Method and System for Language and Domain Acceleration with Embedding Evaluation |
| US20200387572A1 (en) * | 2019-06-04 | 2020-12-10 | International Business Machines Corporation | Automatic detection of context switch triggers |
| US20200410050A1 (en) * | 2019-06-27 | 2020-12-31 | International Business Machines Corporation | Deep learning approach to computing spans |
| US20210110895A1 (en) * | 2018-06-19 | 2021-04-15 | Ellipsis Health, Inc. | Systems and methods for mental health assessment |
| US20210149994A1 (en) * | 2019-11-15 | 2021-05-20 | 42 Maru Inc. | Device and method for machine reading comprehension question and answer |
| US20210183484A1 (en) * | 2019-12-06 | 2021-06-17 | Surgical Safety Technologies Inc. | Hierarchical cnn-transformer based machine learning |
| US20210295822A1 (en) * | 2020-03-23 | 2021-09-23 | Sorcero, Inc. | Cross-context natural language model generation |
| US20210366303A1 (en) * | 2018-02-28 | 2021-11-25 | Obrizum Group Ltd. | Learning management systems and methods therefor |
| US20210383711A1 (en) * | 2018-06-07 | 2021-12-09 | Thinkster Learning Inc. | Intelligent and Contextual System for Test Management |
| US20220100676A1 (en) * | 2020-09-29 | 2022-03-31 | Microsoft Technology Licensing, Llc | Dynamic cache management in beam search |
| US20220108195A1 (en) * | 2020-10-01 | 2022-04-07 | Crowdsmart, Inc. | Infinitely scaling a/b testing |
| US20220121824A1 (en) * | 2019-11-25 | 2022-04-21 | Boe Technology Group Co., Ltd. | Method for determining text similarity, method for obtaining semantic answer text, and question answering method |
| US20220188661A1 (en) * | 2020-11-05 | 2022-06-16 | Birdview Films, LLC | Stateful, Real-Time, Interactive, and Predictive Knowledge Pattern Machine |
| US20220309356A1 (en) * | 2021-03-29 | 2022-09-29 | Citrix Systems, Inc. | Web elements-based virtual assistant for distributed applications |
| US20220415203A1 (en) * | 2021-06-28 | 2022-12-29 | ACADEMIC MERIT LLC d/b/a FINETUNE LEARNING | Interface to natural language generator for generation of knowledge assessment items |
| US20230419044A1 (en) | 2022-06-27 | 2023-12-28 | Prometric Llc | Tagging for subject matter or learning schema |
-
2022
- 2022-06-27 US US17/809,217 patent/US12277870B2/en active Active
- 2022-06-27 WO PCT/US2022/073192 patent/WO2023278980A1/en not_active Ceased
Patent Citations (40)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20030216913A1 (en) * | 2002-05-14 | 2003-11-20 | Microsoft Corporation | Natural input recognition tool |
| US9098492B2 (en) | 2005-08-01 | 2015-08-04 | Amazon Technologies, Inc. | Knowledge repository |
| US10504376B2 (en) | 2006-10-25 | 2019-12-10 | Reliant Exams As | System and method for improving the quality of computer generated exams |
| US8838659B2 (en) | 2007-10-04 | 2014-09-16 | Amazon Technologies, Inc. | Enhanced knowledge repository |
| US8517738B2 (en) | 2008-01-31 | 2013-08-27 | Educational Testing Service | Reading level assessment method, system, and computer program product for high-stakes testing applications |
| US10664862B1 (en) * | 2011-06-21 | 2020-05-26 | Contextlogic, Inc. | Topic inference based contextual content |
| US9858828B1 (en) | 2013-03-15 | 2018-01-02 | Querium Corporation | Expert systems and methods for dynamic assessment and assessment authoring |
| US9082309B1 (en) | 2013-03-15 | 2015-07-14 | Querium Corporation | Dynamic question template system and architecture |
| US20150324459A1 (en) * | 2014-05-09 | 2015-11-12 | Chegg, Inc. | Method and apparatus to build a common classification system across multiple content entities |
| US20190272325A1 (en) * | 2014-11-25 | 2019-09-05 | Truthful Speaking, Inc. | Written word refinement system & method for truthful transformation of spoken and written communications |
| US20160162456A1 (en) * | 2014-12-09 | 2016-06-09 | Idibon, Inc. | Methods for generating natural language processing systems |
| US20170169829A1 (en) * | 2015-12-11 | 2017-06-15 | Microsoft Technology Licensing, Llc | Language and domain independent model based approach for on-screen item selection |
| US20170177660A1 (en) * | 2015-12-16 | 2017-06-22 | Adobe Systems Incorporated | Natural language embellishment generation and summarization for question-answering systems |
| US10249207B2 (en) | 2016-01-19 | 2019-04-02 | TheBeamer, LLC | Educational teaching system and method utilizing interactive avatars with learning manager and authoring manager functions |
| US20170371861A1 (en) * | 2016-06-24 | 2017-12-28 | Mind Lakes, Llc | Architecture and processes for computer learning and understanding |
| US20180101598A1 (en) * | 2016-10-10 | 2018-04-12 | International Business Machines Corporation | Personalized Approach to Handling Hypotheticals in Text |
| US10504268B1 (en) | 2017-04-18 | 2019-12-10 | Educational Testing Service | Systems and methods for generating facial expressions in a user interface |
| US20200193317A1 (en) * | 2017-05-19 | 2020-06-18 | Riiid Inc. | Method, device and computer program for estimating test score |
| US20180336183A1 (en) * | 2017-05-22 | 2018-11-22 | International Business Machines Corporation | Deep Embedding for Natural Language Content Based on Semantic Dependencies |
| US20200356729A1 (en) * | 2017-12-15 | 2020-11-12 | Microsoft Technology Licensing, Llc | Generation of text from structured data |
| US20210366303A1 (en) * | 2018-02-28 | 2021-11-25 | Obrizum Group Ltd. | Learning management systems and methods therefor |
| US20190378429A1 (en) * | 2018-06-07 | 2019-12-12 | SelectQ Corporation | Intelligent and Contextual System for Knowledge Progression and Quiz Management |
| US20210383711A1 (en) * | 2018-06-07 | 2021-12-09 | Thinkster Learning Inc. | Intelligent and Contextual System for Test Management |
| US20210110895A1 (en) * | 2018-06-19 | 2021-04-15 | Ellipsis Health, Inc. | Systems and methods for mental health assessment |
| US20200218988A1 (en) * | 2019-01-08 | 2020-07-09 | International Business Machines Corporation | Generating free text representing semantic relationships between linked entities in a knowledge graph |
| US20200334331A1 (en) * | 2019-04-17 | 2020-10-22 | International Business Machines Corporation | Machine learned sentence span inclusion judgments |
| US20200342054A1 (en) * | 2019-04-29 | 2020-10-29 | International Business Machines Corporation | Traversal-based sentence span judgements |
| US20200372106A1 (en) * | 2019-05-24 | 2020-11-26 | International Business Machines Corporation | Method and System for Language and Domain Acceleration with Embedding Evaluation |
| US20200387572A1 (en) * | 2019-06-04 | 2020-12-10 | International Business Machines Corporation | Automatic detection of context switch triggers |
| US20200410050A1 (en) * | 2019-06-27 | 2020-12-31 | International Business Machines Corporation | Deep learning approach to computing spans |
| US20210149994A1 (en) * | 2019-11-15 | 2021-05-20 | 42 Maru Inc. | Device and method for machine reading comprehension question and answer |
| US20220121824A1 (en) * | 2019-11-25 | 2022-04-21 | Boe Technology Group Co., Ltd. | Method for determining text similarity, method for obtaining semantic answer text, and question answering method |
| US20210183484A1 (en) * | 2019-12-06 | 2021-06-17 | Surgical Safety Technologies Inc. | Hierarchical cnn-transformer based machine learning |
| US20210295822A1 (en) * | 2020-03-23 | 2021-09-23 | Sorcero, Inc. | Cross-context natural language model generation |
| US20220100676A1 (en) * | 2020-09-29 | 2022-03-31 | Microsoft Technology Licensing, Llc | Dynamic cache management in beam search |
| US20220108195A1 (en) * | 2020-10-01 | 2022-04-07 | Crowdsmart, Inc. | Infinitely scaling a/b testing |
| US20220188661A1 (en) * | 2020-11-05 | 2022-06-16 | Birdview Films, LLC | Stateful, Real-Time, Interactive, and Predictive Knowledge Pattern Machine |
| US20220309356A1 (en) * | 2021-03-29 | 2022-09-29 | Citrix Systems, Inc. | Web elements-based virtual assistant for distributed applications |
| US20220415203A1 (en) * | 2021-06-28 | 2022-12-29 | ACADEMIC MERIT LLC d/b/a FINETUNE LEARNING | Interface to natural language generator for generation of knowledge assessment items |
| US20230419044A1 (en) | 2022-06-27 | 2023-12-28 | Prometric Llc | Tagging for subject matter or learning schema |
Non-Patent Citations (3)
| Title |
|---|
| International Search Report and Written Opinion dated Nov. 21, 2023 in connection with International Patent Application No. PCT/US2023/031447, 9 pages. |
| International Search Report and Written Opinion of International Patent Application No. PCT/US2022/73192, mailed Dec. 2, 2022. |
| Khan, Saad M. et al. "Catalog: An Educational Content Tagging System," 14th Intl Conf on Educational Data Mining, EDM 2021, 4 pages. |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20240378397A1 (en) * | 2023-05-09 | 2024-11-14 | Jim Liu | Dynamic content generation method |
Also Published As
| Publication number | Publication date |
|---|---|
| WO2023278980A1 (en) | 2023-01-05 |
| US20220415203A1 (en) | 2022-12-29 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12277870B2 (en) | Interface to natural language generator for generation of knowledge assessment items | |
| US11816439B2 (en) | Multi-turn dialogue response generation with template generation | |
| Tripathi et al. | Naive Bayes classification model for the student performance prediction | |
| US20200134103A1 (en) | Visualization-dashboard narration using text summarization | |
| Quamar et al. | Conversational BI: An ontology-driven conversation system for business intelligence applications | |
| JP7422767B2 (en) | natural language solutions | |
| Akhuseyinoglu et al. | Data-driven modeling of learners’ individual differences for predicting engagement and success in online learning | |
| US20230419044A1 (en) | Tagging for subject matter or learning schema | |
| CN117795581A (en) | Systems and methods for educational and psychological modeling and assessment | |
| JP2022082524A (en) | Information provision methods and devices using learning models through machine learning | |
| CN116956866A (en) | Scenario data processing method, apparatus, device, storage medium and program product | |
| Bobek et al. | KnAC: an approach for enhancing cluster analysis with background knowledge and explanations | |
| WO2024192213A2 (en) | Apparatus and method for an education platform and ecosystem using extended reality | |
| CN113761193A (en) | Log classification method and device, computer equipment and storage medium | |
| CN116956934A (en) | Task processing method, device, equipment and storage medium | |
| CN120387432A (en) | Project text intelligent generation method and device | |
| CN120277199A (en) | Child education knowledge boundary management method, system and equipment based on large model | |
| Santamarıa-Bonfil et al. | A learning ecosystem for linemen training based on big data components and learning analytics | |
| Brown et al. | Assessing changes in reliability methods over time: An unsupervised text mining approach | |
| Hilmarsson | Using machine learning for predicting the likelihood of upper secondary school student dropout | |
| CN120725836B (en) | Group teaching strategy generation method and device based on intelligent learning cloud platform | |
| KR102737471B1 (en) | Method for developing management system | |
| JP7442430B2 (en) | Examination support system and examination support method | |
| WO2025081750A1 (en) | Method and apparatus for programming | |
| Chen et al. | Studying the Effects of Collaboration in Interactive Theme Discovery Systems |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: ACADEMIC MERIT LLC D/B/A FINETUNE LEARNING, MASSACHUSETTS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HAMER, JESSE ANDREW LEWIS;ALMEIDA, TIAGO LIMA;FOSTER, CHARLES;AND OTHERS;SIGNING DATES FROM 20220701 TO 20220713;REEL/FRAME:060565/0646 |
|
| AS | Assignment |
Owner name: ACADEMIC MERIT LLC D/B/A FINETUNE LEARNING, MAINE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KHAN, SAAD MASOOD;REEL/FRAME:060676/0279 Effective date: 20210628 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| AS | Assignment |
Owner name: PROMETRIC LLC, MARYLAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ACADEMIC MERIT LLC D/B/A FINETUNE LEARNING;REEL/FRAME:070131/0637 Effective date: 20220801 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| AS | Assignment |
Owner name: PROMETRIC LLC, MARYLAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VISPOEL, SARA;REEL/FRAME:070505/0365 Effective date: 20250313 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| CC | Certificate of correction |