Disclosure of Invention
The embodiment of the application provides a data generation method, a device, electronic equipment and a computer readable storage medium, which can improve the effectiveness and the integrity of extracted gesture fragment data and further improve the precision of an acquired gesture data set.
In a first aspect, an embodiment of the present application provides a data generating method, including:
acquiring gesture motion data;
identifying at least one target peak from the gesture motion data;
Extracting at least one gesture fragment data from the gesture motion data according to each target peak value;
labeling the gesture fragment data to obtain a target gesture data set.
In a second aspect, an embodiment of the present application further provides a data generating apparatus, where the apparatus includes:
The acquisition module is used for acquiring gesture motion data;
The identification module is used for identifying at least one target peak value from the gesture motion data;
The extraction module is used for extracting at least one gesture fragment data from the gesture motion data according to each target peak value;
And the labeling module is used for labeling each gesture fragment data to obtain a target gesture data set.
Optionally, in some embodiments of the present application, the identifying at least one target peak from the gesture motion data includes:
Identifying a local peak of the gesture motion data over at least one time window;
and screening at least one target peak value from the local peaks according to a peak value threshold.
Optionally, in some embodiments of the application, the gesture motion data includes acceleration data;
The identifying the local peak of the gesture motion data over at least one time window comprises:
For each time window, deriving the analog length data corresponding to the acceleration data in the time window to obtain a derivative data set, wherein the derivative data set comprises derivative data corresponding to the analog length data of at least one time stamp in the time window;
taking the derivative data which is larger than zero in the derivative data set as candidate peak values;
and screening the maximum value from the candidate peaks to obtain the local peak value.
Optionally, in some embodiments of the present application, the selecting a maximum value from the candidate peaks to obtain the local peak includes:
comparing adjacent candidate peaks according to the sequence of the time stamps, and comparing the larger candidate peak with the next candidate peak until the comparison of the last candidate peak in the time window is finished, so as to obtain the largest candidate peak;
the largest candidate peak is set as the local peak.
Optionally, in some embodiments of the present application, the selecting a maximum value from the candidate peaks to obtain the local peak includes:
sequentially adding each candidate peak to a peak stack in the order of time stamps, and
If the current candidate peak value is larger than the previous candidate peak value, removing the previous candidate peak value from the peak value stack, and reserving the current candidate peak value;
If the current candidate peak is smaller than the previous candidate peak, retaining the current candidate peak and the previous candidate peak, and if the next candidate peak is larger than the previous candidate peak and the next candidate peak is larger than the current candidate peak, removing the current candidate peak and the previous candidate peak, and retaining the next candidate peak;
and after the last candidate peak value is added to the peak value stack in the time window and the sizes are compared, setting the candidate peak value reserved in the peak value stack or the candidate peak value at the stack top of the peak value stack as the local peak value, wherein the sizes of the candidate peak values of the peak value stack from the stack top to the stack tail are sequentially decreased.
Optionally, in some embodiments of the present application, the extracting at least one gesture segment data from the gesture motion data according to each target peak value includes:
And extracting gesture fragment data which takes the target peak value as a center and has a preset duration from the gesture motion data aiming at each target peak value.
Optionally, in some embodiments of the application, the method is applied to an augmented reality device;
labeling each gesture fragment data to obtain a target gesture data set, including:
Labeling each gesture fragment data through a pre-deployment model to obtain an original labeling result of each gesture fragment data;
Displaying the original labeling result through a virtual display screen;
obtaining a preliminary verification result in response to receiving first modification information input for the original labeling result;
the gesture fragment data, the gesture video data and the preliminary verification result are displayed in an aligned mode, and the gesture video data are synchronously shot aiming at the gesture motion data;
responding to the received second modification information input for the preliminary verification result, and obtaining a target verification result;
and generating the target gesture data set according to the target verification result.
In a third aspect, an embodiment of the present application further provides an electronic device, where the electronic device includes a memory, a processor, and a computer program stored in the memory and executable on the processor, and the computer program when executed by the processor implements the steps in the data generating method described above.
In a fourth aspect, an embodiment of the present application further provides a computer readable storage medium, where a computer program is stored, where the computer program when executed by a processor implements the steps in the data generating method described above.
In a fifth aspect, embodiments of the present application also provide a computer program product or computer program comprising computer instructions stored in a computer readable storage medium. The processor of the computer device reads the computer instructions from the computer-readable storage medium and executes the computer instructions to cause the computer device to perform the methods provided in the various alternative implementations described in the embodiments of the present application.
In summary, the embodiment of the application acquires gesture motion data, identifies at least one target peak value from the gesture motion data, extracts at least one gesture fragment data from the gesture motion data according to each target peak value, and marks each gesture fragment data to obtain a target gesture data set.
The method comprises the steps of extracting a target peak value from gesture motion data, extracting gesture fragment data from the gesture motion data based on the target peak value, and improving the effectiveness and accuracy of the gesture fragment data, so that after a target gesture data set is obtained by labeling based on the gesture fragment data, the accuracy of the target gesture data set is improved.
Detailed Description
The following description of the embodiments of the present application will be made more apparent and fully hereinafter with reference to the accompanying drawings, in which some, but not all embodiments of the application are shown. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to fall within the scope of the application.
The embodiment of the application provides a data generation method, a data generation device, electronic equipment and a computer readable storage medium. Specifically, the embodiment of the application provides a data generating device suitable for an electronic device, wherein the electronic device comprises a terminal device or a server, the terminal device comprises a mobile phone, a tablet computer, a notebook computer, a smart watch, a smart bracelet or an extended reality device, the extended reality device comprises a Virtual Reality (VR) device, an Augmented Reality (AR) device, a Mixed Reality (MR) device and the like, and the extended reality device at least comprises a product in the form of glasses or a wearable electronic device in the form of a head-mounted display and the like. The server may be an independent physical server, a server cluster or a distributed system formed by a plurality of physical servers, or a cloud server or the like for providing cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, content delivery networks (CDNs, content Delivery Network), basic cloud computing services such as big data and artificial intelligent platforms, and the servers may be directly or indirectly connected through wired or wireless communication modes.
For example, referring to fig. 1, taking an example that an augmented reality device executes the data generating method as an example, fig. 1 is a schematic view of a scene in which the augmented reality device executes the data generating method, and specifically, a specific execution procedure of the augmented reality device executing the data generating method is as follows:
The method comprises the steps that the augmented reality device 101 obtains gesture motion data, wherein the gesture motion data are acquired by the intelligent watch 102 through a built-in IMU sensor, the intelligent watch 102 sends the acquired gesture motion data to the augmented reality device 101, then, after the gesture motion data are received by the augmented reality device 101, at least one target peak value is identified from the gesture motion data, at least one gesture fragment data are extracted from the gesture motion data according to the target peak value, and the gesture fragment data are marked, so that a target gesture data set is obtained.
For example, the gesture segment data is intercepted from the gesture motion data by taking acceleration peak value, speed peak value, angular speed peak value and the like of the gesture motion as centers, so that complete and effective gesture segment data is obtained.
In summary, the embodiment of the application extracts the target peak value from the gesture motion data, and extracts the gesture fragment data from the gesture motion data based on the target peak value, so that the validity and the accuracy of the gesture fragment data are improved, and further, after the target gesture data set is obtained by labeling based on the gesture fragment data, the accuracy of the target gesture data set is also improved.
The following will describe in detail. It should be noted that the following description order of embodiments is not a limitation of the priority order of embodiments.
Referring to fig. 2, fig. 2 is a flowchart of a data generating method according to an embodiment of the present application, in which although a logic sequence is shown in the flowchart, in some cases, the steps shown or described may be performed in a different order than shown in the flowchart. Specifically, the specific flow of the data generation method is as follows:
201. gesture motion data is acquired.
The gesture motion data are acquired through an IMU sensor, for example, the gesture motion data are acquired through a smart watch and a smart bracelet which are internally provided with the IMU sensor.
Accordingly, in the embodiment of the application, the gesture motion data can be obtained by receiving the smart watch or the smart bracelet.
202. At least one target peak is identified from the gesture motion data.
The peak value refers to a local maximum value (also referred to as a maximum point) of a certain physical quantity in time or space, and for example, for gesture motion data, the peak value includes a peak value based on acceleration, a peak value of motion speed, a peak value of angular speed, and the like. The target peak is a specific number of peaks selected from the plurality of peaks. It will be appreciated that the target peaks are the result of screening, filtering, or both, of the peaks. For example, filtering out invalid peaks based on a peak threshold value results in a number of target peaks that can characterize the gesture motion, e.g., target peaks include peaks of acceleration, velocity, or magnitude of angular velocity at the vertex at the time of gesture motion, etc. It will be appreciated that each target peak corresponds to an effective gesture motion.
203. And extracting at least one gesture fragment data from the gesture motion data according to each target peak value.
It can be appreciated that, since the target peak is a valid peak after filtering, the validity and accuracy of the gesture fragment data extracted based on the target peak are ensured. For example, gesture segment data including peak values of each target are extracted from gesture motion data, so that each gesture segment data corresponds to a valid gesture motion.
It can be understood that the extraction of the gesture fragment data is performed by taking the target peak value as a reference, so that the accuracy of gesture fragment data extraction is improved, and the extraction of the gesture fragment data is facilitated to be improved to complete and effective gesture fragment data.
204. Labeling the gesture fragment data to obtain a target gesture data set.
The labeling refers to labeling of gesture types corresponding to each gesture segment data, for example, labeling gesture types such as finger click, finger movement, palm detection or palm stretching to which each gesture segment data belongs.
It can be understood that after labeling each piece of gesture segment data, gesture segment data labeled with a gesture type is obtained, and the data can be used as a target gesture data set for training a gesture recognition model, so that the trained gesture recognition model can learn the relationship between the gesture segment data and the gesture type, and further can recognize and classify various pieces of gesture segment data, analyze the gesture type corresponding to the gesture segment data, and realize the function of gesture recognition.
In summary, the embodiment of the application extracts the target peak value from the gesture motion data, and extracts the gesture fragment data from the gesture motion data based on the target peak value, so that the validity and the accuracy of the gesture fragment data are improved, and further, after the target gesture data set is obtained by labeling based on the gesture fragment data, the accuracy of the target gesture data set is also improved.
Optionally, in an embodiment of the present application, in order to further improve the validity and accuracy of the collected gesture fragment data, in an embodiment of the present application, the extracting of the gesture fragment data is controlled by taking the target peak value as a center according to a preset duration, that is, optionally, in some embodiments of the present application, the step of "extracting at least one gesture fragment data from the gesture motion data according to each target peak value" includes:
And extracting gesture fragment data which takes the target peak value as a center and has a preset duration from the gesture motion data aiming at each target peak value.
For example, for each target peak, data segments with time lengths of 500ms before and 500ms after the target peak are extracted from gesture motion data by taking the target peak as a center, and gesture segment data with time length of about 1s are obtained by combining.
It can be understood that the gesture fragment data is obtained by combining the front and rear intercepting part data fragments by taking the target peak value as the center, so that the gesture fragment data taking the gesture action as the center can be obtained, and the effectiveness and the accuracy of the gesture fragment data can be improved.
Optionally, in embodiments of the present application, local peaks of each time window may be extracted according to the time window, and an effective target peak may be obtained by means of peak threshold filtering, that is, optionally, in some embodiments of the present application, the step of "identifying at least one target peak from the gesture motion data" includes:
Identifying a local peak of the gesture motion data over at least one time window;
and screening at least one target peak value from the local peaks according to a peak value threshold.
The local peak value is the largest peak value in a plurality of time stamps corresponding to each time window, and belongs to the local maximum value. Wherein the length of the time window may be set based on the general duration of the gesture, e.g., the size of the time window is set according to the general duration of the gesture.
The peak threshold is set based on the peak values of the gestures, for example, the peak values of a plurality of gestures are collected, and the minimum peak value corresponding to each gesture is set as the peak threshold. It will be appreciated that the peak threshold is set to filter out invalid local peaks to obtain valid target peaks corresponding to a certain gesture. For example, filtering local peaks corresponding to common general gestures and general gesture actions which do not contain specific instructions can avoid obtaining invalid gesture fragment data based on invalid local peak interception, and the validity and accuracy of the gesture fragment data acquired by the scheme are improved.
For example, taking the peak for acceleration as an example, i.e. the gesture motion data comprises acceleration data, the step of "identifying a local peak of the gesture motion data over at least one time window" comprises:
For each time window, deriving the analog length data corresponding to the acceleration data in the time window to obtain a derivative data set, wherein the derivative data set comprises derivative data corresponding to the analog length data of at least one time stamp in the time window;
taking the derivative data which is larger than zero in the derivative data set as candidate peak values;
and screening the maximum value from the candidate peaks to obtain the local peak value.
For example, the derivative of the analog length data or the analog length signal corresponding to the acceleration data is performed according to the time window, and the derivative data of the zero crossing point is set as the candidate peak value. And screening the maximum value from the plurality of candidate peaks to be used as the maximum peak value corresponding to the time window, namely the local peak value.
Optionally, in the embodiment of the present application, a local peak value with a largest value may be screened out by comparing candidate peak values, that is, step "screening out a largest value from the candidate peak values to obtain the local peak value" includes:
comparing adjacent candidate peaks according to the sequence of the time stamps, and comparing the larger candidate peak with the next candidate peak until the comparison of the last candidate peak in the time window is finished, so as to obtain the largest candidate peak;
the largest candidate peak is set as the local peak.
For example, each candidate peak is compared in the order of the time stamps, and the larger candidate peak in the comparison result is compared with the next candidate peak in the order, so as to obtain the maximum value of the plurality of candidate peaks.
Optionally, in the embodiment of the present application, a local peak value with the largest value may also be selected from a plurality of candidate peak values by using a principle of a monotonic stack, that is, optionally, in some embodiments of the present application, the step of "selecting a maximum value from the candidate peak values to obtain the local peak value" includes:
sequentially adding each candidate peak to a peak stack in the order of time stamps, and
If the current candidate peak value is larger than the previous candidate peak value, removing the previous candidate peak value from the peak value stack, and reserving the current candidate peak value;
If the current candidate peak is smaller than the previous candidate peak, retaining the current candidate peak and the previous candidate peak, and if the next candidate peak is larger than the previous candidate peak and the next candidate peak is larger than the current candidate peak, removing the current candidate peak and the previous candidate peak, and retaining the next candidate peak;
and after the last candidate peak value is added to the peak value stack in the time window and the sizes are compared, setting the candidate peak value reserved in the peak value stack or the candidate peak value at the stack top of the peak value stack as the local peak value, wherein the sizes of the candidate peak values of the peak value stack from the stack top to the stack tail are sequentially decreased.
The peak stack has a monotonic attribute, and the element in the stack is always kept to be monotonically increased or monotonically decreased.
For example, in the embodiment of the present application, each candidate peak is sequentially added to a peak stack according to the order of the time stamps, and the added candidate peak and the candidate peak which is previously added and reserved are compared in size by using the peak stack, so that a larger candidate peak is always reserved in the stack by removing the smaller candidate peak to the maximum, and further, after each candidate peak is added to the peak stack and compared, the local peak with the maximum value is obtained.
For example, taking local peaks of 0.10, 0.15, 0.20, 0.18 and 0.22 respectively and corresponding time stamps of 1.0s, 1.2s, 1.3s, 1.5s and 1.7s respectively as examples, the processing steps of the peak stack specifically include:
The first step, 0.10 candidate peak value is received, so that 0.10 candidate peak value is pushed into the stack;
Step two, receiving candidate peak value 0.15, comparing candidate peak value 0.15 with candidate peak value 0.10, removing candidate peak value 0.10 and retaining candidate peak value 0.15 or pushing candidate peak value 0.15 into stack because of 0.15> 0.10;
Step three, receiving candidate peak value 0.20, comparing candidate peak value 0.20 with candidate peak value 0.15, removing candidate peak value 0.15 and retaining candidate peak value 0.20 or pushing candidate peak value 0.20 into stack because of 0.20> 0.15;
fourthly, receiving a candidate peak value 0.18, and comparing the candidate peak value 0.20 with the candidate peak value 0.18, wherein the candidate peak value 0.18 is 0.18<0.20, and the candidate peak value 0.18 is the newly added candidate peak value, and then the candidate peak value 0.18 is put into a stack, and the candidate peak value 0.20 and the candidate peak value 0.18 are reserved in the stack;
Fifth, candidate peak 0.22 is received, candidate peak 0.22 and candidate peaks 0.20 and 0.18 are compared, and due to 0.22>0.20, and 0.22>0.18, candidate peak 0.20 is removed and candidate peak 0.18 is removed, and candidate peak 0.22 is reserved, or candidate peak 0.22 is pushed onto the stack.
After the processing, only the candidate peak value 0.22 with the largest finger is reserved in the peak stack, and the candidate peak value 0.22 is taken as the current local peak value. Correspondingly, the corresponding time stamp is 1.7s.
For another example, if the target peak values are respectively 0.22 (corresponding time stamp is 1.7 s), 0.25 (corresponding time stamp is 3.2 s), 0.26 (corresponding time stamp is 5.4 s) and 0.27 (corresponding time stamp is 6.9 s) after filtering according to the peak value threshold value, four gesture fragment data with the duration of about 1s are obtained by intercepting from gesture motion data with the time stamp of 1.7s, 3.2s, 5.4s and 6.9s as the center respectively, and correspondingly, after marking each gesture fragment data, a gesture data set containing four marked gesture data sets is obtained. And further can be used for training a gesture recognition model.
Optionally, in the embodiment of the present application, after the gesture fragment data is obtained, gesture prediction reasoning may be performed by using a local lightweight model preferentially to obtain an original labeling result of the gesture fragment data, and then manually verified labeling data is obtained by combining with a manual confirmation mode, that is, optionally, in some embodiments of the present application, the step of labeling each gesture fragment data to obtain a target gesture data set includes:
Labeling each gesture fragment data through a pre-deployment model to obtain an original labeling result of each gesture fragment data;
Displaying the original labeling result through a virtual display screen;
obtaining a preliminary verification result in response to receiving first modification information input for the original labeling result;
the gesture fragment data, the gesture video data and the preliminary verification result are displayed in an aligned mode, and the gesture video data are synchronously shot aiming at the gesture motion data;
responding to the received second modification information input for the preliminary verification result, and obtaining a target verification result;
and generating the target gesture data set according to the target verification result.
It should be noted that the pre-deployment model includes a local lightweight model, including a pre-deployed gesture recognition model, where the pre-deployment model may be the same or different from a gesture recognition model to be trained by the target gesture data set. For example, when the pre-deployment model is different from the gesture recognition model to be trained, reasoning is achieved by using other models, the marking of gesture data is quickened by combining a manual confirmation mode, and further after the marked gesture data is obtained through manual calibration, the gesture recognition model to be trained can be trained. For another example, the pre-deployment model, that is, the gesture recognition model to be trained, performs preliminary reasoning based on the general reasoning capability of the pre-deployment model, and performs labeling in combination with manual, so that the workload of manual labeling can be reduced as well, and further, after labeled gesture data is obtained after labeling, the pre-deployment model can be trained so as to improve the reasoning capability of the pre-deployment model.
It may be understood that in the embodiment of the present application, the labeling result may be calibrated by combining the video data of the gesture, for example, when the hand performs the gesture to generate the gesture motion data, the gesture video data is synchronously captured by using the camera, and accordingly, the gesture video data and the gesture fragment data may be displayed in alignment, so that the user calibrates the gesture fragment data and the labeling result corresponding to the gesture fragment data through the displayed gesture video data.
For example, the original labeling result labeled by the pre-deployment model is displayed through a virtual display screen of the augmented reality device, and modification confirmation of the original labeling result can be manually realized through gesture operation, a matched ring or other tools or based on eye movement data, so that a preliminary verification result is obtained. Furthermore, the gesture video data and the gesture fragment data can be displayed through the virtual display screen of the augmented reality device in an aligned mode, so that a user browses the gesture video data to recalibrate the gesture fragment data and the preliminary verification result corresponding to the gesture fragment data, and a more accurate target verification result is obtained.
It can be understood that the data generation method of the embodiment of the application is suitable for watches, mobile phones, computers, and augmented reality devices or data generation systems composed of watches, mobile phones, computers, or watches, augmented reality devices, and computers.
For example, taking a data generation system consisting of a watch, a mobile phone and a computer as an example, the mobile phone is connected with the watch through the BLE, a user issues an instruction for starting to collect data through the mobile phone, the mobile phone controls the watch through the BLE to collect gesture motion data of the hand of the user based on an IMU sensor, the mobile phone also shoots simultaneously to obtain gesture video data of the hand motion, the watch can directly obtain local peaks of all time windows through peak detection, further obtain target peaks through peak stacks, obtain a plurality of gesture fragment data from the gesture motion data according to all the target peaks, then the watch invokes a pre-deployment model to identify original labeling results corresponding to all the gesture fragment data, and the watch sends the gesture fragment data, the gesture motion data and the original labeling results to the mobile phone.
After the mobile phone receives the gesture fragment data, the gesture motion data and the original labeling result sent by the watch, the original labeling result corresponding to the gesture fragment data is displayed in the mobile phone interface, and a user can confirm or modify the original labeling result through the mobile phone interface to obtain a primary verification result confirmed by a person for the first time.
Furthermore, the mobile phone can send the gesture fragment data, the gesture motion data, the original labeling result and the gesture video data to a computer (wherein the computer can also upload the cloud end first and download the gesture video data from the cloud end), the computer can display the gesture video data and the gesture fragment data in an aligned manner through an ELAN tool, and a user can check the initial checking result of the gesture fragment data again through the ELAN tool and by utilizing the gesture video data to obtain a target checking result. And then, obtaining the marked target gesture data set based on the target verification result.
In summary, the embodiment of the application extracts the target peak value from the gesture motion data, and extracts the gesture fragment data from the gesture motion data based on the target peak value, so that the validity and the accuracy of the gesture fragment data are improved, and further, after the target gesture data set is obtained by labeling based on the gesture fragment data, the accuracy of the target gesture data set is also improved.
And, by screening the target peak value based on the monotonic peak value stack, the screening efficiency of the target peak value is improved.
And extracting gesture fragment data by taking the target peak value as the center according to the preset time length, so that the extracted gesture fragment data are more effective and complete.
In order to facilitate better implementation of the data generation method, the application also provides a data generation device based on the data generation method. Wherein the meaning of nouns is the same as in the data generation method described above, specific implementation details may be referred to the description in the method embodiment.
Referring to fig. 3, fig. 3 is a schematic structural diagram of a data generating device according to an embodiment of the present application, where the data generating device may specifically be as follows:
an acquiring module 301, configured to acquire gesture motion data;
An identification module 302, configured to identify at least one target peak from the gesture motion data;
an extracting module 303, configured to extract at least one gesture segment data from the gesture motion data according to each target peak value;
and the labeling module 304 is configured to label each gesture segment data to obtain a target gesture data set.
Optionally, in some embodiments of the present application, the identifying at least one target peak from the gesture motion data includes:
Identifying a local peak of the gesture motion data over at least one time window;
and screening at least one target peak value from the local peaks according to a peak value threshold.
Optionally, in some embodiments of the application, the gesture motion data includes acceleration data;
The identifying the local peak of the gesture motion data over at least one time window comprises:
For each time window, deriving the analog length data corresponding to the acceleration data in the time window to obtain a derivative data set, wherein the derivative data set comprises derivative data corresponding to the analog length data of at least one time stamp in the time window;
taking the derivative data which is larger than zero in the derivative data set as candidate peak values;
and screening the maximum value from the candidate peaks to obtain the local peak value.
Optionally, in some embodiments of the present application, the selecting a maximum value from the candidate peaks to obtain the local peak includes:
comparing adjacent candidate peaks according to the sequence of the time stamps, and comparing the larger candidate peak with the next candidate peak until the comparison of the last candidate peak in the time window is finished, so as to obtain the largest candidate peak;
the largest candidate peak is set as the local peak.
Optionally, in some embodiments of the present application, the selecting a maximum value from the candidate peaks to obtain the local peak includes:
sequentially adding each candidate peak to a peak stack in the order of time stamps, and
If the current candidate peak value is larger than the previous candidate peak value, removing the previous candidate peak value from the peak value stack, and reserving the current candidate peak value;
If the current candidate peak is smaller than the previous candidate peak, retaining the current candidate peak and the previous candidate peak, and if the next candidate peak is larger than the previous candidate peak and the next candidate peak is larger than the current candidate peak, removing the current candidate peak and the previous candidate peak, and retaining the next candidate peak;
and after the last candidate peak value is added to the peak value stack in the time window and the sizes are compared, setting the candidate peak value reserved in the peak value stack or the candidate peak value at the stack top of the peak value stack as the local peak value, wherein the sizes of the candidate peak values of the peak value stack from the stack top to the stack tail are sequentially decreased.
Optionally, in some embodiments of the present application, the extracting at least one gesture segment data from the gesture motion data according to each target peak value includes:
And extracting gesture fragment data which takes the target peak value as a center and has a preset duration from the gesture motion data aiming at each target peak value.
Optionally, in some embodiments of the application, the method is applied to an augmented reality device;
labeling each gesture fragment data to obtain a target gesture data set, including:
Labeling each gesture fragment data through a pre-deployment model to obtain an original labeling result of each gesture fragment data;
Displaying the original labeling result through a virtual display screen;
obtaining a preliminary verification result in response to receiving first modification information input for the original labeling result;
the gesture fragment data, the gesture video data and the preliminary verification result are displayed in an aligned mode, and the gesture video data are synchronously shot aiming at the gesture motion data;
responding to the received second modification information input for the preliminary verification result, and obtaining a target verification result;
and generating the target gesture data set according to the target verification result.
In the embodiment of the application, gesture motion data is firstly acquired by an acquisition module 301, at least one target peak value is identified from the gesture motion data by an identification module 302, at least one gesture fragment data is extracted from the gesture motion data by an extraction module 303 according to each target peak value, and each gesture fragment data is marked by a marking module 304, so that a target gesture data set is obtained.
According to the embodiment of the application, the target peak value is extracted from the gesture motion data, the gesture fragment data is extracted from the gesture motion data based on the target peak value, the effectiveness and the accuracy of the gesture fragment data are improved, and the accuracy of the target gesture data set is improved after the target gesture data set is obtained by labeling based on the gesture fragment data.
In addition, the application further provides an electronic device, as shown in fig. 4, which shows a schematic structural diagram of the electronic device provided by the embodiment of the application, specifically:
The electronic device may include one or more processing cores 'processors 401, one or more computer-readable storage media's memory 402, power supply 403, and input unit 404, among other components. Those skilled in the art will appreciate that the electronic device structure shown in fig. 4 is not limiting of the electronic device and may include more or fewer components than shown, or may combine certain components, or may be arranged in different components. Wherein:
The processor 401 is a control center of the electronic device, connects various parts of the entire electronic device using various interfaces and lines, and performs various functions of the electronic device and processes data by running or executing software programs and/or modules stored in the memory 402, and calling data stored in the memory 402, thereby performing overall monitoring of the electronic device. Optionally, the processor 401 may include one or more processing cores, and preferably the processor 401 may integrate an application processor and a modem processor, wherein the application processor mainly processes an operating system, a user interface, an application program, etc., and the modem processor mainly processes wireless communication. It will be appreciated that the modem processor described above may not be integrated into the processor 401.
The memory 402 may be used to store software programs and modules, and the processor 401 executes various functional applications and data processing by executing the software programs and modules stored in the memory 402. The memory 402 may mainly include a storage program area that may store an operating system, application programs required for at least one function (such as a sound playing function, an image playing function, etc.), etc., and a storage data area that may store data created according to the use of the electronic device, etc. In addition, memory 402 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device. Accordingly, the memory 402 may also include a memory controller to provide the processor 401 with access to the memory 402.
The electronic device further comprises a power supply 403 for supplying power to the various components, preferably the power supply 403 may be logically connected to the processor 401 by a power management system, so that functions of managing charging, discharging, and power consumption are performed by the power management system. The power supply 403 may also include one or more of any of a direct current or alternating current power supply, a recharging system, a power device commissioning circuit, a power converter or inverter, a power status indicator, etc.
The electronic device may further comprise an input unit 404, which input unit 404 may be used for receiving input digital or character information and generating keyboard, mouse, joystick, optical or trackball signal inputs in connection with user settings and function control.
Although not shown, the electronic device may further include a display unit or the like, which is not described herein. In this embodiment, the processor 401 in the electronic device loads executable files corresponding to the processes of one or more application programs into the memory 402 according to the following instructions, and the processor 401 executes the application programs stored in the memory 402, so as to implement the steps in any data generating method provided in the embodiment of the present application.
According to the embodiment of the application, gesture motion data are acquired, at least one target peak value is identified from the gesture motion data, at least one gesture fragment data is extracted from the gesture motion data according to each target peak value, each gesture fragment data is marked, and a target gesture data set is obtained.
The method comprises the steps of extracting a target peak value from gesture motion data, extracting gesture fragment data from the gesture motion data based on the target peak value, and improving the effectiveness and accuracy of the gesture fragment data, so that after a target gesture data set is obtained by labeling based on the gesture fragment data, the accuracy of the target gesture data set is improved.
The specific implementation of each operation above may be referred to the previous embodiments, and will not be described herein.
Those of ordinary skill in the art will appreciate that all or a portion of the steps of the various methods of the above embodiments may be performed by instructions, or by instructions controlling associated hardware, which may be stored in a computer-readable storage medium and loaded and executed by a processor.
To this end, the present application provides a computer readable storage medium having stored thereon a computer program that can be loaded by a processor to perform the steps of any of the data generation methods provided by the present application.
The specific implementation of each operation above may be referred to the previous embodiments, and will not be described herein.
The computer readable storage medium may include, among others, read Only Memory (ROM), random access Memory (RAM, random Access Memory), magnetic or optical disks, and the like.
Because the instructions stored in the computer readable storage medium can execute the steps in any data generating method provided by the present application, the beneficial effects that any data generating method provided by the present application can achieve can be achieved, and detailed descriptions are omitted herein.
The foregoing has outlined rather broadly the principles and embodiments of the present application in order that the detailed description of the application may be better understood, and in order that the present application may be better understood, the present application should not be construed as being limited to the details of the embodiments and applications of the present application.