[go: up one dir, main page]

CN120029466B - Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking - Google Patents

Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking

Info

Publication number
CN120029466B
CN120029466B CN202510504620.7A CN202510504620A CN120029466B CN 120029466 B CN120029466 B CN 120029466B CN 202510504620 A CN202510504620 A CN 202510504620A CN 120029466 B CN120029466 B CN 120029466B
Authority
CN
China
Prior art keywords
image
screen
module
image plane
lens
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202510504620.7A
Other languages
Chinese (zh)
Other versions
CN120029466A (en
Inventor
乐孜纯
张晓源
沈晨航
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University of Technology ZJUT
Original Assignee
Zhejiang University of Technology ZJUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University of Technology ZJUT filed Critical Zhejiang University of Technology ZJUT
Priority to CN202510504620.7A priority Critical patent/CN120029466B/en
Publication of CN120029466A publication Critical patent/CN120029466A/en
Application granted granted Critical
Publication of CN120029466B publication Critical patent/CN120029466B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0093Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/28Quantising the image, e.g. histogram thresholding for discrimination between background and foreground patterns
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/34Smoothing or thinning of the pattern; Morphological operations; Skeletonisation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0138Head-up displays characterised by optical features comprising image capture systems, e.g. camera
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/014Head-up displays characterised by optical features comprising information/image processing systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Optics & Photonics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Ophthalmology & Optometry (AREA)
  • Processing Or Creating Images (AREA)

Abstract

一种基于多像面眼动追踪的动态虚拟像面调节装置,属于视觉训练技术领域,多像面眼动追踪系统利用红外摄像机捕捉用户眼部图像,处理得到精准的注视点坐标,并将数据传输至控制系统;控制系统接收注视点坐标,解析当前所需的像面深度,并生成透镜调整指令,发送至透镜‑屏幕调节系统;透镜‑屏幕调节系统驱动透镜和屏幕进行物理调整,同时利用传感反馈机制实时监测调整精度,确保虚像位置与用户注视焦点匹配;显示系统结合控制系统提供的深度信息,动态调整动画画面,使其与用户的注视深度同步变化。以及提供一种基于多像面眼动追踪的动态虚拟像面调节方法。本发明实现用户视觉焦点的实时追踪与动态调节,满足多深度场景下的精准视觉交互需求。

A dynamic virtual image plane adjustment device based on multi-image plane eye tracking belongs to the field of visual training technology. The multi-image plane eye tracking system uses an infrared camera to capture the user's eye image, processes it to obtain accurate gaze point coordinates, and transmits the data to the control system; the control system receives the gaze point coordinates, analyzes the current required image plane depth, and generates lens adjustment instructions, which are sent to the lens-screen adjustment system; the lens-screen adjustment system drives the lens and screen to make physical adjustments, and uses a sensor feedback mechanism to monitor the adjustment accuracy in real time to ensure that the virtual image position matches the user's gaze focus; the display system dynamically adjusts the animation screen in combination with the depth information provided by the control system, so that it changes synchronously with the user's gaze depth. And a dynamic virtual image plane adjustment method based on multi-image plane eye tracking is provided. The present invention realizes real-time tracking and dynamic adjustment of the user's visual focus, meeting the needs of accurate visual interaction in multi-depth scenes.

Description

Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking
Technical Field
The invention belongs to the technical field of vision training, and particularly provides a dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking.
Background
In recent years, virtual Reality (VR) and Augmented Reality (AR) technologies have been rapidly developed, and have been widely used in various fields such as medical treatment, education, entertainment, and the like. However, the technical limitations of the existing VR/AR devices in display form remain significant, especially in terms of focal plane adjustment, which presents a number of bottlenecks. Most VR/AR devices in the current market adopt a fixed focal plane design, so that the image planes watched by users in the virtual environment are always at the same depth, and the positions of the image planes cannot be dynamically adjusted according to the change of the watching point of the users. The design of the fixed focal length not only limits the authenticity of the immersive experience, but also can cause visual fatigue after long-term use, affecting the visual health of the user.
Some devices have a zooming function, but the adjustment mode often depends on manual operation of a user, so that the process is complicated, and real-time and intelligent focal length adjustment is difficult to realize. For example, some zoom devices require a user to adjust the lens position through a mechanical knob or button, while a few devices with electronic focusing function rely on a preset focusing mode, which not only lacks an active feedback mechanism, but also does not have a dynamic adjusting function, so that it is difficult to adapt to the individual requirements of the user in different scenes.
The current eye tracking technology is mainly used in the fields of sight line position analysis, interactive control and the like, and is not applied to focal length dynamic adjustment. The mainstream eye tracking method comprises a pupil positioning technology based on an infrared camera and a sight prediction method based on machine learning, and the pupil positioning technology and the sight prediction method have greatly progressed in accuracy and real-time. However, these techniques have not been effectively combined and applied to optimize the visual experience of virtual reality devices, especially intelligent applications in focus adjustment, remain in the infancy.
On the other hand, the multi-image-plane display technology is a method for alleviating visual fatigue, and the core principle of the multi-image-plane display technology is that a plurality of virtual image planes are generated at different depths, so that a user can obtain visual experience which is closer to reality when watching virtual objects at different depths. However, the implementation of this technology often relies on complex optical designs such as multi-layer displays, optical layering devices, or light field based dynamic reconstruction systems. The application of the schemes in the existing equipment is limited mainly due to the fact that the hardware structure is complex, the calculation cost is high, and meanwhile, a large optimization space still exists in the aspects of synchronism with the change of the user's sight and coordination with the dynamic adjustment of the image plane.
For the currently existing related patents, the analysis is as follows:
Compared with the patent technology CN209014753U (variable focus lens and VR equipment), the variable focus lens has the advantages that the refractive index is adjusted by adopting the thickness change of liquid in the elastic optical surface cavity, so that the zoom function is realized. However, the zooming mode still depends on passive manual adjustment, a fixed focal length is required to be preset and focusing is performed through an analog circuit, and active feedback based on real-time gazing information of a user is absent.
Compared with the patent technology CN208823365U (variable-focus VR eye vision instrument), the variable-focus VR eye vision instrument adopts a motor driving mode to adjust the distance between two lenses so as to change the optical path between the virtual image plane and the human eye. Although the method can realize the adjustment of near vision and far vision to a certain extent, manual operation or adjustment depending on fixed time intervals is needed, and the self-adaptive adjustment on the real-time fixation requirement of the user is lacking.
Compared with the patent technology CN114895793A (an active self-adaptive eye tracking method and AR glasses), the method utilizes eye tracking to capture the point of gaze of a user, and combines a TOF (Time-of-Flight) sensor to detect the distance of the actual observation point of the user so as to adjust the imaging position and focusing distance of an image module and avoid visual discomfort caused by frequent change of a virtual scene. However, this approach focuses mainly on the adjustment of the image projection position, and does not optimize the focus adjustment in combination with the multi-image plane display technique.
Most of the devices on the market today use a fixed focal plane display design, which is prone to visual fatigue and discomfort when used for a long period of time by the user. The main reason for this is that existing VR/AR devices typically only support a single fixed focal plane, and cannot dynamically adjust the position of the image plane according to the depth of the user's gaze point, or manually focus by the user. Therefore, the image plane of the user looking at in the virtual scene is always at the same depth, the natural focusing requirement of the sight cannot be met, the lens can be kept in a fixed state continuously after long-time use, the visual burden is increased, the use experience of the user cannot be met, and accordingly visual fatigue is caused and adverse effects are caused on eye health.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention provides a dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking, which realize real-time tracking and dynamic adjustment of a user visual focus through multi-system cooperation and meet the accurate visual interaction requirement under a multi-depth scene, thereby dynamically adjusting the image plane position according to the change of a user fixation point.
The technical scheme adopted for solving the technical problems is as follows:
A dynamic virtual image plane regulating device based on multi-image plane eye tracking comprises a multi-image plane eye tracking system, a lens-screen regulating system, a control system and a display system,
The multi-image-plane eye tracking system captures eye images of a user by using an infrared camera, calculates through an image acquisition module, a pupil detection module, an ellipse fitting module, a calibration module and a polynomial fitting module, converts the eye images to obtain accurate gaze point coordinates, and transmits data to a control system;
The control system receives the gaze point information, analyzes the current required image plane depth, generates a lens adjustment instruction and sends the lens adjustment instruction to the lens-screen adjustment system;
The lens-screen adjusting system drives the lens and the screen to carry out physical adjustment according to the control instruction, and simultaneously monitors adjustment accuracy in real time by utilizing a sensing feedback mechanism to ensure that the virtual image position is matched with the user gazing focus;
The display system combines the depth information provided by the control system to dynamically adjust the animation picture to enable the animation picture to synchronously change with the gazing depth of the user, so that visual immersion is enhanced.
Further, in the multi-image-plane eye tracking system, the image acquisition module acquires pupil images of eyes of a user by calling the active infrared camera, the pupil detection module is used for calibrating pupil outlines, the ellipse fitting module is used for detecting pupil center positions with highest confidence degrees, the calibration module is used for carrying out coordinate conversion by combining known reference points, and the polynomial fitting module is used for carrying out optimization processing on data and finally converting pupil data into gaze point coordinates.
Preferably, the implementation flow of the image acquisition module is as follows:
Step 111, selecting a suitable 940nm short-focus camera and an infrared lamp bead;
112, connecting the 940nm short-focus camera with the infrared lamp beads, so as to facilitate subsequent collection;
and 113, adjusting the positions of the 940nm short-focus camera and the infrared lamp beads to ensure that the eyes are completely illuminated and avoid pupil detection failure.
Still preferably, the implementation flow of the pupil detection module is as follows:
step 121, judging whether pupil detection fails according to a pupil threshold algorithm, if so, repeating step 113 to ensure success of pupil detection;
Step 122, converting the input video into video frames, and processing frame by frame:
The input original image is subjected to cropping and scaling operations to unify the image size and facilitate subsequent processing. Calculating a New width new_width=int (Desire _ratio) by calculating the current_ratio (current_width/current_height) of the Current image and the target width ratio Desire _ratio (preset as width/height, for example, width=580, height=480), if the current_ratio > Desire _ratio indicates that the image is too wide, calculating a New width new_width=int (Desire _ratio) by calculating the current_ratio, clipping the current_width from the position of the center offset loc_w= (current_width-new_width)// 2 in the image width direction, and obtaining a clipped image size cut_image if current_ratio is excessively high, namely, calculating a New height new_height=int (current_width/Desire _ratio), clipping the image from the center offset loc_w= (current_width-new_width)// 2, and finally obtaining the clipping the image size of the current_image, [ current_image ]: desire _ratio ];
Step 123, storing A, B, C binary images with similar three different areas in the image:
converting the cropped and scaled image into a gray image, applying a strict binarization threshold to the gray image by adding a threshold increment to the darkest pixel value Darkest _value Calculating to obtain a Threshold value Threshold= Darkest _value +Binarizing the image using a cv2.Threshold function, setting pixels smaller than the threshold to white (255) and black (0) to obtain a strictly thresholded image Thresholded _image_1, and centering around the darkest point Darkest _point with a side lengthCreating a square mask, setting the pixel value outside the square area asPerforming mask processing on Thresholded _image_1;
Adding the darkest pixel value Darkest _value to the threshold increment Calculating to obtain a Threshold value Threshold= Darkest _value +Binarizing the gray level image to obtain Thresholded _image_2, and taking the darkest point as the center and the side lengthPerforming mask treatment;
Adding the darkest pixel value Darkest _value to the threshold increment Calculating to obtain a Threshold value Threshold= Darkest _value +Binarizing the gray image to obtain Thresholded _image_3, and centering on the darkest pointPerforming mask treatment;
And Thresholded _image_1 Thresholded _image_2 Thresholded _image_3;
step 124, detecting A, B, C color threshold ranges of three different areas;
at step 125, the pupil image with less fluctuation of the average color threshold is selected as the most reliable pupil image.
More preferably, the implementation flow of the ellipse fitting module is as follows:
Step 131, performing dilation processing on each binary image to enhance the target area:
For the image after multi-thresholding and masking, morphological dilation operation (such as cv2.Dilate function) is first used to enhance the contour features in the image, then contours in the image are found by cv2. Final contours function, then the found contours are filtered, all contours are traversed, the area of each contour is calculated as area=cv2. Final area (contour), if area > =pixel (e.g. pixel=10), the width w and height h of the bounding rectangle of the contour are further calculated (by cv2. Final features function), and the aspect ratio length=max (w, h), width=min (w, h), current_ratio=max (length/width, width/length) are calculated. If current_ratio < = ratio_thresh (e.g. ratio_thresh=5), the contour is eligible, returning to the eligible contour with the largest area;
step 132, extracting an external contour in the expanded image;
step 133, screening the profile through area and number limitation, and returning to the maximum profile;
Step 134, not searching within D pixels of the image edge;
Step 135, checking the brightness every E pixels in the area;
Step 136, edge Shaft and method for producing the sameThe axis samples once every F pixels, ignoring the boundary;
step 137, updating the pixel point with the maximum threshold value between 0 and 255 color channels;
More preferably, the implementation flow of the calibration module is as follows:
step 141, selecting the screen size as a calibration screen, inputting screen parameters into a control system and projecting the screen parameters into human eyes through a lens;
step 142, selecting 9-12 points uniformly paved on the screen as standard points;
step 143, the user wears the device and adjusts the size of the eye image in the picture;
Step 144, performing autonomous calibration through a peripheral device such as a keyboard or a handle, to ensure that the point at which the user gazes in the current state is a calibration point:
In the calibration mode, images are acquired in real time through a camera, the center coordinates (x_eye, y_eye) of pupils are obtained through the pupil detection algorithm, a currently calibrated target point (displayed in red) is drawn on each frame of image, and when a user presses a space key, the current pupil coordinates (x_eye, y_eye) and corresponding screen coordinates (namely, the screen coordinates (x_screen, y_screen) of the currently calibrated target point) are recorded into a calibration data dictionary calibration_data. After the recording of all calibration points is completed, storing the calibration_data into a file (such as a calibration_data. Json);
step 145, adjusting the distance between the lens and the screen, thereby changing the distance between the image plane and the human eye, and repeating steps 142 to 144;
And step 146, saving the weight information marked by the different depth indexes.
Preferably, the implementation flow of the polynomial fitting module is as follows:
step 151, obtaining pixel point coordinates of 9-12 calibration points;
step 152, outputting the coordinates of the pixel point at the pupil center from the video frame;
step 153, coordinates of the pixel points in step 151 and step 152 And (3) withSplit intoOne-dimensional data;
Step 154, selecting a polynomial to perform one-dimensional data fitting on coordinate points;
Step 155, fitting Polynomial coefficients in the function;
Step 156, fitting Polynomial coefficients in the function;
step 157, save the fitted Shaft and method for producing the sameCoefficients of one-dimensional polynomial of axes:
Pupil coordinate data eye_coordinates and screen coordinate data screen_coordinates are loaded from the stored calibration data file. Extracting x-component eye_x and y-component eye_y of the pupil coordinate, and performing polynomial fitting on the pupil x-coordinate and the screen x-coordinate and the pupil y-coordinate and the screen y-coordinate respectively by using functions to obtain fitting coefficients coefficis_x and coefficis_y;
Step 158, starting to predict the gaze point coordinates by inputting pupil center point coordinates;
Still further, the lens-screen adjusting system comprises a lens, a screen, a linear slide rail, a micro stepping motor and a driving circuit, wherein the driving circuit is connected with the micro stepping motor, the action end of the micro stepping motor is linked with the screen, the screen is slidably arranged on the linear slide rail, the screen and the lens are respectively arranged on the front side and the rear side of the linear slide rail, all parts of the system are highly integrated, when a fixation point enters a target area, the driving circuit sends a control signal to the micro stepping motor, and the micro driving motor drives the screen to perform stable and accurate linear displacement along the linear slide rail, so that the geometric distance between the lens and the screen is adjusted. In this scheme, the introduction of miniature stepper motor has realized the lightweight design of structure, possesses accurate feedback adjustment ability simultaneously, and after receiving the data from many image plane eye tracking system, the system can carry out dynamic adjustment through miniature stepper motor drive lens-screen unit to ensure the real-time matching of user's visual focus and image plane degree of depth.
The lens-screen adjusting system further comprises an adjusting module and a sensing feedback module, wherein the adjusting module is used for selecting the distance to be adjusted in real time according to the predicted gaze point coordinates, and the sensing feedback module is used for driving the micro stepping motor to adjust the distance between the lens and the screen.
Preferably, the implementation flow of the adjusting module is as follows:
Step 211, selecting a picture area according to the predicted gaze point coordinates output in step 158;
in step 212, the previously set travel range is adjusted when this area is marked.
Still preferably, the implementation flow of the sensing feedback module is as follows:
step 221, selecting a miniature stepping motor with a stroke of 8 mm;
step 222, applying PWM waves to control the stroke of the micro stepping motor according to different gazing areas.
The control system comprises a data processing module, a calculation module and a control instruction module, wherein the data processing module is used for receiving the gaze point coordinates from the eye tracking system, the calculation module is used for analyzing the current required image plane depth, and the control instruction module is used for sending a motion instruction to the lens-screen adjusting system.
Preferably, the implementation flow of the data processing module is as follows:
step 311, receiving pupil center coordinates output by an eye tracking system;
step 312, receiving and analyzing the calibration fixation point coordinates;
step 313, performing coordinate correction and filtering based on the pupil center coordinates and the calibration gaze point coordinates, removing noise, and improving data precision;
step 314, storing the corrected gaze point coordinates and transmitting the data to a computing module;
Still preferably, the implementation flow of the computing module is as follows:
Step 321, receiving gaze point coordinate information from a data processing module;
Step 322, calculating the current required image plane depth based on a gaze point-depth mapping model preset by the system;
Step 323, detecting abnormal value of the calculation result, and performing interpolation processing to optimize data smoothness;
Step 324, storing the calculated depth information and sending it to the control instruction module;
More preferably, the implementation flow of the control instruction module is as follows:
Step 331, receiving image plane depth data transmitted by a computing module;
step 332, generating control instructions of the stepper motor according to the depth requirement, including forward, backward and stroke length;
step 333, controlling the moving step length and speed of the micro stepping motor by using a PWM (pulse width modulation) signal;
step 334, monitoring the state of the stepping motor in real time to ensure accurate adjustment of the lens position;
step 335, feeding back the execution result, and performing secondary adjustment if necessary to improve the system response accuracy.
The display system comprises an animation preprocessing module and a fixation point sensing module, wherein the animation preprocessing module is used for setting virtual image depths corresponding to different areas in a picture in advance, and the fixation point sensing module is used for ensuring that picture updating and lens adjustment are synchronously carried out.
Preferably, the implementation flow of the animation preprocessing module is as follows:
step 411, setting depth information in the picture according to the depth information in real life;
step 412, presetting depth information for different regions of each frame of picture in the animation;
step 413, presetting different depth information for the pictures at different depths;
Step 414, setting preset depth information as a feedback terminal;
Preferably, the implementation flow of the gaze point sensing module is as follows:
Step 421, setting a region threshold range;
step 422, presetting different area threshold ranges for the pictures at different depths;
step 423, when the gaze point is transferred to the region threshold, feedback information is generated, and the micro stepping motor of step 222 is driven to realize feeding.
A dynamic virtual image plane adjusting method based on multi-image plane eye tracking comprises the following steps:
Step 1, wearing equipment by a user, namely, correctly wearing eye tracking equipment by the user, ensuring that the equipment stably fits the face, adjusting the position, enabling a camera to clearly capture the eye area of the user;
step 2, eye movement calibration, namely entering a calibration mode, guiding a user to sequentially watch a plurality of preset calibration points to establish a user eye movement model, and precisely fitting personalized eye movement parameters of the user by calculating the mapping relation between the pupil center offset and screen coordinates;
step 3, gaze point prediction, namely continuously collecting eye movement data of a user, including pupil positions and sight line directions, under a normal operation mode, calculating and predicting current gaze point coordinates of the user in real time based on an established eye movement model, and determining a specific target area of the sight line of the user on a screen, wherein the prediction data is used for driving a follow-up self-adaptive vision adjustment mechanism;
Detecting micro movement of eyeballs of the user, and dynamically adjusting calculation accuracy of the gaze point so as to cope with eye movement characteristics of different users;
And 5, adjusting the lens-screen distance, namely when confirming that the gaze point of the user is positioned in the threshold range of the area set by the gaze point sensing module, generating a driving signal by the driving circuit, transmitting the driving signal to the micro stepping motor, and accurately moving the screen along the linear sliding rail by the micro stepping motor to change the geometric distance between the screen and the lens so as to adjust the visual focal length, thereby matching the current vision requirement of the user.
The invention provides an intelligent real-time response virtual image plane adjusting device by combining an eye movement tracking technology with a multi-image plane dynamic adjusting scheme. The device can dynamically adjust the lens position according to the depth information of the user's gaze point, so that the virtual image plane can realize intelligent focusing on different depths, thereby effectively relieving visual fatigue, improving immersive experience, and providing a more advanced solution for focal length adjustment of future VR/AR equipment. Compared with the prior art CN208823365U, the invention combines the eye movement tracking with the multi-image-plane dynamic adjustment technology, so that the focal length adjustment process is more natural and more intelligent, and visual fatigue can be effectively relieved, and compared with the prior art CN114895793A, the invention further provides a virtual image plane dynamic adjustment scheme combining the multi-image-plane eye movement tracking technology on the basis of the prior art, so that the system can dynamically adjust the relative positions of the lens and the screen according to the depth information of the user's gaze point, and more accurate focal length adjustment is realized.
The invention senses the sight line position of the user in real time through the multi-image-plane eye tracking module, combines the lens-screen adjusting module to dynamically adjust the light path length, and cooperates with the efficient data processing of the control system module and the depth image presentation of the display system module to achieve the effect of dynamically adjusting the depth of the virtual image plane, thereby providing more comfortable and natural visual experience for the user.
The beneficial effects of the invention are mainly shown in the following steps:
1) The invention discloses a dynamic virtual image plane regulating device and a dynamic virtual image plane regulating method based on multi-image plane eye movement tracking, which are integrated with eye movement tracking according to a regulating principle of a crystalline lens, skillfully and efficiently combine optics and ophthalmology according to ophthalmic medical requirements and optical basic knowledge.
2) Multi-image-plane eye tracking, namely, multi-image-plane eye tracking can meet the requirement that a user can realize eye tracking functions under different image plane depths, and compared with the traditional VR/AR equipment which usually only performs eye tracking on a single image plane, the multi-image-plane eye tracking cannot be realized.
3) Setting the depth of the animation, namely setting the depth information corresponding to different positions in each frame of animation in advance, and realizing a feedback mechanism by combining the depth information with a multi-image-plane eye tracking system to help a user to realize a more real interaction scene.
4) And the active feedback is realized by judging the position of the eye point and changing the physical distance between the lens and the screen in real time so as to change the distance between the virtual image surface and the human eye without manual adjustment.
Drawings
FIG. 1 is a schematic block diagram of a dynamic virtual image plane adjustment device based on multi-image plane eye tracking;
FIG. 2 is a flow chart of a dynamic virtual image plane adjustment method based on multi-image plane eye tracking;
FIG. 3 is a schematic diagram of the optical path of a multi-image-plane eye tracking system;
FIG. 4 is a schematic diagram of a preset depth of the display system;
FIG. 5 is a schematic diagram of a front view of a dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to the present invention;
fig. 6 is a schematic diagram of a back structure of a dynamic virtual image plane adjustment device based on multi-image plane eye tracking in the present invention.
Wherein 101 is a screen, 102 is a lens 102,103 is a human eye, 104 is an active infrared camera, 105 is a baffle, 106 is a miniature stepper motor, 201 is a near-looking cylindrical obstacle, 202 is a far-looking cylindrical obstacle, 203 is the sun, 204 is a cloud, 301 is a shell, 302 is a linear slide rail, 303 is a lens frame, 304 is a clamping groove, 305 is an active infrared light source, and 401 is a screen frame.
Detailed Description
The invention is further described below with reference to the accompanying drawings.
Referring to fig. 1-6, a dynamic virtual image plane adjusting device combining a multi-image plane eye tracking technology detects the gaze point position of a user in real time by presetting depth information of different positions in a display animation, dynamically adjusts the distance between a lens and a screen, changes the position of the virtual image plane, realizes active feedback, does not need manual adjustment of the user, and therefore relieves visual fatigue and improves user experience.
Fig. 1 is a schematic block diagram of a dynamic virtual image plane adjusting device based on multi-image plane eye tracking in the present invention, and as shown in fig. 1, the dynamic virtual image plane adjusting device based on multi-image plane eye tracking includes a multi-image plane eye tracking system, a control system, a lens-screen adjusting system, and a display system.
The optical path of the multi-image-plane eye tracking system is shown in fig. 3, light emitted by the screen 101 is emitted into the human eye 103 after passing through the lens 102, and eye images are collected by the active infrared camera 104 to perform multi-image-plane eye tracking, wherein the partition plate 105 is used for ensuring that eyes are split, and the distance between the screen and the lens is changed through the micro stepping motor 106 according to the predicted position of the fixation point, so that the distance between an image generated by the light emitted by the screen and the human eye is changed, and a real scene can be simulated more truly.
The multi-image-plane eye tracking system comprises an image acquisition module, a pupil detection module, an ellipse fitting module, a calibration module and a polynomial fitting module, wherein:
The image acquisition module is used for acquiring a high-resolution image of the eyes of a user so as to ensure the accuracy of subsequent processing, and the implementation flow is as follows:
step 111, selecting a suitable 940nm short-focus camera and an infrared light source as data acquisition equipment to enhance the contrast ratio of pupil detection;
112, correctly connecting the 940nm short-focus camera with an infrared light source to ensure the stable operation of the acquisition system;
Step 113, optimizing the positions of the camera and the infrared light source, so that the eye area is fully illuminated, pupil detection failure is avoided, and the quality of data acquisition is improved;
the pupil detection module is used for identifying a pupil area from the acquired image, extracting pupil edge information, and preliminarily determining the pupil center position, and the implementation flow is as follows:
Step 121, evaluating the validity of the current detection result based on a pupil threshold algorithm. If the detection fails, the camera and the light source are required to be adjusted (the environment adjustment step 113 is repeated) so as to ensure that a stable pupil image is obtained;
Step 122, converting the input video stream into independent video frames, and analyzing each frame by frame;
step 123, binarizing A, B, C different areas in the image, and storing the binary image to enhance pupil characteristics;
Step 124, detecting a color threshold range of A, B, C areas and analyzing the variation trend thereof;
Step 125, selecting a region with the minimum fluctuation of the color threshold value from the A, B, C region as the most reliable pupil image so as to improve the detection stability;
The ellipse fitting module is used for performing ellipse fitting on the pupil area based on the pupil detection result so as to optimize the positioning accuracy of the pupil center, and the implementation flow is as follows:
step 131, performing expansion processing on the binarized image to enhance the visibility of the target area;
step 132, extracting an external contour from the inflated image to obtain a candidate pupil region;
Step 133, screening out the maximum contour most conforming to the pupil shape based on the contour area and the number features;
step 134, performing no pupil search in the area within D pixels of the image edge to reduce noise interference;
Step 135, in the specific area, brightness detection is performed every E pixels to enhance the robustness of pupil detection;
136, sampling every F pixels along the X axis and the Y axis, and ignoring boundary noise to improve the calculation efficiency;
Step 137, updating the pixel point with the largest threshold change in the 0-255 color interval to improve the pupil detection precision;
The calibration module is used for establishing a mapping relation between the pupil center and the screen fixation point, ensuring the accuracy of eye movement tracking, and the implementation flow is as follows:
Step 141, selecting the screen size as a calibration screen, and inputting parameters thereof into a control system to ensure that the lens projection imaging meets the visual requirement of a user;
Step 142, uniformly arranging 9-12 calibration points on a screen to ensure uniform data distribution so as to improve calibration accuracy;
Step 143, after the user wears the device, the size of the eye image on the screen is adjusted to meet the calibration requirement;
Step 144, the user performs calibration through external equipment (keyboard or handle) to ensure that the current gaze point corresponds to the calibration point correctly;
step 145, adjusting the distance between the lens and the screen, thereby changing the distance between the image plane and the human eye, and repeating steps 142 to 144;
Step 146, recording calibration weight information under different depth conditions, and providing reference data for subsequent eye tracking;
the polynomial fitting module is responsible for fitting the mathematical relationship between the pupil center coordinates and the screen fixation point, and constructing a mathematical model for predicting the user fixation position, and the implementation flow is as follows:
Step 151, obtaining pixel point coordinates of 9-12 calibration points ;
Step 152, outputting the coordinates of the pixel point at the pupil center from the video frame;
Step 153, coordinates of the pixel points in step 151 and step 152And (3) withSplit intoOne-dimensional data;
step 154, selecting polynomial orders, fitting coordinate point data to construct an optimal mapping function;
step 155, training based on data The polynomial coefficient of (2) ensures the conversion accuracy in the X-axis direction;
step 156, calculate To ensure the mapping accuracy in the Y-axis direction;
step 157, save the fitted Shaft and method for producing the sameCoefficients of a one-dimensional polynomial of the axis;
Step 158, in the actual running process, inputting pupil center coordinates detected in real time, and predicting gaze point coordinates on a screen by using a fitting function to realize high-precision eye tracking;
The display system comprises an animation preprocessing module and a fixation point sensing module, fig. 4 is a schematic diagram of a preset depth image in the display system, wherein a scene on a road is simulated in the image, the scene comprises a near-looking cylindrical obstacle 201, a far-looking cylindrical obstacle 202, a sun 203 and a cloud 204, according to a geometric perspective relationship, a human eye spontaneously considers that the near-looking cylindrical obstacle 201 is closer than the far-looking cylindrical obstacle 202, and the sun 203 and the cloud 204 can be regarded as infinity, so that when a predicted fixation point reaches a certain position in 201-204, a signal is sent to a screen-lens adjusting system, and the distance between the two is changed through a micro stepping motor.
The display system comprises an animation preprocessing module and a fixation point sensing module, wherein:
The animation preprocessing module is used for setting virtual image depths corresponding to different areas in a picture in advance, and the implementation flow is as follows:
Step 211, setting a depth parameter in the virtual picture according to the real-world depth information;
Step 212, allocating depth information of different areas for each frame of picture in the animation;
step 213, setting corresponding depth parameters for different depth ranges;
step 214, transmitting the calculated depth information to a feedback terminal for dynamic adjustment in visual display;
The gaze point sensing module is used for ensuring that the picture update and the lens adjustment are synchronously carried out, and the implementation flow is as follows:
Step 221, setting a region threshold range of a screen for gaze detection;
Step 222, presetting different area thresholds according to different depth ranges;
step 223, when the gaze point of the user enters the set region range, triggering feedback information to drive the micro stepping motor to adjust the focal length so as to optimize the viewing experience.
According to the flow chart shown in fig. 2, the tester needs to wear the housing 301 shown in fig. 5 on the head correctly after starting and can see the screen 101 shown in fig. 6 through the lens 102, the screen 101 is mounted in the screen frame 401, the screen frame 401 is slidably mounted on one side of the linear slide rail 302, the lens 102 is fixedly mounted on the other side of the linear slide rail 302 through the lens frame 303 and is connected with the partition 105 through the clamping groove 304, during the process of looking at the screen 101, the partition 105 separates the eyes, so that the eyes can see the images seen by the two eyes relatively independently, the active infrared light source 305 in fig. 5 illuminates the pupils after the test adjustment is finished, and the pupil images are shot by the active infrared camera 104 for eye movement calibration, after the calibration is finished, the micro stepping motor 106 in fig. 6 drives the screen frame 401 to move, thereby changing the distance between the screen and the lens, during the distance change, the partition 105 and the lens frame 303 are connected with the partition 105 through the clamping groove 304, the eyes are still independent from each other, and the circulation is completed according to the flow chart shown after the end of the distance change.
Further, the micro stepping motor 106 drives the screen frame 401 to move, which is accomplished by a lens-screen adjusting system.
The lens-screen adjusting system comprises a lens 102, a screen 101, a linear slide rail 302, a micro stepping motor 106 and a driving circuit, wherein the driving circuit is connected with the micro stepping motor 106, the action end of the micro stepping motor 106 is linked with the screen 101, the screen 101 is slidably arranged on the linear slide rail 302, the front side and the rear side of the linear slide rail 302 are respectively provided with the screen 101 and the lens 102, all parts of the system are highly integrated, when a fixation point enters a target area, the driving circuit sends a control signal to the micro stepping motor 106, and the driving motor drives the screen 101 to perform stable and accurate linear displacement along the linear slide rail 302, so that the geometric distance between the lens 102 and the screen 101 is adjusted. In this scheme, the introduction of miniature stepper motor 106 has realized the lightweight design of structure, possesses accurate feedback adjustment ability simultaneously, and after receiving the data from many image plane eye tracking system, the system can carry out dynamic adjustment through miniature stepper motor drive lens-screen unit to ensure the real-time matching of user's visual focus and image plane degree of depth.
The lens-screen adjustment system includes an adjustment module and a sensory feedback module, wherein:
the adjusting module is used for selecting the distance to be adjusted in real time according to the predicted fixation point coordinates, and the implementation flow is as follows:
step 311, selecting a picture area according to the outputted predicted gaze point coordinates;
step 312, when the gaze point falls into the region, the system adjusts the focal length according to the preset travel range to match the corresponding visual requirement;
The sensing feedback module is used for driving the miniature stepping motor to adjust the distance between the lens and the screen, and the implementation flow is as follows:
step 321, adopting a high-precision miniature stepping motor with a stroke range of 8mm to ensure the accuracy and stability of adjustment;
step 322, according to the difference of the user's gazing areas, the system generates PWM (pulse width modulation) control signals to drive the micro stepping motor to execute corresponding displacement adjustment, so as to realize focal length optimization;
the circulation flow is realized by a control system, the control system comprises a data processing module, a calculation module and a control instruction module, wherein:
the data processing module is used for receiving the fixation point coordinates from the eye tracking system, and the implementation flow is as follows:
Step 411, receiving pupil center coordinates output by an eye tracking system;
Step 412, receiving and resolving the calibration gaze point coordinates;
step 413, performing coordinate correction and filtering based on the pupil center coordinate and the calibration gaze point coordinate, removing noise, and improving data precision;
step 414, storing the corrected gaze point coordinates and transmitting the data to the computing module;
The calculation module is used for analyzing the current required image plane depth, and the implementation flow is as follows:
step 421, receiving gaze point coordinate information from a data processing module;
step 422, calculating the current required image plane depth based on the gaze point-depth mapping model preset by the system;
Step 423, detecting an abnormal value of the calculation result, and performing interpolation processing to optimize the data smoothness;
step 424, storing the calculated depth information and sending it to the control instruction module;
the control instruction module is used for sending a motion instruction to the lens-screen adjusting system, and the implementation flow is as follows:
step 431, receiving the image plane depth data transmitted by the computing module;
Step 432, generating control instructions of the stepper motor according to the depth requirement, including forward, backward and stroke length;
Step 433, controlling the moving step length and speed of the micro stepping motor by adopting PWM (pulse width modulation) signals;
step 434, monitoring the state of the stepper motor in real time to ensure accurate adjustment of the lens position;
in step 435, the execution result is fed back, and secondary adjustment is performed if necessary, so as to improve the system response accuracy.
Referring to fig. 2, a dynamic virtual image plane adjustment method based on multi-image plane eye tracking is provided, the system is started, the system is initialized and all core systems are started, at this stage, all hardware is ensured to work normally, and software enters a standby state to prepare for subsequent operation;
the adjusting method comprises the following steps:
step 1, wearing equipment by a user, namely, correctly wearing the eye movement tracking equipment by the user, ensuring that the equipment stably fits the face, and adjusting the equipment to a proper position so that the camera can clearly capture the eye area of the user. The system self-checks whether the pupil of the user is completely visible or not, and adjusts the infrared illumination intensity so as to optimize the pupil imaging quality and ensure the accuracy of subsequent data acquisition;
Step 2, eye movement calibration, namely, the system enters a calibration mode, guides a user to sequentially watch a plurality of preset calibration points to establish a user eye movement model, and accurately fits the personalized eye movement parameters of the user by calculating the mapping relation between the pupil center offset and the screen coordinates to ensure higher precision of subsequent eye movement point prediction;
Step 3, gaze point prediction, in which under a normal operation mode, the system continuously collects eye movement data of a user, including pupil position, gaze direction and other characteristic information, based on an established eye movement model, the system calculates and predicts current gaze point coordinates of the user in real time, and determines a specific target area of the gaze of the user on a screen, wherein the prediction data is used for driving a subsequent self-adaptive vision adjustment mechanism;
Step 4, gaze point sensing, wherein the system further accurately measures the real-time gaze point of the user and carries out error correction on the predicted data; the system can detect the micro-movement of eyeballs of users and dynamically adjust the calculation precision of the fixation point so as to cope with the eye movement characteristics of different users;
Step 5, lens-screen distance adjustment, namely when the system confirms that the gaze point of the user is positioned in the threshold range of the region set by the gaze point sensing module, the driving circuit generates a driving signal and transmits the driving signal to the micro stepping motor, the micro stepping motor accurately moves the screen along the linear sliding rail to change the geometric distance between the screen and the lens so as to adjust the visual focal length to match the current vision requirement of the user;
The system releases hardware resources when exiting, saves personalized calibration parameters of the user, can be quickly loaded in subsequent use, and improves the response efficiency of the system;
The embodiments described in this specification are merely illustrative of the manner in which the inventive concepts may be implemented. The scope of the present invention should not be construed as being limited to the specific forms set forth in the embodiments, but the scope of the present invention and the equivalents thereof as would occur to one skilled in the art based on the inventive concept.

Claims (10)

1.一种基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述调节装置包括多像面眼动追踪系统、透镜-屏幕调节系统、控制系统以及显示系统,1. A dynamic virtual image plane adjustment device based on multi-image plane eye tracking, characterized in that the adjustment device comprises a multi-image plane eye tracking system, a lens-screen adjustment system, a control system and a display system, 所述多像面眼动追踪系统利用红外摄像机捕捉用户眼部图像,经过图像采集模块、瞳孔检测模块、椭圆拟合模块、标定模块和多项式拟合模块计算,转换得到精准的注视点坐标,并将数据传输至控制系统;The multi-image eye tracking system uses an infrared camera to capture the user's eye image, and converts it into accurate gaze point coordinates through calculations by an image acquisition module, a pupil detection module, an ellipse fitting module, a calibration module, and a polynomial fitting module, and transmits the data to a control system; 所述控制系统接收注视点信息,解析当前所需的像面深度,并生成透镜调整指令,发送至透镜-屏幕调节系统;The control system receives the gaze point information, analyzes the currently required image plane depth, generates a lens adjustment instruction, and sends it to the lens-screen adjustment system; 所述透镜-屏幕调节系统根据控制指令,驱动透镜和屏幕进行物理调整,同时利用传感反馈机制实时监测调整精度,确保虚像位置与用户注视焦点匹配;The lens-screen adjustment system drives the lens and the screen to make physical adjustments according to the control instructions, and uses a sensor feedback mechanism to monitor the adjustment accuracy in real time to ensure that the virtual image position matches the user's gaze focus; 所述显示系统结合控制系统提供的深度信息,动态调整动画画面,使其与用户的注视深度同步变化,增强视觉沉浸感。The display system dynamically adjusts the animation picture in combination with the depth information provided by the control system so that it changes synchronously with the user's gaze depth, thereby enhancing the visual immersion. 2.如权利要求1所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述多像面眼动追踪系统中,所述图像采集模块通过调用主动式红外摄像机获取用户眼部瞳孔图像,所述瞳孔检测模块用于进行瞳孔轮廓的标定,所述椭圆拟合模块用于检测置信度最高的瞳孔中心位置,所述标定模块结合已知参考点进行坐标转换,所述多项式拟合模块对数据进行优化处理,并最终将瞳孔数据转换为注视点坐标。2. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking as described in claim 1 is characterized in that, in the multi-image plane eye tracking system, the image acquisition module obtains the user's eye pupil image by calling an active infrared camera, the pupil detection module is used to calibrate the pupil contour, the ellipse fitting module is used to detect the pupil center position with the highest confidence, the calibration module performs coordinate conversion in combination with known reference points, the polynomial fitting module optimizes the data, and finally converts the pupil data into gaze point coordinates. 3.如权利要求2所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述的图像采集模块的实施流程如下:3. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to claim 2, characterized in that the implementation process of the image acquisition module is as follows: 步骤111,选用短焦摄像头以及红外灯珠;Step 111, select a short-focus camera and infrared lamp beads; 步骤112,将短焦摄像头及红外灯珠进行连接;Step 112, connecting the short-focus camera and the infrared lamp beads; 步骤113,调整短焦摄像头以及红外灯珠的位置,确保眼部被完全照亮;Step 113, adjusting the position of the short-focus camera and the infrared lamp beads to ensure that the eyes are fully illuminated; 所述的瞳孔检测模块的实施流程如下:The implementation process of the pupil detection module is as follows: 步骤121,根据瞳孔阈值算法判断瞳孔检测是否失败,若检测失败则需要重复步骤113,确保瞳孔检测的成功;Step 121, judging whether pupil detection fails according to the pupil threshold algorithm, if the detection fails, step 113 needs to be repeated to ensure the success of pupil detection; 步骤122,将输入的视频转化为视频帧,逐帧处理:Step 122, convert the input video into video frames and process them frame by frame: 对输入的原始图像进行裁剪和缩放操作,以统一图像尺寸并便于后续处理,通过计算当前图像的宽高比 Current_ratio与目标宽高比 Desire_ratio,若Current_ratio>Desire_ratio,表明图像过宽,则计算新的宽度 New_width = int(Desire_ratio *Current_ratio),并在图像宽度方向上从中心偏移 loc_w = (current_width - New_width)//2 的位置进行裁剪,得到裁剪后的图像cropped_img若 Current_ratio<=Desire_ratio,即图像过高,计算新的高度 New_height = int(current_width/Desire_ratio),在图像高度方向上从中心偏移loc_h= (current_height - New_height)//2 的位置裁剪,得到 cropped_img = image[loc_h: loc_h + New_height, :];最后,使用 cv2.resize 函数将裁剪后的图像调整为目标尺寸 (width, height);Crop and scale the input original image to unify the image size and facilitate subsequent processing. By calculating the aspect ratio of the current image Current_ratio and the target aspect ratio Desire_ratio, if Current_ratio>Desire_ratio, it indicates that the image is too wide, then calculate the new width New_width = int(Desire_ratio *Current_ratio), and crop it at the position offset loc_w = (current_width - New_width)//2 from the center in the image width direction to obtain the cropped image cropped_img. If Current_ratio<=Desire_ratio, that is, the image is too high, calculate the new height New_height = int(current_width/Desire_ratio), and crop it at the position offset loc_h = (current_height - New_height)//2 from the center in the image height direction to obtain cropped_img = image[loc_h: loc_h + New_height, :]; Finally, use cv2.resize The function resizes the cropped image to the target size (width, height); 步骤123,存储图像中A、B、C三种不同区域类似的二值图像:Step 123, storing similar binary images of three different regions A, B, and C in the image: 将裁剪并缩放后的图像转换为灰度图像,对灰度图像应用严格的二值化阈值;通过将最暗像素值Darkest_ value加上阈值增量,计算得到阈值Threshold = Darkest_ value+;使用cv2.threshold函数将图像二值化,小于该阈值的像素置为255,其余置为0,得到严格阈值处理后的图像Thresholded_image_1,即A;然后,以最暗点Darkest_point为中心,边长为创建一个正方形掩膜,将该正方形区域外的像素值设为,对 Thresholded_image_1进行掩膜处理;Convert the cropped and scaled image to a grayscale image and apply a strict binarization threshold to the grayscale image; add the darkest pixel value Darkest_ value to the threshold increment , the calculated threshold is Threshold = Darkest_ value + ; Use cv2.threshold function to binarize the image, set the pixels less than the threshold to 255, and the rest to 0, and get the strictly thresholded image Thresholded_image_1, i.e. A; then, with the darkest point Darkest_point as the center and the side length Create a square mask and set the pixel values outside the square area to , perform mask processing on Thresholded_image_1; 将最暗像素值Darkest_value加上阈值增量,计算得到阈值Threshold=Darkest_value+,对灰度图像进行二值化处理得到 Thresholded_image_2,即B,以最暗点为中心,边长进行掩膜处理;Add the darkest pixel value Darkest_value to the threshold increment , calculate the threshold Threshold = Darkest_value + , the grayscale image is binarized to obtain Thresholded_image_2, i.e. B, with the darkest point as the center and the side length Perform masking; 将最暗像素值Darkest_value加上阈值增量,计算得到阈值Threshold=Darkest_value+,对灰度图像进行二值化处理得到 Thresholded_image_3,即C;再以最暗点为中心,边长进行掩膜处理;Add the darkest pixel value Darkest_value to the threshold increment , calculate the threshold Threshold = Darkest_value + , binarize the grayscale image to get Thresholded_image_3, i.e. C; then take the darkest point as the center and the side length Perform masking; 并将Thresholded_image_1、Thresholded_image_2、Thresholded_image_3进行保存;And save Thresholded_image_1, Thresholded_image_2, and Thresholded_image_3; 步骤124,检测A、B、C三种不同区域的颜色阈值范围;Step 124, detecting the color threshold ranges of three different areas A, B, and C; 步骤125,选择平均颜色阈值波动较小的作为最可信的瞳孔图像。Step 125 , selecting the pupil image with the smallest fluctuation in average color threshold as the most credible pupil image. 4.如权利要求2所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述的椭圆拟合模块的实施流程如下:4. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to claim 2, characterized in that the implementation process of the ellipse fitting module is as follows: 步骤131,对每种二值图像进行膨胀处理,以增强目标区域:Step 131, dilation processing is performed on each binary image to enhance the target area: 对经过多阈值处理和掩膜后的图像,首先使用形态学膨胀操作来增强图像中的轮廓特征;然后,通过cv2.findContours函数查找图像中的轮廓;接着,对找到的轮廓进行过滤,遍历所有轮廓,计算每个轮廓的面积area=cv2.contourArea(contour),若 area>=pixel,pixel为预设阈值,则进一步计算轮廓的外接矩形的宽w和高h,并计算长宽比 length =max(w, h),width=min(w, h),Current_ratio=max(length/width, width/length);若Current_ratio<= ratio_thresh, ratio_thresh为阈值,则该轮廓符合条件,返回面积最大的符合条件的轮廓;For the image after multi-threshold processing and masking, first use the morphological dilation operation to enhance the contour features in the image; then, use the cv2.findContours function to find the contours in the image; then, filter the found contours, traverse all contours, and calculate the area of each contour area=cv2.contourArea(contour); if area>=pixel, pixel is the preset threshold, then further calculate the width w and height h of the circumscribed rectangle of the contour, and calculate the aspect ratio length =max(w, h), width=min(w, h), Current_ratio=max(length/width, width/length); if Current_ratio<= ratio_thresh, ratio_thresh is the threshold, then the contour meets the conditions, and the contour with the largest area that meets the conditions is returned; 步骤132,提取膨胀图像中的外部轮廓;Step 132, extracting the outer contour in the dilated image; 步骤133,通过面积和数量限制筛选轮廓,并返回最大轮廓;Step 133, filter the contours by area and quantity restrictions, and return the largest contour; 步骤134,在图像边缘的D个像素内不进行搜索;Step 134, not searching within D pixels at the edge of the image; 步骤135,在区域内每隔E个像素检查一次亮度;Step 135, checking the brightness every E pixels in the area; 步骤136,沿轴和轴每隔F个像素采样一次,忽略边界;Step 136, along Axis and The axis is sampled every F pixels, ignoring the boundaries; 步骤137,更新出0-255颜色通道之间阈值最大的像素点。Step 137, updating the pixel with the maximum threshold value between the 0-255 color channels. 5.如权利要求2所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述的标定模块的实施流程如下:5. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to claim 2, characterized in that the implementation process of the calibration module is as follows: 步骤141,选定屏幕尺寸作为标定屏幕,将屏幕参数输入控制系统并通过透镜投射进入人眼;Step 141, select a screen size as a calibration screen, input screen parameters into a control system and project them into human eyes through a lens; 步骤142,选择均匀铺满屏幕的9-12个点作为标定点;Step 142, selecting 9-12 points that evenly cover the screen as calibration points; 步骤143,用户佩戴好装置并调整画面中眼部图像的大小;Step 143, the user wears the device and adjusts the size of the eye image in the screen; 步骤144,通过外设如键盘或手柄进行自主标定,确保当前状态下所注视的点为标定点:Step 144, perform autonomous calibration through an external device such as a keyboard or a handle to ensure that the point being looked at in the current state is a calibration point: 在标定模式下,通过摄像头实时获取图像,利用上述瞳孔检测算法得到瞳孔的中心坐标 (x_eye, y_eye);在每一帧图像上绘制当前标定的目标点;当用户按下空格键时,记录当前瞳孔坐标 (x_eye, y_eye) 和对应的屏幕坐标到标定数据字典calibration_data中;当完成所有标定用点的记录后,将calibration_data保存到文件中;In calibration mode, the camera acquires images in real time, and the pupil center coordinates (x_eye, y_eye) are obtained using the pupil detection algorithm. The current calibration target point is drawn on each frame of the image. When the user presses the space bar, the current pupil coordinates (x_eye, y_eye) and the corresponding screen coordinates are recorded in the calibration data dictionary calibration_data. After all calibration points are recorded, calibration_data is saved to a file. 步骤145,调节透镜与屏幕之间的距离,从而改变像面距离人眼的远近,重复步骤142至步骤144;Step 145, adjusting the distance between the lens and the screen, thereby changing the distance between the image plane and the human eye, and repeating steps 142 to 144; 步骤146,保存不同深度下标定出的权重信息。Step 146, saving the weight information calibrated at different depths. 6.如权利要求2所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述的多项式拟合模块的实施流程如下:6. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to claim 2, characterized in that the implementation process of the polynomial fitting module is as follows: 步骤151,获取9-12个标定点的像素点坐标;Step 151, obtaining pixel coordinates of 9-12 calibration points; 步骤152,从视频帧中输出瞳孔中心的像素点坐标;Step 152, outputting the pixel coordinates of the pupil center from the video frame; 步骤153,将步骤151和步骤152中的像素点坐标拆分为一维数据;Step 153: The pixel coordinates in step 151 and step 152 are and Split into One-dimensional data; 步骤154,选取多项式对坐标点进行一维数据拟合;Step 154, selecting a polynomial to perform one-dimensional data fitting on the coordinate points; 步骤155,拟合函数中的多项式系数;Step 155, Fitting Polynomial coefficients in functions; 步骤156,拟合函数中的多项式系数;Step 156, Fitting Polynomial coefficients in functions; 步骤157,保存拟合出的轴和轴的一维多项式的系数:Step 157, save the fitted Axis and Coefficients of the one-dimensional polynomial of the axis: 从保存的标定数据文件中加载瞳孔坐标数据eye_coords和屏幕坐标数据screen_coords,提取瞳孔坐标的x分量eye_x和y分量eye_y,以及屏幕坐标的x分量screen_x和y分量screen_y使用函数分别对瞳孔x坐标与屏幕x坐标、瞳孔y坐标与屏幕y坐标进行多项式拟合,得到拟合系数coefficients_x和coefficients_y;Load pupil coordinate data eye_coords and screen coordinate data screen_coords from the saved calibration data file, extract the x component eye_x and y component eye_y of the pupil coordinate, and the x component screen_x and y component screen_y of the screen coordinate, and use functions to perform polynomial fitting on the pupil x coordinate and the screen x coordinate, and the pupil y coordinate and the screen y coordinate, respectively, to obtain fitting coefficients coefficients_x and coefficients_y; 步骤158,通过输入瞳孔中心点坐标,开始预测注视点坐标。Step 158, start predicting the gaze point coordinates by inputting the pupil center coordinates. 7.如权利要求1~6之一所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述透镜-屏幕调节系统包括透镜、屏幕、线性滑轨、微型步进电机和驱动电路,所述驱动电路与所述微型步进电机连接,所述微型步进电机的动作端与屏幕联动,所述屏幕可滑动地安装在所述线性滑轨上,所述线性滑轨的前后侧分别布置所述屏幕和透镜;7. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to any one of claims 1 to 6, characterized in that the lens-screen adjustment system comprises a lens, a screen, a linear slide rail, a micro stepping motor and a driving circuit, the driving circuit is connected to the micro stepping motor, the action end of the micro stepping motor is linked to the screen, the screen is slidably mounted on the linear slide rail, and the screen and the lens are arranged on the front and rear sides of the linear slide rail respectively; 所述的透镜-屏幕调节系统还包括调节模块和传感反馈模块,所述调节模块用于根据预测的注视点坐标实时选择需要调整的距离;所述传感反馈模块,用于驱动微型步进电机调节透镜与屏幕之间的距离;The lens-screen adjustment system further includes an adjustment module and a sensor feedback module, wherein the adjustment module is used to select the distance to be adjusted in real time according to the predicted gaze point coordinates; the sensor feedback module is used to drive the micro-stepping motor to adjust the distance between the lens and the screen; 所述的调节模块的实施流程如下:The implementation process of the adjustment module is as follows: 步骤211,根据多项式拟合模块所输出的预测的注视点坐标选定画面区域;Step 211, selecting a picture area according to the predicted gaze point coordinates output by the polynomial fitting module; 步骤212,当此区域被标记后调节提前设置好的行程范围;Step 212, when the area is marked, adjust the travel range set in advance; 所述的传感反馈模块的实施流程如下:The implementation process of the sensor feedback module is as follows: 步骤221,选用8mm行程的微型步进电机;Step 221, select a micro stepper motor with 8mm stroke; 步骤222,根据不同的注视区域施加PWM波控制微型步进电机行程。Step 222: Apply PWM waves to control the travel of the micro-stepping motor according to different gaze areas. 8.如权利要求1~6之一所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述的控制系统包括数据处理模块、计算模块和控制指令模块,所述数据处理模块用于接收来自眼动追踪系统的注视点坐标,所述计算模块用于解析当前所需的像面深度,所述控制指令模块用于发送运动指令给透镜-屏幕调节系统;8. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to any one of claims 1 to 6, characterized in that the control system comprises a data processing module, a calculation module and a control instruction module, the data processing module is used to receive the gaze point coordinates from the eye tracking system, the calculation module is used to analyze the currently required image plane depth, and the control instruction module is used to send motion instructions to the lens-screen adjustment system; 所述的数据处理模块的实施流程如下:The implementation process of the data processing module is as follows: 步骤311,接收眼动追踪系统输出的瞳孔中心坐标;Step 311, receiving pupil center coordinates output by the eye tracking system; 步骤312,接收并解析标定注视点坐标;Step 312, receiving and parsing the calibrated gaze point coordinates; 步骤313,基于瞳孔中心坐标与标定注视点坐标,进行坐标校正与滤波,去除噪声,提高数据精度;Step 313, based on the pupil center coordinates and the calibrated gaze point coordinates, coordinate correction and filtering are performed to remove noise and improve data accuracy; 步骤314,存储校正后的注视点坐标,并将数据传输至计算模块;Step 314, storing the corrected gaze point coordinates and transmitting the data to the calculation module; 所述的计算模块的实施流程如下:The implementation process of the calculation module is as follows: 步骤321,接收来自数据处理模块的注视点坐标信息;Step 321, receiving the gaze point coordinate information from the data processing module; 步骤322,基于系统预设的注视点-深度映射模型,计算当前所需的像面深度;Step 322, based on the gaze point-depth mapping model preset by the system, calculate the currently required image plane depth; 步骤323,对计算结果进行异常值检测,并进行插值处理以优化数据平滑度;Step 323, performing outlier detection on the calculation results and performing interpolation processing to optimize data smoothness; 步骤324,存储计算出的深度信息,并将其发送至控制指令模块;Step 324, storing the calculated depth information and sending it to the control instruction module; 所述的控制指令模块的实施流程如下:The implementation process of the control instruction module is as follows: 步骤331,接收计算模块传输的像面深度数据;Step 331, receiving image plane depth data transmitted by a calculation module; 步骤332,根据深度需求,生成步进电机的控制指令,包括前进、后退及行程长度;Step 332, generating control instructions for the stepper motor according to the depth requirement, including forward, backward and stroke length; 步骤333,采用脉冲宽度调制PWM信号控制微型步进电机的移动步长和速度;Step 333, using a pulse width modulation (PWM) signal to control the moving step length and speed of the micro-stepping motor; 步骤334,实时监测步进电机状态,确保透镜位置精准调整;Step 334, real-time monitoring of the stepper motor status to ensure accurate adjustment of the lens position; 步骤335,反馈执行结果,并在必要时进行二次调整,以提高系统响应精度。Step 335, feedback the execution result, and make secondary adjustments if necessary to improve the system response accuracy. 9.如权利要求1~6之一所述的基于多像面眼动追踪的动态虚拟像面调节装置,其特征在于,所述的显示系统包括动画预处理模块和注视点感应模块,所述动画预处理模块用于提前设定好画面中不同区域对应的虚像深度;所述注视点感应模块用于确保画面更新与透镜调整同步进行;9. The dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to any one of claims 1 to 6, characterized in that the display system comprises an animation preprocessing module and a gaze point sensing module, the animation preprocessing module is used to set the virtual image depth corresponding to different areas in the picture in advance; the gaze point sensing module is used to ensure that the picture update and the lens adjustment are synchronized; 所述的动画预处理模块的实施流程如下:The implementation process of the animation preprocessing module is as follows: 步骤411,根据现实生活中的深度信息设定画面中的深度信息;Step 411, setting the depth information in the picture according to the depth information in real life; 步骤412,对动画中的每一帧画面的不同区域预设深度信息;Step 412, presetting depth information for different areas of each frame of the animation; 步骤413,对不同深度下的画面预设不同的深度信息;Step 413, presetting different depth information for images at different depths; 步骤414,将预设的深度信息设为反馈终端;Step 414, setting the preset depth information as a feedback terminal; 所述的注视点感应模块的实施流程如下:The implementation process of the gaze point sensing module is as follows: 步骤421,设定区域阈值范围;Step 421, setting the area threshold range; 步骤422,对不同深度下的画面预设不同的区域阈值范围;Step 422, presetting different area threshold ranges for images at different depths; 步骤423,当注视点转移至区域阈值后产生反馈信息,驱动透镜-屏幕调节系统的微型步进电机实现进给。Step 423, when the gaze point is transferred to the regional threshold, feedback information is generated to drive the micro-stepping motor of the lens-screen adjustment system to achieve feeding. 10.一种如权利要求1所述的基于多像面眼动追踪的动态虚拟像面调节装置实现的方法,其特征在于,所述方法包括以下步骤:10. A method for implementing the dynamic virtual image plane adjustment device based on multi-image plane eye tracking according to claim 1, characterized in that the method comprises the following steps: 步骤1、用户佩戴设备:用户正确佩戴眼动追踪设备,确保设备稳固贴合面部,并调整位置,使摄像头能够清晰捕捉用户的眼部区域;自检用户的瞳孔是否完全可见,并调整红外光照强度;Step 1: The user wears the device: The user wears the eye tracking device correctly, ensures that the device fits the face firmly, and adjusts the position so that the camera can clearly capture the user's eye area; self-check whether the user's pupil is completely visible, and adjust the infrared light intensity; 步骤2、眼动校准:进入校准模式,引导用户依次注视多个预设校准点,以建立用户眼动模型;通过计算瞳孔中心偏移量与屏幕坐标之间的映射关系,精准拟合用户的个性化眼动参数;校准过程中,根据采样误差动态调整参数;Step 2, eye movement calibration: Enter the calibration mode and guide the user to look at multiple preset calibration points in turn to establish the user's eye movement model; accurately fit the user's personalized eye movement parameters by calculating the mapping relationship between the pupil center offset and the screen coordinates; during the calibration process, dynamically adjust the parameters according to the sampling error; 步骤3、注视点预测:在正常运行模式下,持续采集用户的眼动数据,包括瞳孔位置和视线方向;基于已建立的眼动模型实时计算并预测用户的当前注视点坐标,确定其视线在屏幕上的具体目标区域;此预测数据用于驱动后续的自适应视觉调整机制;Step 3, gaze point prediction: In normal operation mode, the user's eye movement data, including pupil position and line of sight direction, is continuously collected; based on the established eye movement model, the user's current gaze point coordinates are calculated and predicted in real time to determine the specific target area of their line of sight on the screen; this prediction data is used to drive the subsequent adaptive visual adjustment mechanism; 步骤4、注视点感应:进一步精确测量用户的实时注视点,并对预测数据进行误差修正;检测用户眼球的微小运动,并动态调整注视点的计算精度,以应对不同用户的眼动特性;Step 4: Gaze point sensing: further accurately measure the user's real-time gaze point and perform error correction on the predicted data; detect the user's eye movements and dynamically adjust the calculation accuracy of the gaze point to cope with the eye movement characteristics of different users; 步骤5、透镜-屏幕距离调整:所述透镜-屏幕调节系统包括透镜、屏幕、线性滑轨、微型步进电机和驱动电路,所述驱动电路与所述微型步进电机连接,所述微型步进电机的动作端与屏幕联动,所述屏幕可滑动地安装在所述线性滑轨上,所述线性滑轨的前后侧分别布置所述屏幕和透镜;当确认用户的注视点位于注视点感应模块所设定区域的阈值范围内时,驱动电路生成驱动信号,传输至微型步进电机,微型步进电机沿线性滑轨精确移动屏幕,使屏幕与透镜之间的几何距离发生变化,以调整视觉焦距,从而匹配用户的当前视线需求。Step 5, lens-screen distance adjustment: the lens-screen adjustment system includes a lens, a screen, a linear slide rail, a micro stepper motor and a drive circuit, the drive circuit is connected to the micro stepper motor, the action end of the micro stepper motor is linked to the screen, the screen is slidably mounted on the linear slide rail, and the screen and the lens are respectively arranged on the front and back sides of the linear slide rail; when it is confirmed that the user's gaze point is within the threshold range of the area set by the gaze point sensing module, the drive circuit generates a drive signal and transmits it to the micro stepper motor, and the micro stepper motor accurately moves the screen along the linear slide rail to change the geometric distance between the screen and the lens to adjust the visual focal length, thereby matching the user's current line of sight requirements.
CN202510504620.7A 2025-04-22 2025-04-22 Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking Active CN120029466B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202510504620.7A CN120029466B (en) 2025-04-22 2025-04-22 Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202510504620.7A CN120029466B (en) 2025-04-22 2025-04-22 Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking

Publications (2)

Publication Number Publication Date
CN120029466A CN120029466A (en) 2025-05-23
CN120029466B true CN120029466B (en) 2025-07-15

Family

ID=95738286

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202510504620.7A Active CN120029466B (en) 2025-04-22 2025-04-22 Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking

Country Status (1)

Country Link
CN (1) CN120029466B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107247511A (en) * 2017-05-05 2017-10-13 浙江大学 A kind of across object exchange method and device based on the dynamic seizure of eye in virtual reality
CN107609516A (en) * 2017-09-13 2018-01-19 重庆爱威视科技有限公司 Adaptive eye moves method for tracing

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6020923B2 (en) * 2013-05-21 2016-11-02 パナソニックIpマネジメント株式会社 Viewer having variable focus lens and video display system
CN110502100B (en) * 2019-05-29 2020-09-29 中国人民解放军军事科学院军事医学研究院 Virtual reality interaction method and device based on eye movement tracking
CN117882032A (en) * 2021-08-18 2024-04-12 元平台技术有限公司 Systems and methods for performing eye tracking
CN217467358U (en) * 2022-06-21 2022-09-20 深圳迈塔兰斯科技有限公司 Eye movement tracking system based on superlens, near-to-eye display optical system and equipment
CN117876426A (en) * 2024-01-18 2024-04-12 深圳市华弘智谷科技有限公司 A near-eye multi-camera eye tracking device and method that does not rely on corneal reflection

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107247511A (en) * 2017-05-05 2017-10-13 浙江大学 A kind of across object exchange method and device based on the dynamic seizure of eye in virtual reality
CN107609516A (en) * 2017-09-13 2018-01-19 重庆爱威视科技有限公司 Adaptive eye moves method for tracing

Also Published As

Publication number Publication date
CN120029466A (en) 2025-05-23

Similar Documents

Publication Publication Date Title
CN108427503B (en) Human eye tracking method and human eye tracking device
Moreno-Arjonilla et al. Eye-tracking on virtual reality: a survey
KR102000865B1 (en) A method for operating an eye tracking device and an eye tracking device
CN109521547B (en) Variable-step-length automatic focusing method and system
US6075557A (en) Image tracking system and method and observer tracking autostereoscopic display
US7783077B2 (en) Eye gaze tracker system and method
US20240385688A1 (en) Eye tracking device and a method thereof
US10048750B2 (en) Content projection system and content projection method
CN114973392B (en) Human eye motion tracking system and method
CN102855471A (en) Remote iris intelligent imaging device and method
US12315178B2 (en) Eye movement analysis method and system
CN112666705A (en) Eye movement tracking device and eye movement tracking method
CN111603134A (en) Eyeball movement testing device and method
US11054659B2 (en) Head mounted display apparatus and distance measurement device thereof
CN105761243A (en) Three-dimensional full face photographing system based on structured light projection and photographing method thereof
CN115022616B (en) A display device and display method for image focus enhancement based on human eye tracking
CN119653067A (en) A smart head-mounted device with a camera that can rotate with the user
CN116382473A (en) Sight calibration, motion tracking and precision testing method based on self-adaptive time sequence analysis prediction
CN120029466B (en) Dynamic virtual image plane adjusting device and method based on multi-image plane eye tracking
KR102085285B1 (en) System for measuring iris position and facerecognition based on deep-learning image analysis
CN114815262B (en) Contact MR (magnetic resonance) lens system based on bidirectional microvibration scanning projection imaging
CN111526286A (en) Method and system for controlling motor motion and terminal equipment
EP4652918A1 (en) Estimating prescription glasses strength for head-mounted display users
CN115019381B (en) Eye movement tracking model correction method and device
RU219079U1 (en) Device for determining the direction of gaze

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant