[go: up one dir, main page]

CN111836031A - VR video rendering method in multi-channel projection virtual reality - Google Patents

VR video rendering method in multi-channel projection virtual reality Download PDF

Info

Publication number
CN111836031A
CN111836031A CN201910315458.9A CN201910315458A CN111836031A CN 111836031 A CN111836031 A CN 111836031A CN 201910315458 A CN201910315458 A CN 201910315458A CN 111836031 A CN111836031 A CN 111836031A
Authority
CN
China
Prior art keywords
data
video
dimensional
dimensional data
perspective
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910315458.9A
Other languages
Chinese (zh)
Inventor
欧剑
郑亚冰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Elephant Virtual Reality Technology Co ltd
Original Assignee
Shenzhen Elephant Virtual Reality Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Elephant Virtual Reality Technology Co ltd filed Critical Shenzhen Elephant Virtual Reality Technology Co ltd
Priority to CN201910315458.9A priority Critical patent/CN111836031A/en
Publication of CN111836031A publication Critical patent/CN111836031A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/122Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/139Format conversion, e.g. of frame-rate or size

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)
  • Image Generation (AREA)

Abstract

The invention discloses a VR video rendering method in multi-channel projection virtual reality, which comprises the following steps: (1) acquiring geometric data of a projection surface by using a laser scanning method for the special-shaped surface; (2) performing space transformation on the geometric data to enable the geometric data to be mapped to a spherical space of the VR video according to perspective of a projector and surface three-dimensional information, wherein the surface data is changed into two-dimensional data of the VR video from the three-dimensional data; (3) the graphics processor performs hardware decoding on the two-dimensional data to form a corrected image; (4) outputting the rectified image to a projector; the beneficial effects are that: the method can effectively reduce the calculation efficiency of perspective transformation in the CAVE video, fully utilize the calculation capability of the graphic processor, convert three-dimensional data into two-dimensional data, and improve the rendering efficiency and the perspective correction effect of the multi-channel visual simulation system.

Description

VR video rendering method in multi-channel projection virtual reality
Technical Field
The invention belongs to the technical field of information, and particularly relates to a VR rendering method in multi-channel projection virtual reality.
Background
Cave (cave Automatic Virtual environment) is a projection-based Virtual reality system, which consists of four projection planes surrounding the viewer, as shown in the following figure. The four projection surfaces form a cubic structure, wherein three wall surfaces adopt a rear projection mode, and the ground surface adopts a front projection mode. The observer wears liquid crystal stereoscopic glasses and a six-degree-of-freedom head tracking device to feed back the viewpoint position of the observer to the computer system in real time and experience a feeling of being personally on the scene. As the viewer walks in CAVE, the system automatically calculates the correct stereoscopic perspective for each projection surface. At the same time, the observer holds a sensor called wands, interacting with the virtual environment.
When rendering 3D objects in CAVE, the method used is the Off Axis perspective (Off Axis Projection) method to render a three dimensional scene to a Projection surface. Moreover, perspective projection conversion is often performed with the viewpoint of the operator as the center; however, with this approach, 3D scene rendering using off-axis perspective transformation for each surface is required. Such as a CAVE containing four surfaces (front left and right), it requires 4 renderings to finally render the 3D scene onto the projection surface, and the rendering resolution of each surface is 1920x1080, which is computationally expensive for the rendering hardware. When a small scene or a VR video is rendered, most hardware can ensure real-time rendering. However, when the VR video exceeds 8K, since the graphics processor needs to complete both decoding and playing of the 8K video and off-axis perspective transformation, rendering efficiency is drastically reduced, and the VR video stream cannot be smoothly played.
Disclosure of Invention
The technical problem to be solved by the present invention is to provide a VR video rendering method in multi-channel projection virtual reality to solve the above mentioned problems in the background art.
In order to solve the technical problems, the technical scheme of the invention is as follows: a VR video rendering method in multi-channel projected virtual reality includes the following steps: (1) acquiring geometric data of a projection surface by using a laser scanning method for the special-shaped surface; (2) performing space transformation on the geometric data to enable the geometric data to be mapped to a spherical space of the VR video according to perspective of a projector and surface three-dimensional information, wherein the surface data is changed into two-dimensional data of the VR video from the three-dimensional data; (3) the graphics processor performs hardware decoding on the two-dimensional data to form a corrected image; (4) outputting the rectified image to a projector.
As a preferred embodiment of the present invention, a manual measurement method can be used to obtain the geometric data of the shaped surface.
As a preferred aspect of the present invention, the geometric data may also be mapped to a hexahedral space of the VR video according to the projector perspective and the surface three-dimensional information.
The invention adopts the technical scheme that the method has the beneficial effects that: the method can effectively reduce the calculation efficiency of perspective transformation in the CAVE video, fully utilize the calculation capability of the graphic processor, convert three-dimensional data into two-dimensional data, and improve the rendering efficiency and the perspective correction effect of the multi-channel visual simulation system.
Drawings
FIG. 1 is a schematic flow diagram of the present invention;
FIG. 2 is a schematic diagram of the principles of the present invention;
in the figure, 1-VR video, 2-spherical space, 3-projection surface.
Detailed Description
The following further describes embodiments of the present invention with reference to the drawings. It should be noted that the description of the embodiments is provided to help understanding of the present invention, but the present invention is not limited thereto. In addition, the technical features involved in the embodiments of the present invention described below may be combined with each other as long as they do not conflict with each other.
Referring to the drawings, a VR video rendering method in multi-channel projected virtual reality includes the following steps: (1) acquiring geometric data of the projection surface 3 by using a laser scanning method for the special-shaped surface; (2) performing spatial transformation on the geometric data to enable the geometric data to be mapped to a spherical space 2 of the VR video 1 according to perspective of a projector and surface three-dimensional information, wherein the surface data is changed into two-dimensional data of the VR video 1 from the three-dimensional data; (3) the graphics processor performs hardware decoding on the two-dimensional data to form a corrected image; (4) outputting the rectified image to a projector.
An artificial measurement method can also be used to form a corrected image for acquiring the geometric data of the profiled surface.
The geometry data may also be mapped to the hexahedral space of the VR video 1 according to the projector perspective and surface three-dimensional information.
VR technology is now widely used in our lives, and is now used in many fields of games, such as projecting an image of a player into a VR scene to form a mirror image model; checking the size data of the player and copying the checking data to the mirror image model; calculating the current action of the player, and setting the action state on the mirror model; taking a mirror image model of a player as a role model; however, when the VR video 1 exceeds 8K, since the graphics processor needs to complete both decoding and playing of the 8K video and off-axis perspective transformation, rendering efficiency is drastically reduced, and the VR video stream cannot be smoothly played.
And carrying out spatial transformation on the geometric data to enable the geometric data to be mapped to a spherical space 2 or a hexahedral space of the VR video 1 according to the perspective of a projector and the surface three-dimensional information, wherein the surface data are changed into 2D texture coordinates of the VR video 1 from 3D coordinates, and the processing speed of the graphics processor on the two-dimensional data is far higher than the processing speed of the three-dimensional data.
The embodiments of the present invention have been described in detail with reference to the accompanying drawings, but the present invention is not limited to the described embodiments. It will be apparent to those skilled in the art that various changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, and the scope of protection is still within the scope of the invention.

Claims (3)

1. A VR video rendering method in multi-channel projected virtual reality is characterized by comprising the following steps: (1) acquiring geometric data of a projection surface by using a laser scanning method for the special-shaped surface; (2) performing space transformation on the geometric data to enable the geometric data to be mapped to a spherical space of the VR video according to perspective of a projector and surface three-dimensional information, wherein the surface data is changed into two-dimensional data of the VR video from the three-dimensional data; (3) the graphics processor performs hardware decoding on the two-dimensional data to form a corrected image; (4) outputting the rectified image to a projector.
2. The VR video rendering method in multi-channel projected virtual reality of claim 1, wherein a manual measurement method is further used to obtain geometric data of the shaped surface.
3. The method of claim 1, wherein the geometry data is further mapped to a hexahedral space of the VR video according to projector perspective and surface three-dimensional information.
CN201910315458.9A 2019-04-18 2019-04-18 VR video rendering method in multi-channel projection virtual reality Pending CN111836031A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910315458.9A CN111836031A (en) 2019-04-18 2019-04-18 VR video rendering method in multi-channel projection virtual reality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910315458.9A CN111836031A (en) 2019-04-18 2019-04-18 VR video rendering method in multi-channel projection virtual reality

Publications (1)

Publication Number Publication Date
CN111836031A true CN111836031A (en) 2020-10-27

Family

ID=72915643

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910315458.9A Pending CN111836031A (en) 2019-04-18 2019-04-18 VR video rendering method in multi-channel projection virtual reality

Country Status (1)

Country Link
CN (1) CN111836031A (en)

Similar Documents

Publication Publication Date Title
US10957011B2 (en) System and method of capturing and rendering a stereoscopic panorama using a depth buffer
US7573475B2 (en) 2D to 3D image conversion
US10839591B2 (en) Stereoscopic rendering using raymarching and a virtual view broadcaster for such rendering
US10474227B2 (en) Generation of virtual reality with 6 degrees of freedom from limited viewer data
US7573489B2 (en) Infilling for 2D to 3D image conversion
US11025882B2 (en) Live action volumetric video compression/decompression and playback
US11282264B2 (en) Virtual reality content display method and apparatus
US9508191B2 (en) Optimal point density using camera proximity for point-based global illumination
US12100106B2 (en) Stereoscopic rendering of virtual 3D objects
US9401044B1 (en) Method for conformal visualization
US20210090322A1 (en) Generating and Modifying Representations of Objects in an Augmented-Reality or Virtual-Reality Scene
CN113366825B (en) Device and method for generating an image signal representing a scene
US6559844B1 (en) Method and apparatus for generating multiple views using a graphics engine
Lee et al. Real time 3D avatar for interactive mixed reality
CA3155612A1 (en) Method and system for providing at least a portion of content having six degrees of freedom motion
CN115841539A (en) Three-dimensional light field generation method and device based on visual shell
Leão et al. Altered reality: Augmenting and diminishing reality in real time
KR101163020B1 (en) Method and scaling unit for scaling a three-dimensional model
CN111836031A (en) VR video rendering method in multi-channel projection virtual reality
TW202046716A (en) Image signal representing a scene
US20200296349A1 (en) Stereo-aware panorama conversion for immersive media
JP2002260003A (en) Video display device
Zellmann et al. A Practical Guide to Implementing Off-Axis Stereo Projection Using Existing Ray Tracing Libraries
CN120411370A (en) Map updating method, image generation method, device and electronic equipment
GB2575932A (en) Method and system for providing at least a portion of content having six degrees of freedom motion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination