EP3032816A1 - Method for adapting a pre-captured videos to current lighting conditions - Google Patents
Method for adapting a pre-captured videos to current lighting conditions Download PDFInfo
- Publication number
- EP3032816A1 EP3032816A1 EP14307011.8A EP14307011A EP3032816A1 EP 3032816 A1 EP3032816 A1 EP 3032816A1 EP 14307011 A EP14307011 A EP 14307011A EP 3032816 A1 EP3032816 A1 EP 3032816A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- image
- video
- pixels
- date
- transformation matrix
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000000034 method Methods 0.000 title claims abstract description 35
- 239000011159 matrix material Substances 0.000 claims abstract description 53
- 230000009466 transformation Effects 0.000 claims abstract description 50
- 230000003190 augmentative effect Effects 0.000 claims description 11
- 238000004590 computer program Methods 0.000 claims description 2
- 238000004891 communication Methods 0.000 description 7
- 230000008569 process Effects 0.000 description 6
- 238000007654 immersion Methods 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 229920001690 polydopamine Polymers 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 239000013598 vector Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/46—Colour picture communication systems
- H04N1/56—Processing of colour picture signals
- H04N1/60—Colour correction or control
- H04N1/6083—Colour correction or control controlled by factors external to the apparatus
- H04N1/6088—Colour correction or control controlled by factors external to the apparatus by viewing conditions, i.e. conditions at picture output
Definitions
- the present disclosure relates generally to the field of video processing.
- AR augmented reality
- CGI computer generated images
- the user who is required to respect a relatively precise position on the site can rotate to change his view point in the 360° scene.
- the pre-captured video shot from the same position is augmented with additional elements that can be either natural or pure graphics.
- This operation can be performed offline through specialized post-production tools or when shooting by adding natural elements in the scene.
- These additional elements are for example texts on parts of the scene in the pre-captured video.
- the text can deal with historical or architectural aspects of some parts of the scene. For example, if the scene shows a skyscraper, the added text can give the height of the building and its date of building.
- Another known example of additional elements is the introduction of a music band playing in the scene.
- the user can be virtually positioned at the middle of the scene and navigate using his device as if he was capturing himself the music band playing in this scene.
- the navigation in this 360° video makes use of different sensors of the smartphone as the compass, the gyroscope or the zoom.
- the 360° video is realistic some differences with the current configuration of the site can sometimes be noticed, especially when light conditions are different. Indeed the conditions at the time of the visit and those at the period of capturing the video can actually be different. In such case, the video that should faithfully reproduce the environment is getting less realistic, and that could even be visually disruptive in case the discrepancy would be too large.
- a purpose of the present disclosure is to propose a method for compensating at least partially this discrepancy in order to improve the immersion of the user in the pre-captured video.
- the consistency of the lighting conditions of the pre-captured video with the current lighting conditions will improve the user's immersion in the pre-captured video.
- the present disclosure proposes a method for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the method comprising:
- the difference of lighting conditions between the first date (date of the first video or pre-captured video) and the second date (which may be the current date) is compensated by applying a color transformation matrix to the images of the first video (pre-captured video).
- the color transformation matrix M is determined by ar gmin M ⁇ x ⁇ C R x - M I x 2 where R(x) is the color value of a pixel x of the reference image R, I(x) is the color value of a pixel x of one image I of the first video and C is the set of pixels x within the images R and I.
- the image I is geometrically aligned on the reference image R and the color transformation matrix M is determined by ar gmin M ⁇ x ⁇ C R x - M I ⁇ x 2 where ⁇ is the aligned image I of the first video.
- the images I and R are geometrically aligned by the steps of:
- the set C comprises pixels other than said modified pixels. For example, if virtual content elements or text elements are added in the first video to obtain an augmented reality video, the pixels from which the color transformation matrix is determined are pixels that are not impacted by these additional elements.
- the color transformation matrix M is determined differently.
- the determination of the color transformation matrix M comprises the steps of:
- said first video is augmented by at least one additional element at a third date comprised between the first date and the second date, said at least additional element modifying pixels of images of the first video, said modified pixels are excluded from the selected pixels of said at least one image of the first video.
- the present disclosure concerns also a device for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the device comprising:
- the present disclosure concerns also a device for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the device comprising:
- the present disclosure also relates to a computer program product comprising instructions of program code for execution by at least one processor to perform the method of displaying at least one image of a video representing a scene, when the program is executed on a computer.
- the present disclosure also relates to a processor readable medium having stored therein instructions for causing a processor to perform the method of displaying at least one image of a video representing a scene.
- the present disclosure aims at improving the user's immersion when displaying at a given date a video of a scene captured at a previous date and more specifically when the user is present on the scene's site.
- a flow chart of the inventive method is shown on Figure 1 .
- the pre-captured video (captured at a first date) is modified in order to make the lighting conditions of the scene in this video correspond substantially to the lighting conditions of the scene at a second date (for example at a current date). These lighting conditions depend from the time of the day or the night the video is captured. They may also depend from the season, the weather... And they can change between two dates.
- first video captured at a first date is a video provided by a server.
- This first video is for example a video captured by a tour operator which is made available to the operator's clients.
- a step S1 at least one image, called reference image, of the scene is acquired at the second date under second lighting conditions.
- This image is captured by the user present on the site via its smartphone or a tablet or a laptop.
- the second date is the current date.
- the reference image is preferably captured from the same geographical position as the one from which the first video has been captured and preferably with the same orientation in the 3D space.
- These spatial coordinates can be captured by the GPS receiver or the gyroscope of the capturing device. They may be provided together with the first video by the tour operator. This position and this orientation may be retrieved via the GPS receiver or the gyroscope of his smartphone, tablet or laptop.
- a color transformation matrix M is determined in order to modify the color values of pixels of at least one image of the first video.
- the color transformation matrix M is applied to pixels of at least one image of the first video.
- the matrix M is determined such that the deviation between the color values of the pixels of the transformed image and the color values of the pixels of the reference image are lower than a predefined threshold T.
- the matrix M is applied to at least one image of the first video in order to make the lighting conditions of the scene in the transformed image (generated at step S3) correspond substantially to the lighting conditions of the scene in the reference image (captured at step S1). Two ways of determining the matrix M will be described later in the present document.
- the transformed image is displayed.
- the transformed image is displayed at the second date or at a later date. It is displayed on a smartphone, a tablet or a laptop of the user.
- the steps S3 and S4 can be applied to a plurality of images or to all the images of the first video.
- the matrix M can be determined in different ways.
- the color transformation matrix M is determined by the following equation ar gmin M ⁇ x ⁇ C R x - M I x 2 where
- the threshold T is substantially equal to zero.
- the matrix M is 4x4 color transformation matrix in homogeneous coordinates.
- the image I is preferably geometrically aligned on the reference image R.
- the color transformation matrix M is determined by the following equation ar gmin M ⁇ x ⁇ C R x - M I ⁇ x 2 where ⁇ is the aligned image I of the first video.
- the images I and R are for example geometrically aligned by the steps of:
- the pixels whose color value is modified by these additional elements are excluded from the set C.
- helps improving the minimization of the difference
- the matrix M is computed based on the mean gray level of selected pixels of the reference image R and the mean gray level of selected pixels of the image I of the first video.
- An illuminant is estimated for each one of the images R and I based on the so called gray world algorithm disclosed in the document " Robust automatic white balance algorithm using gray color points in images", Huo, J. Y., Chang, Y. L., Wang, J., & Wei, X. X., IEEE Transactions on Consumer Electronics, Vol.52, No.2, pages 541-546, May 2006 . and the matrix M is determined based on these two estimated illuminants.
- pixels in the YUV color space meeting the following condition are selected in the following image: U + V Y ⁇ T where T is a threshold between 0 and 1.
- the YUV color space used to extract gray points could be replaced by any other color space. In that case the above condition must be modified accordingly.
- the condition becomes respectively: a + b L ⁇ T And 2 R 2 + G 2 + B 2 - RG - RB - GB 3 ⁇ T
- the mean gray level of the selected pixels of image R and the mean gray level of the selected pixels of image I are computed.
- the mean gray level of the image R is computed once while a mean gray level will be possibly computed for a plurality of images I of the first video (pre-captured video).
- the matrix M is applied to at least one image I of the pre-captured video.
- some pixels of the image I can be excluded from the pixels selection process when they have been modified by the addition of elements (virtual elements, text elements%) in the image I.
- the method can be implemented by the user device which is for example a smartphone or a laptop.
- the first video pre-captured
- the first video is downloaded on the user device. In the case of a first video on a touristic site, this download can be done before the trip or on the site.
- the user when the user is on the site, he captures at least one image (reference image) of the site via his smartphone or laptop, and the transformation of the first video (computation and application of the matrix M) is performed on the fly by the user device itself as the first video is played.
- all the computation operations are performed on a distant computer of the provider of the first video (for example, the tour operator) or on any computer or server on the cloud.
- the reference image R is downloaded to this computer.
- the computer performs the transformation of the first video and the transformed video is then downloaded to the user device.
- the user captures a reference image at a current date (when present on the site) and this current image is used for amending the color values of the first video captured at a previous date.
- the reference image is not an image captured at a current date (when present on the site) but is an image captured by webcams present on the site some days before the trip or the visit.
- the user takes preferably an image corresponding to the time at which the visit will be done.
- This reference image is retrieved from the web some days before the visit such that the transformation operations of the first video can be performed offline. And the transformed image is downloaded on the user device before or during the visit.
- the reference image can be picked up from a database of images of the site under different lighting or weather conditions.
- Figure 2 illustrates schematically an exemplary hardware embodiment of a device 1 arranged to compute and display a transformed image as described before.
- This embodiment comprises a plurality of modules that can be present in a single device, i.e. the user device, or on a plurality of devices.
- the device 1 comprises:
- the reference image R is captured by the picture capturing circuit 10 and the first video is downloaded in the device via the communication interface 12. These images are stored in the local memory 11.
- the color transformation matrix M is determined and applied to at least one image I of the first video by the processing unit 14.
- the transformed video is stored in the local memory 11 and displayed on the display circuit 13.
- the first video can be transformed in a distant server.
- the reference image R captured by picture capturing circuit 10 is transmitted to the distant server via the communication interface 12.
- the color transformation to be applied to the first video is determined by the processing units of the distant server. This color transformation is applied to the first video on the distant server and the transformed video is then transmitted to the user device via the communication interface 12. The transformed video is then played by the user device.
- the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method or a device), the implementation of features discussed may also be implemented in other forms (for example a program).
- An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
- the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, Smartphones, tablets, computers, mobile phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
- PDAs portable/personal digital assistants
- Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications, particularly, for example, equipment or applications associated with data encoding, data decoding, view generation, texture processing, and other processing of images and related texture information and/or depth information.
- equipment include an encoder, a decoder, a post-processor processing output from a decoder, a pre-processor providing input to an encoder, a video coder, a video decoder, a video codec, a web server, a set-top box, a laptop, a personal computer, a cell phone, a PDA, and other communication devices.
- the equipment may be mobile and even installed in a mobile vehicle.
- the methods may be implemented by instructions being performed by a processor, and such instructions (and/or data values produced by an implementation) may be stored on a processor-readable medium such as, for example, an integrated circuit, a software carrier or other storage device such as, for example, a hard disk, a compact diskette (“CD"), an optical disc (such as, for example, a DVD, often referred to as a digital versatile disc or a digital video disc), a random access memory (“RAM”), or a read-only memory (“ROM”).
- the instructions may form an application program tangibly embodied on a processor-readable medium. Instructions may be, for example, in hardware, firmware, software, or a combination.
- a processor may be characterized, therefore, as, for example, both a device configured to carry out a process and a device that includes a processor-readable medium (such as a storage device) having instructions for carrying out a process. Further, a processor-readable medium may store, in addition to or in lieu of instructions, data values produced by an implementation.
- implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted.
- the information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
- a signal may be formatted to carry as data the rules for writing or reading the syntax of a described embodiment, or to carry as data the actual syntax-values written by a described embodiment.
- Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
- the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
- the information that the signal carries may be, for example, analog or digital information.
- the signal may be transmitted over a variety of different wired or wireless links, as is known.
- the signal may be stored on a processor-readable medium.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Processing (AREA)
Abstract
The present disclosure relates to a method for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the method comprising:
- acquiring (S1) at least one image, called reference image, of said scene at a second date under second lighting conditions, said second date being subsequent to said first date,
- determining (S2) a color transformation matrix M to be applied to the color values of pixels of at least one image of the first video,
- applying (S3) said color transformation matrix to said at least one image of the first video in order to make the lighting conditions of the scene in the transformed image correspond substantially to the second lighting conditions, and
- displaying (S4) the transformed image.
- acquiring (S1) at least one image, called reference image, of said scene at a second date under second lighting conditions, said second date being subsequent to said first date,
- determining (S2) a color transformation matrix M to be applied to the color values of pixels of at least one image of the first video,
- applying (S3) said color transformation matrix to said at least one image of the first video in order to make the lighting conditions of the scene in the transformed image correspond substantially to the second lighting conditions, and
- displaying (S4) the transformed image.
Description
- The present disclosure relates generally to the field of video processing.
- Accessing to information via the Internet from anywhere and at any time, coupled with the facility and intelligence of portable devices such as smartphones and tablets has led to new applications including that of augmented reality (AR). AR consists of adding virtual content as computer generated images (CGI) to natural content like video that can be captured by the user with his device. This technique is quite complex because requiring a precise estimation and tracking of the camera position in real time in order to get a reliable and realistic mixing of the video and CGI content.
- So a new way to simulate an AR application is to play an interactive video that has been pre-captured at the same location and that can be preloaded on the end user device. Such content can be 360° videos as proposed by the application developed by the company Onspot (see https://www.facebook.com/onspotapp). This application allows a viewer to watch videos shot in the exact position where he is located, thanks to a camera capable of shooting in all directions. The viewer is then completely immersed in the video and can turn its attention where he wants. Whatever the movements of the viewer, the real and the virtual are synchronous thanks to the gyroscope and the compass of his smartphone. This application is particularly useful for tourism industry. For example, it allows visitors immersing in virtual reconstructions in historical sites.
- In such application, the user who is required to respect a relatively precise position on the site can rotate to change his view point in the 360° scene. The pre-captured video shot from the same position is augmented with additional elements that can be either natural or pure graphics. This operation can be performed offline through specialized post-production tools or when shooting by adding natural elements in the scene. These additional elements are for example texts on parts of the scene in the pre-captured video. The text can deal with historical or architectural aspects of some parts of the scene. For example, if the scene shows a skyscraper, the added text can give the height of the building and its date of building. Another known example of additional elements is the introduction of a music band playing in the scene. The user can be virtually positioned at the middle of the scene and navigate using his device as if he was capturing himself the music band playing in this scene. The navigation in this 360° video makes use of different sensors of the smartphone as the compass, the gyroscope or the zoom.
- Although the 360° video is realistic some differences with the current configuration of the site can sometimes be noticed, especially when light conditions are different. Indeed the conditions at the time of the visit and those at the period of capturing the video can actually be different. In such case, the video that should faithfully reproduce the environment is getting less realistic, and that could even be visually disruptive in case the discrepancy would be too large.
- Current based AR video solutions do not propose any processing to compensate this discrepancy.
- A purpose of the present disclosure is to propose a method for compensating at least partially this discrepancy in order to improve the immersion of the user in the pre-captured video.
- According to the present disclosure, it is proposed to modify the color values of the pre-captured video in order to adapt it to other lighting conditions.
- The consistency of the lighting conditions of the pre-captured video with the current lighting conditions will improve the user's immersion in the pre-captured video.
- The present disclosure proposes a method for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the method comprising:
- acquiring at least one image, called reference image, of said scene at a second date under second lighting conditions, said second date being subsequent to said first date,
- determining a color transformation matrix M to be applied to the color values of pixels of at least one image of the first video, said color transformation matrix M being determined such that the deviation between the color values of the pixels of the transformed image and the color values of the pixels of said at least one reference image are lower than a predefined threshold,
- applying said color transformation matrix to said at least one image of the first video in order to make the lighting conditions of the scene in the transformed image correspond substantially to the second lighting conditions, and
- displaying the transformed image.
- Thus the difference of lighting conditions between the first date (date of the first video or pre-captured video) and the second date (which may be the current date) is compensated by applying a color transformation matrix to the images of the first video (pre-captured video).
-
-
- In one particular embodiment, the images I and R are geometrically aligned by the steps of:
- detecting at least one interest point in the reference image R and at least one interest point in the image I, and
- matching said at least one interest point in the image I with said at least one interest point in the reference image R.
- In one particular embodiment, if said first video is augmented by at least one additional element at a third date comprised between the first date and the second date, said at least additional element modifying pixels of images of the first video, the set C comprises pixels other than said modified pixels. For example, if virtual content elements or text elements are added in the first video to obtain an augmented reality video, the pixels from which the color transformation matrix is determined are pixels that are not impacted by these additional elements.
- In another embodiment, the color transformation matrix M is determined differently. In this embodiment, the determination of the color transformation matrix M comprises the steps of:
- selecting a group of pixels of said at least one image of the first video according to a first selection criterion,
- computing the mean gray level of the selected pixels of said at least one image of the first video;
- selecting a group of pixels of said at least one reference image according to the first selection criterion,
- computing the mean gray level of the selected pixels of said at least one reference image, and
- determining the color transformation matrix from the mean gray level of the selected pixels of said at least one image of the first video and the mean gray level of the selected pixels of said at least one reference image.
- Advantageously, if said first video is augmented by at least one additional element at a third date comprised between the first date and the second date, said at least additional element modifying pixels of images of the first video, said modified pixels are excluded from the selected pixels of said at least one image of the first video.
- The present disclosure concerns also a device for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the device comprising:
- means for acquiring at least one image, called reference image, of said scene at a second date under second lighting conditions, said second date being subsequent to said first date,
- means for determining a color transformation matrix M to be applied to the color values of pixels of at least one image of the first video, said color transformation matrix M being determined such that the deviation between the color values of the pixels of the transformed image and the color values of the pixels of said at least one reference image are lower than a predefined threshold,
- means for applying said color transformation matrix to said at least one image of the first video in order to make the lighting conditions of the scene in the transformed image correspond substantially to the second lighting conditions, and
- means for displaying the transformed image.
- The present disclosure concerns also a device for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the device comprising:
- a picture capturing device, e.g. a photosensor array, acquiring at least one image, called reference image, of said scene at a second date under second lighting conditions, said second date being subsequent to said first date,
- at least one processor determining a color transformation matrix M to be applied to the color values of pixels of at least one image of the first video, said color transformation matrix M being determined such that the deviation between the color values of the pixels of the transformed image and the color values of the pixels of said at least one reference image are lower than a predefined threshold,
- at least one processor applying said color transformation matrix to said at least one image of the first video in order to make the lighting conditions of the scene in the transformed image correspond substantially to the second lighting conditions, and
- a display device displaying the transformed image.
- The present disclosure also relates to a computer program product comprising instructions of program code for execution by at least one processor to perform the method of displaying at least one image of a video representing a scene, when the program is executed on a computer.
- The present disclosure also relates to a processor readable medium having stored therein instructions for causing a processor to perform the method of displaying at least one image of a video representing a scene.
- While not explicitly described, the present embodiments may be employed in any combination or subcombination.
- The present disclosure can be better understood with reference to the following description and drawings, given by way of example and not limiting the scope of protection, and in which:
-
Figure 1 is a flow chart of the steps of the inventive method; and -
Figure 2 is a schematic view of a device for implementing the method ofFigure 1 . - The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the present disclosure.
- The present disclosure aims at improving the user's immersion when displaying at a given date a video of a scene captured at a previous date and more specifically when the user is present on the scene's site. A flow chart of the inventive method is shown on
Figure 1 . The pre-captured video (captured at a first date) is modified in order to make the lighting conditions of the scene in this video correspond substantially to the lighting conditions of the scene at a second date (for example at a current date). These lighting conditions depend from the time of the day or the night the video is captured. They may also depend from the season, the weather... And they can change between two dates. - In the video captured at the first date, the scene is acquired under first lighting conditions. The lighting conditions at a second date subsequent to the first date are likely to be different. This second date may be the current date. The video, called first video, captured at a first date is a video provided by a server. This first video is for example a video captured by a tour operator which is made available to the operator's clients.
- According to a step S1, at least one image, called reference image, of the scene is acquired at the second date under second lighting conditions. This image is captured by the user present on the site via its smartphone or a tablet or a laptop. In that case, the second date is the current date. The reference image is preferably captured from the same geographical position as the one from which the first video has been captured and preferably with the same orientation in the 3D space. These spatial coordinates can be captured by the GPS receiver or the gyroscope of the capturing device. They may be provided together with the first video by the tour operator. This position and this orientation may be retrieved via the GPS receiver or the gyroscope of his smartphone, tablet or laptop.
- According to a step S2, a color transformation matrix M is determined in order to modify the color values of pixels of at least one image of the first video.
- According to a step S3, the color transformation matrix M is applied to pixels of at least one image of the first video.
- The matrix M is determined such that the deviation between the color values of the pixels of the transformed image and the color values of the pixels of the reference image are lower than a predefined threshold T. The matrix M is applied to at least one image of the first video in order to make the lighting conditions of the scene in the transformed image (generated at step S3) correspond substantially to the lighting conditions of the scene in the reference image (captured at step S1). Two ways of determining the matrix M will be described later in the present document.
- In the last step, referenced S4, the transformed image is displayed. The transformed image is displayed at the second date or at a later date. It is displayed on a smartphone, a tablet or a laptop of the user.
- The steps S3 and S4 can be applied to a plurality of images or to all the images of the first video.
- As mentioned above, the matrix M can be determined in different ways.
-
- R(x) is the color value of a pixel x of the reference image R,
- I(x) is the color value of a pixel x of one image I of the first video, and
- C is a set of pixels x within the images R and I.
- In this embodiment, the threshold T, defined above, is substantially equal to zero. The matrix M is 4x4 color transformation matrix in homogeneous coordinates.
-
- Different known methods can be used for aligning the images R and I. The images I and R are for example geometrically aligned by the steps of:
- detecting at least one interest point in the reference image R and at least one interest point in the image I, and
- matching said at least one interest point in the image I with said at least one interest point in the reference image R.
- This geometrical alignment helps improving the minimization of the difference |R(x)-M(I(x))|.
- In addition, if the first video is augmented (at a third date comprised between the first date and the second date) by additional elements such as virtual content elements or text elements to achieve an augmented reality video, the pixels whose color value is modified by these additional elements are excluded from the set C. This also helps improving the minimization of the difference |R(x)-M(I(x))|. For example, if a music band is added in the first video by the tour operator, the pixels associated with the objects that will not be present in the reference image are excluded from the set C.
- According to another embodiment, the matrix M is computed based on the mean gray level of selected pixels of the reference image R and the mean gray level of selected pixels of the image I of the first video. An illuminant is estimated for each one of the images R and I based on the so called gray world algorithm disclosed in the document "Robust automatic white balance algorithm using gray color points in images", Huo, J. Y., Chang, Y. L., Wang, J., & Wei, X. X., IEEE Transactions on Consumer Electronics, Vol.52, No.2, pages 541-546, May 2006. and the matrix M is determined based on these two estimated illuminants.
-
-
- Then the mean gray level of the selected pixels of image R and the mean gray level of the selected pixels of image I are computed. The mean gray level of the image R is computed once while a mean gray level will be possibly computed for a plurality of images I of the first video (pre-captured video).
-
- MA is a CAT transformation matrix to LMS cone space such as the matrix CAT02, Bradford, CMCCAT2000 and Sharp well-known in the literature, and
- the row vectors (XWS, YWS, ZWS) and (XWD, YWD, ZWD) are respectively the estimated white points of the images I and R in XYZ color space.
- In this embodiment, the matrix M is applied to at least one image I of the pre-captured video.
- More details for this transformation, with illustrated examples can be found in the website of Jason Su's CAT Project: http://web.stanford.edu/-sujason/ColorBalancing/adaptation.h tml
- As a generalization of this embodiment, it is possible to compute the mean of the p-norm in finite dimensions of the gray levels satisfying the conditions (1) or (2) or (3) instead of computing their mean gray level.
- As for the previous embodiment, some pixels of the image I can be excluded from the pixels selection process when they have been modified by the addition of elements (virtual elements, text elements...) in the image I.
- The implementation of the above-described method can be done in various ways.
- The method can be implemented by the user device which is for example a smartphone or a laptop. The first video (pre-captured) is downloaded on the user device. In the case of a first video on a touristic site, this download can be done before the trip or on the site. Then, when the user is on the site, he captures at least one image (reference image) of the site via his smartphone or laptop, and the transformation of the first video (computation and application of the matrix M) is performed on the fly by the user device itself as the first video is played.
- In another solution, all the computation operations are performed on a distant computer of the provider of the first video (for example, the tour operator) or on any computer or server on the cloud. In that case, when captured, the reference image R is downloaded to this computer. The computer performs the transformation of the first video and the transformed video is then downloaded to the user device. The advantage of this solution is that it does not require a powerful user device since the computations are made on distant computers.
- For these two solutions, the user captures a reference image at a current date (when present on the site) and this current image is used for amending the color values of the first video captured at a previous date.
- In a variant, the reference image is not an image captured at a current date (when present on the site) but is an image captured by webcams present on the site some days before the trip or the visit. The user takes preferably an image corresponding to the time at which the visit will be done. This reference image is retrieved from the web some days before the visit such that the transformation operations of the first video can be performed offline. And the transformed image is downloaded on the user device before or during the visit.
- In another variant, the reference image can be picked up from a database of images of the site under different lighting or weather conditions.
-
Figure 2 illustrates schematically an exemplary hardware embodiment of adevice 1 arranged to compute and display a transformed image as described before. This embodiment comprises a plurality of modules that can be present in a single device, i.e. the user device, or on a plurality of devices. - The
device 1 comprises: - a
picture capturing circuit 10 such as a CCD sensor or a CMOS sensor; - a
local memory 11, e.g. a video memory and/or a RAM memory and/or a ROM memory and/or a flash memory; - a
communication interface 12, e.g. a wired interface (for example a bus interface, a wide area network interface, a local area network interface) or a wireless interface (such as a IEEE 802.11 interface or a Bluetooth interface); - a
display circuit 13; and - one or
more processing units 14, for example a CPU and/or a GPU and/or a DSP. - All these circuits are connected via a
bus 15. - The reference image R is captured by the
picture capturing circuit 10 and the first video is downloaded in the device via thecommunication interface 12. These images are stored in thelocal memory 11. The color transformation matrix M is determined and applied to at least one image I of the first video by theprocessing unit 14. The transformed video is stored in thelocal memory 11 and displayed on thedisplay circuit 13. - As explained before, the first video can be transformed in a distant server. In that case, the reference image R captured by
picture capturing circuit 10 is transmitted to the distant server via thecommunication interface 12. The color transformation to be applied to the first video is determined by the processing units of the distant server. This color transformation is applied to the first video on the distant server and the transformed video is then transmitted to the user device via thecommunication interface 12. The transformed video is then played by the user device. - The implementations described herein may be implemented in, for example, a method or a process, an apparatus, a software program, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method or a device), the implementation of features discussed may also be implemented in other forms (for example a program). An apparatus may be implemented in, for example, appropriate hardware, software, and firmware. The methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, Smartphones, tablets, computers, mobile phones, portable/personal digital assistants ("PDAs"), and other devices that facilitate communication of information between end-users.
- Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications, particularly, for example, equipment or applications associated with data encoding, data decoding, view generation, texture processing, and other processing of images and related texture information and/or depth information. Examples of such equipment include an encoder, a decoder, a post-processor processing output from a decoder, a pre-processor providing input to an encoder, a video coder, a video decoder, a video codec, a web server, a set-top box, a laptop, a personal computer, a cell phone, a PDA, and other communication devices. As should be clear, the equipment may be mobile and even installed in a mobile vehicle.
- Additionally, the methods may be implemented by instructions being performed by a processor, and such instructions (and/or data values produced by an implementation) may be stored on a processor-readable medium such as, for example, an integrated circuit, a software carrier or other storage device such as, for example, a hard disk, a compact diskette ("CD"), an optical disc (such as, for example, a DVD, often referred to as a digital versatile disc or a digital video disc), a random access memory ("RAM"), or a read-only memory ("ROM"). The instructions may form an application program tangibly embodied on a processor-readable medium. Instructions may be, for example, in hardware, firmware, software, or a combination. Instructions may be found in, for example, an operating system, a separate application, or a combination of the two. A processor may be characterized, therefore, as, for example, both a device configured to carry out a process and a device that includes a processor-readable medium (such as a storage device) having instructions for carrying out a process. Further, a processor-readable medium may store, in addition to or in lieu of instructions, data values produced by an implementation.
- As will be evident to one of skill in the art, implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted. The information may include, for example, instructions for performing a method, or data produced by one of the described implementations. For example, a signal may be formatted to carry as data the rules for writing or reading the syntax of a described embodiment, or to carry as data the actual syntax-values written by a described embodiment. Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal. The formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream. The information that the signal carries may be, for example, analog or digital information. The signal may be transmitted over a variety of different wired or wireless links, as is known. The signal may be stored on a processor-readable medium.
- A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made. For example, elements of different implementations may be combined, supplemented, modified, or removed to produce other implementations. Additionally, one of ordinary skill will understand that other structures and processes may be substituted for those disclosed and the resulting implementations will perform at least substantially the same function(s), in at least substantially the same way(s), to achieve at least substantially the same result(s) as the implementations disclosed. Accordingly, these and other implementations are contemplated by this application.
Claims (15)
- Method for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the method comprising:- acquiring (S1) at least one image, called reference image, of said scene at a second date under second lighting conditions, said second date being subsequent to said first date,- determining (S2) a color transformation matrix M to be applied to the color values of pixels of at least one image of the first video, said color transformation matrix M being determined such that the deviation between the color values of the pixels of the transformed image and the color values of the pixels of said at least one reference image are lower than a predefined threshold,- applying (S3) said color transformation matrix to said at least one image of the first video in order to make the lighting conditions of the scene in the transformed image correspond substantially to the second lighting conditions, and- displaying (S4) the transformed image.
- Method according to claim 3, wherein the images I and R are geometrically aligned by:- detecting at least one interest point in the reference image R and at least one interest point in the image I, and- matching said at least one interest point in the image I with said at least one interest point in the reference image R.
- Method according to any one of claims 2 to 4, wherein, if said first video is augmented by at least one additional element at a third date comprised between the first date and the second date, said at least one additional element modifying pixels of images of the first video, the set C comprises pixels other than said modified pixels.
- Method according to claim 1, wherein the determination of the color transformation matrix M comprises:- selecting a group of pixels of said at least one image of the first video according to a first selection criterion,- computing the mean gray level of the selected pixels of said at least one image of the first video;- selecting a group of pixels of said at least one reference image according to the first selection criterion,- computing the mean gray level of the selected pixels of said at least one reference image, and- determining the color transformation matrix from the mean gray level of the selected pixels of said at least one image of the first video and the mean gray level of the selected pixels of said at least one reference image.
- Method according to claim 6, wherein, if said first video is augmented by at least one additional element at a third date comprised between the first date and the second date, said at least one additional element modifying pixels of images of the first video, said modified pixels are excluded from the selected pixels of said at least one image of the first video.
- Device configured for displaying at least one image of a video representing a scene, said video being derived from a first video representing said scene acquired at a first date under first lighting conditions, the device comprising:- means for acquiring at least one image, called reference image, of said scene at a second date under second lighting conditions, said second date being subsequent to said first date,- means for determining a color transformation matrix M to be applied to the color values of pixels of at least one image of the first video, said color transformation matrix M being determined such that the deviation between the color values of the pixels of the transformed image and the color values of the pixels of said at least one reference image are lower than a predefined threshold,- means for applying said color transformation matrix to said at least one image of the first video in order to make the lighting conditions of the scene in the transformed image correspond substantially to the second lighting conditions, and- means for displaying the transformed image.
- Device according to claim 10, wherein the images I and R are geometrically aligned by:- detecting at least one interest point in the reference image R and at least one interest point in the image I, and- matching said at least one interest point in the image I with said at least one interest point in the reference image R.
- Device according to any one of claims 9 to 11, wherein, if said first video is augmented by at least one additional element at a third date comprised between the first date and the second date, said at least one additional element modifying pixels of images of the first video, the set C comprises pixels other than said modified pixels.
- Device according to claim 8, wherein the determination of the color transformation matrix M comprises:- selecting a group of pixels of said at least one image of the first video according to a first selection criterion,- computing the mean gray level of the selected pixels of said at least one image of the first video;- selecting a group of pixels of said at least one reference image according to the first selection criterion,- computing the mean gray level of the selected pixels of said at least one reference image, and- determining the color transformation matrix from the mean gray level of the selected pixels of said at least one image of the first video and the mean gray level of the selected pixels of said at least one reference image.
- Device according to claim 12, wherein, if said first video is augmented by at least one additional element at a third date comprised between the first date and the second date, said at least one additional element modifying pixels of images of the first video, said modified pixels are excluded from the selected pixels of said at least one image of the first video.
- Computer program product comprising instructions of program code for executing steps of the method according to one of claims 1 to 7, when said program is executed on a computer.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP14307011.8A EP3032816A1 (en) | 2014-12-12 | 2014-12-12 | Method for adapting a pre-captured videos to current lighting conditions |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP14307011.8A EP3032816A1 (en) | 2014-12-12 | 2014-12-12 | Method for adapting a pre-captured videos to current lighting conditions |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| EP3032816A1 true EP3032816A1 (en) | 2016-06-15 |
Family
ID=52302098
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| EP14307011.8A Withdrawn EP3032816A1 (en) | 2014-12-12 | 2014-12-12 | Method for adapting a pre-captured videos to current lighting conditions |
Country Status (1)
| Country | Link |
|---|---|
| EP (1) | EP3032816A1 (en) |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20040056857A1 (en) * | 2002-04-24 | 2004-03-25 | Zhengyou Zhang | System and method for expression mapping |
| US20080101690A1 (en) * | 2006-10-26 | 2008-05-01 | De Dzwo Hsu | Automatic White Balance Statistics Collection |
| US20140055645A1 (en) * | 2012-08-21 | 2014-02-27 | Canon Kabushiki Kaisha | Image processing apparatus and method |
| WO2014064327A1 (en) * | 2012-10-26 | 2014-05-01 | Nokia Corporation | Method and apparatus for obtaining an image associated with a location of a mobile terminal |
-
2014
- 2014-12-12 EP EP14307011.8A patent/EP3032816A1/en not_active Withdrawn
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20040056857A1 (en) * | 2002-04-24 | 2004-03-25 | Zhengyou Zhang | System and method for expression mapping |
| US20080101690A1 (en) * | 2006-10-26 | 2008-05-01 | De Dzwo Hsu | Automatic White Balance Statistics Collection |
| US20140055645A1 (en) * | 2012-08-21 | 2014-02-27 | Canon Kabushiki Kaisha | Image processing apparatus and method |
| WO2014064327A1 (en) * | 2012-10-26 | 2014-05-01 | Nokia Corporation | Method and apparatus for obtaining an image associated with a location of a mobile terminal |
Non-Patent Citations (3)
| Title |
|---|
| HUO, J. Y.; CHANG, Y. L.; WANG, J.; WEI, X. X.: "Robust automatic white balance algorithm using gray color points in images", IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, vol. 52, no. 2, May 2006 (2006-05-01), pages 541 - 546 |
| JUN-YAN HUO ET AL: "Robust automatic white balance algorithm using gray color points in images", IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, IEEE SERVICE CENTER, NEW YORK, NY, US, vol. 52, no. 2, 1 May 2006 (2006-05-01), pages 541 - 546, XP007903612, ISSN: 0098-3063, DOI: 10.1109/TCE.2006.1649677 * |
| YU Y ET AL: "recovering photometric properties of architectural scenes from photographs", COMPUTER GRAPHICS. SIGGRAPH 98 CONFERENCE PROCEEDINGS. ORLANDO, FL, JULY 19- 24, 1998; [COMPUTER GRAPHICS PROCEEDINGS. SIGGRAPH], ACM, NEW YORK, NY, US, 19 July 1998 (1998-07-19), pages 207 - 218, XP002258538, ISBN: 978-0-89791-999-9, DOI: 10.1145/280814.280874 * |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10645298B2 (en) | Methods, devices and systems for automatic zoom when playing an augmented scene | |
| US9305339B2 (en) | Multi-feature image haze removal | |
| US10521468B2 (en) | Animated seek preview for panoramic videos | |
| US11222409B2 (en) | Image/video deblurring using convolutional neural networks with applications to SFM/SLAM with blurred images/videos | |
| US10467737B2 (en) | Method and device for adjusting grayscale values of image | |
| KR20190052089A (en) | Method and device for reconfiguring a point cloud representing scene using bright field data | |
| US10417276B2 (en) | Thumbnail generation from panoramic images | |
| KR20150140392A (en) | Estimating depth from a single image | |
| CN109754464B (en) | Method and apparatus for generating information | |
| CN110336973B (en) | Information processing method and device, electronic device and medium | |
| US20170236331A1 (en) | Method and system for geographic map overlay | |
| CN110222726A (en) | Image processing method, device and electronic equipment | |
| US20210158490A1 (en) | Joint rolling shutter correction and image deblurring | |
| CN111626919B (en) | Image synthesis method and device, electronic equipment and computer readable storage medium | |
| CN108352086A (en) | Determine and present solar flux information | |
| CN111429501A (en) | Depth map prediction model generation method and device and depth map prediction method and device | |
| CN113590878B (en) | Method, device and terminal equipment for planning path on video picture | |
| CN114399696A (en) | Target detection method and device, storage medium and electronic equipment | |
| CN117078574B (en) | Image deraining method and device | |
| CN113739797B (en) | Visual positioning method and device | |
| CN111932466B (en) | Image defogging method, electronic equipment and storage medium | |
| US11721006B2 (en) | Method and device for generating virtual reality data | |
| CN112714263A (en) | Video generation method, device, equipment and storage medium | |
| JP2016081225A (en) | Information presentation system | |
| CN111818265B (en) | Interaction method and device based on augmented reality model, electronic equipment and medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
| AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
| AX | Request for extension of the european patent |
Extension state: BA ME |
|
| 18D | Application deemed to be withdrawn |
Effective date: 20161216 |
|
| STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |