WO2016035500A1 - 検出機能付き投射型表示装置 - Google Patents
検出機能付き投射型表示装置 Download PDFInfo
- Publication number
- WO2016035500A1 WO2016035500A1 PCT/JP2015/072258 JP2015072258W WO2016035500A1 WO 2016035500 A1 WO2016035500 A1 WO 2016035500A1 JP 2015072258 W JP2015072258 W JP 2015072258W WO 2016035500 A1 WO2016035500 A1 WO 2016035500A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- detection
- light
- projection
- power saving
- projection display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/12—Picture reproducers
- H04N9/31—Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
- H04N9/3191—Testing thereof
- H04N9/3194—Testing thereof including sensor feedback
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B21/00—Projectors or projection-type viewers; Accessories therefor
- G03B21/14—Details
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3234—Power saving characterised by the action undertaken
- G06F1/325—Power saving in peripheral device
- G06F1/3265—Power saving in display device
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/042—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
- G06F3/0425—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means using a single imaging device like a video camera for tracking the absolute position of a single or a plurality of objects with respect to an imaged reference surface, e.g. video camera imaging a display or a projection screen, a table or a wall surface, on which a computer generated image is displayed or projected
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/20—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from infrared radiation only
- H04N23/21—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from infrared radiation only from near infrared [NIR] radiation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/63—Generation or supply of power specially adapted for television receivers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/12—Picture reproducers
- H04N9/31—Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
- H04N9/3141—Constructional details thereof
- H04N9/315—Modulator illumination systems
- H04N9/3155—Modulator illumination systems for controlling the light source
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B17/00—Details of cameras or camera bodies; Accessories therefor
- G03B17/48—Details of cameras or camera bodies; Accessories therefor adapted for combination with other photographic or optical apparatus
- G03B17/54—Details of cameras or camera bodies; Accessories therefor adapted for combination with other photographic or optical apparatus with projector
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/12—Picture reproducers
- H04N9/31—Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
- H04N9/3141—Constructional details thereof
- H04N9/315—Modulator illumination systems
- H04N9/3161—Modulator illumination systems using laser light sources
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/12—Picture reproducers
- H04N9/31—Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
- H04N9/3141—Constructional details thereof
- H04N9/315—Modulator illumination systems
- H04N9/3164—Modulator illumination systems using multiple light sources
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/12—Picture reproducers
- H04N9/31—Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
- H04N9/3141—Constructional details thereof
- H04N9/315—Modulator illumination systems
- H04N9/3167—Modulator illumination systems for polarizing the light beam
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D30/00—Reducing energy consumption in communication networks
- Y02D30/50—Reducing energy consumption in communication networks in wire-line communication networks, e.g. low power modes or reduced link rate
Definitions
- the present disclosure relates to a projection display device with a detection function having a function of detecting an object on or near a projection surface.
- Japanese Patent Application Laid-Open No. 2004-133620 proposes a power saving method by disposing a proximity sensor in the vicinity of the touch panel in a display device having a touch panel and shutting off the power supply to the touch panel according to the detection result of the proximity sensor. Yes.
- a projector also performs a pointing operation such that a projected image is operated by a hand according to a person's intuition like a tablet terminal.
- handy-type small projectors have recently appeared on the market, and it is desired to perform pointing operations on projected images projected on the order of 20 inches to 30 inches. Since the touch panel is not incorporated, it is necessary to detect a manual operation by another means.
- some projectors can move an image by operating a remote controller (remote controller) or the like.
- remote controller remote controller
- the device itself is small and the operation by the remote controller is not smart.
- a projection display device with a detection function includes a display light source that emits light that serves as illumination light, a projection display unit that projects an image on a projection surface using the illumination light, and a detection display A light source for detection that emits light, and a detection unit that detects an object on or near the projection surface, and the detection unit instructs the shift to the power saving mode according to the detection state of the object.
- a power instruction signal is output to the projection display unit, and the projection display unit turns off the light source for display based on the power saving instruction signal.
- the detection unit outputs a power saving instruction signal instructing a shift to the power saving mode to the projection display unit according to the detection state of the object. .
- the projection display unit turns off the display light source based on the power saving instruction signal.
- the projection display unit turns off the display light source based on the power saving instruction signal output from the detection unit according to the detection state of the object. Therefore, power saving can be achieved.
- the effects described here are not necessarily limited, and may be any of the effects described in the present disclosure.
- First embodiment projection type display device having a touch detection function
- FIGS. 1 to 6 Overall configuration and operation of the projection display optical system
- FIG. 6 Example of overall configuration of optical system of projection display device 1.1.2
- Other Modified Examples of Optical System 1.2 Configuration and Operation of Control System (FIGS. 7 to 9) 1.2.1 Configuration example of control system 1.2.2 Power-saving control operation 1.3 Effects Second embodiment (projection type display device having a power saving function according to gesture operation) (FIG.
- FIG. 1 shows an example of the overall configuration of an optical system of a projection display device (projector) according to the first embodiment of the present disclosure.
- This projection type display device has a function (touch detection function) for performing active object detection using near infrared light as well as video display.
- FIG. 2 shows an example of a state where video display and object detection are performed in the projection display device.
- FIG. 3 shows an example of a state in which the projection display device shown in FIG. 2 is viewed from the side surface direction.
- FIG. 4 shows an example of light incident on the light valve 21 and the image sensor 22 in the projection display device shown in FIG.
- FIG. 5 schematically shows the concept of image display and object detection by the projection display device.
- the projection display device includes an illumination unit 1, a light valve 21, an image sensor 22, a wire grid 27 as a polarization separation element, a projection lens 24, and a polarization as a polarization member.
- a child 25S, an image processing / touch detection unit 123, and an LD (Laser Diode) driver 122 for detection are provided.
- the illumination unit 1 emits the illumination light L1 from the first direction Z1 toward the wire grid 27 as shown in FIG.
- the illumination unit 1 includes a light source for display and a plurality of illumination optical members that generate illumination light L1 based on light from the light source and guide the illumination light L1 to the light valve 21.
- the light source for display may be a plurality of light sources arranged on different optical paths.
- the illumination unit 1 also includes an optical path combining element that combines optical paths of two or more light sources among the plurality of light sources.
- the illumination unit 1 includes a blue laser 11B, a green laser 11G, and a red laser 11R as a plurality of light sources arranged on different optical paths.
- the illumination unit 1 also includes a first coupling lens 12B, a second coupling lens 12G, a third coupling lens 12R, a drive optical element 14, and a mirror 18 as a plurality of illumination optical members.
- the third condenser lens 163 and the fourth condenser lens 164 are provided.
- the blue laser 11B is a laser light source that emits blue light having a wavelength of about 450 nm, for example.
- the green laser 11G is a laser light source that emits green light having a wavelength of about 520 nm, for example.
- the red laser 11R is a laser light source that emits red light having a wavelength of about 640 nm, for example.
- the detection LD driver 122 performs light emission control of the first light source (for example, the blue laser 11B), the second light source (for example, the green laser 11G), and the third light source (for example, the red laser 11R). For example, the detection LD driver 122 controls light emission of the first to third light sources by a field sequential method.
- the first light source for example, the blue laser 11B
- the second light source for example, the green laser 11G
- the third light source for example, the red laser 11R.
- the detection LD driver 122 controls light emission of the first to third light sources by a field sequential method.
- the second coupling lens 12G is a lens (coupled lens) for collimating the green light emitted from the green laser 11G (as parallel light) and coupling it with the first dichroic prism 131.
- the first coupling lens 12B is a lens (coupling lens) for collimating the blue light emitted from the blue laser 11B and coupling it with the first dichroic prism 131.
- the third coupling lens 12R is a lens (coupling lens) for collimating the red light emitted from the red laser 11R and coupling it with the second dichroic prism 132.
- Each of the first dichroic prism 131 and the second dichroic prism 132 is an optical path combining element that combines optical paths of two or more light sources.
- the first dichroic prism 131 selectively transmits the blue light incident through the first coupling lens 12B while selectively transmitting the green light incident through the second coupling lens 12G. It is a reflecting prism.
- the second dichroic prism 132 selectively transmits the blue light and the green light emitted from the first dichroic prism 131, while selectively transmitting the red light incident through the third coupling lens 12R. It is a reflecting prism.
- color synthesis optical path synthesis
- the drive optical element 14 is an optical element for reducing speckle noise and interference fringes in the illumination light L ⁇ b> 1, and is disposed on the optical path between the first condenser lens 161 and the second condenser lens 162. .
- the drive optical element 14 changes the state of the light beam passing therethrough, for example, by microvibration in a direction along the optical axis or in a direction perpendicular to the optical axis, thereby reducing speckle noise and interference fringes in the illumination light L1. It is possible to make it.
- Each of the first fly-eye lens 151 and the second fly-eye lens 152 is an optical member (integrator) in which a plurality of lenses are two-dimensionally arranged on a substrate. Are divided and emitted.
- the first fly-eye lens 151 is disposed on the optical path between the second dichroic prism 132 and the first condenser lens 161.
- the second fly-eye lens 152 is disposed on the optical path between the second condenser lens 162 and the third condenser lens 163.
- the mirror 18 is an element that bends the optical path of the illumination light L1.
- the mirror 18 is disposed on the optical path between the first condenser lens 161 and the drive optical element 14.
- the first condenser lens 161 is a lens that collects the light emitted from the first fly-eye lens 151 and makes it incident on the drive optical element 14 via the mirror 18.
- the second condenser lens 162 is a lens that collects the light emitted from the drive optical element 14 and makes it incident on the second fly-eye lens 152.
- the third condenser lens 163 and the fourth condenser lens 164 are lenses for collecting the emitted light from the second fly-eye lens 152 and emitting it as the illumination light L1 toward the wire grid 27.
- the wire grid 27 is formed, for example, by forming a metal lattice on a glass substrate at a minute interval. As shown in FIG. 4, the illumination light L1 is incident on the wire grid 27 from the first direction Z1. A light valve 21 is arranged in the second direction Z2. In the third direction Z3, the polarizer 25S and the image sensor 22 are arranged. The projection lens 24 is disposed in the fourth direction Z4.
- the wire grid 27 is a polarization separation element that separates incident light into a first polarization component (for example, a P polarization component) and a second polarization component (for example, an S polarization component) and emits them in different directions. .
- the wire grid 27 selectively reflects a specific first polarization component and selectively transmits a specific second polarization component.
- the wire grid 27 emits (reflects) most of the P-polarized component Lp1 included in the illumination light L1 incident from the first direction Z1 in the second direction Z2 and also S-polarized light. Most of the component Ls1 is emitted (transmitted) in the third direction Z3.
- the wire grid 27 also emits (reflects) most of the P-polarized component Lp3 contained in the detection light L2 incident from the direction opposite to the fourth direction Z4 in the third direction Z3. It is like that.
- the light valve 21 is a reflective liquid crystal element such as LCOS (Liquid Crystal On Silicon).
- the light valve 21 images the first polarization component (for example, the P polarization component Lp1) included in the illumination light L1 that is incident from the second direction Z2 via the wire grid 27. Modulation is based on the data.
- the light valve 21 also emits the modulated light through the wire grid 27 in the fourth direction Z4.
- the light valve 21 emits, for example, an S-polarized component Ls2 whose polarization state has been rotated as incident light as modulated light. Note that the light valve 21 can perform black display by returning the incident P-polarized component Lp1 to the wire grid 27 in the polarization state as it is.
- the projection lens 24 projects the modulated light from the light valve 21 incident from the fourth direction Z4 via the wire grid 27 onto the projection surface 30A of the screen 30. Further, as shown in FIG. 4, the detection light L2 is incident on the projection lens 24 from the direction opposite to the traveling direction of the modulated light.
- the projection lens 24 is a projection optical system for projecting an image and functions as an imaging optical system for object detection.
- the image sensor 22 is composed of a solid-state image sensor such as a CMOS (Complementary Metal-Oxide Semiconductor) or a CCD (Charge Coupled Device).
- the image sensor 22 is disposed at a position optically conjugate with the light valve 21. More specifically, when the light valve 21 is a reflective liquid crystal element, the display surface (liquid crystal surface) for creating an image and the image pickup surface of the image pickup element 22 are arranged at an optically conjugate position. As shown in FIG. 4, the detection light L ⁇ b> 2 is incident on the image sensor 22 from the third direction Z ⁇ b> 3 via the projection lens 24 and the wire grid 27.
- CMOS Complementary Metal-Oxide Semiconductor
- CCD Charge Coupled Device
- the polarizer 25S is a polarizing member that is one of the optical members that reduces the second polarization component contained in the illumination light L1.
- the polarizer 25 ⁇ / b> S is disposed between the image sensor 22 and the wire grid 27.
- the polarizer 25S removes a second polarization component (for example, an S polarization component) included in the incident light.
- the polarizer 25 ⁇ / b> S removes at least the S-polarized component Ls ⁇ b> 1 included in the illumination light L ⁇ b> 1 incident through the wire grid 27 as the second polarized component.
- the image processing / touch detection unit 123 based on the detection signal from the image sensor 22 (photographing result by the image sensor 22), for example, as shown in FIG. 2, FIG. 3, and FIG.
- the position P1 of the feature point of the indicator (object) 71 is detected in association with the coordinates of the projection image V2 projected on the projection plane 30A.
- 2, 3, and 5 show the position of the tip of the human finger as an example of the feature point, but the present invention is not limited to this, and the center of gravity of the human finger, the center of gravity of the hand, and the like can be selected as appropriate.
- FIG. 2 and FIG. 3 show configurations assuming that the projection display device is a short focus type.
- the projection display device includes a near-infrared light projector 40 at the bottom of the main body 100.
- the projection surface 30A is, for example, a flat floor surface.
- the near-infrared light projecting unit 40 is a detection light source unit that emits detection near-infrared light 41 as invisible light for detection from a predetermined height h with respect to the projection surface 30A.
- the near-infrared light projector 40 may include an infrared laser 121 shown in FIG. 7 to be described later as a detection light source.
- Near-infrared light projector 40 emits detection near-infrared light 41 at a predetermined height h from projection surface 30A so as to cover at least an area corresponding to video projection area 31.
- Near-infrared scattered light La diffused by the indicator 71 is incident on the image sensor 22 as detection light L ⁇ b> 2 through the projection lens 24 and the wire grid 27.
- the near-infrared light projecting unit 40 may irradiate the projection surface 30A with detection near-infrared light 41 having a thickness in the height h direction as invisible light for detection.
- the near-infrared light for detection 41 and the projection surface 30A do not necessarily have to be completely separated at a predetermined height h.
- the projection lens 24 may be an ultra short focus lens having a throw ratio of 0.38 or less.
- the slow ratio is represented by L / H, where L is the distance from the projection lens 24 to the projection surface 30A and H is the width of the projection area, as shown in FIGS.
- the projection surface 30A does not have to be a special screen. Further, the projection surface 30A is not limited to the floor surface as shown in FIGS. 2 and 3, and may be a wall surface or the like.
- the image information V1 formed on the light valve 21 is projected onto the projection surface 30A of the screen 30 by the projection lens 24, and is enlarged and displayed as a projection image V2.
- the projection display apparatus also detects the position of an object on the projection surface 30 ⁇ / b> A, for example, the position P ⁇ b> 1 of a feature point of an indicator (object) 71 such as a human finger or a pointer, using the imaging element 22.
- the imaging element 22 performs imaging using an imaging area 32 that is substantially the same area as the projection area 31 on the projection surface 30A.
- the first polarization component may be 99% or more, more preferably 99.5% or more.
- the dominant first polarization component either the S-polarization component Ls1 or the P-polarization component Lp1 can be selected in accordance with the characteristics of the polarization conversion element.
- the wire grid 27 reflects most of the P-polarized component and transmits most of the S-polarized component when the first polarized component is the P-polarized component and the second polarized component is the S-polarized component. For this reason, for example, 99.5% of the illumination light L1 is set to be the P polarization component Lp1, and the remaining 0.5% is set to the S polarization component Ls1. For example, as shown in FIG. 4, the wire grid 27 reflects most of the dominant P-polarized light component Lp ⁇ b> 1 and emits it to the light valve 21.
- the P-polarized light component Lp1 incident on the light valve 21 is modulated (rotated) by the light valve 21 to become modulated light of the S-polarized light component Ls2, and then enters the projection lens 24 via the wire grid 27.
- the S-polarized light component Ls2 that is the modulated light is projected as a projected image V2 onto the projection surface 30A of the screen 30 via the projection lens 24 as shown in FIG.
- the image sensor 22 is disposed at a position optically conjugate with the light valve 21.
- the projection lens 24 is a projection optical system for projecting an image and functions as an imaging optical system for object detection. For this reason, as shown in FIG. 5, it is possible to take an image with the imaging element 22 with the same area as the projection area 31 as the imaging area 32. Since the light valve 21 and the image sensor 22 are conjugate positions, the position P1 of the feature point of the indicator 71 such as a human finger or a pointer on the projection surface 30A is superimposed on the projection image V2 via the projection lens 24. Can be monitored.
- the image processing / touch detection unit 123 performs image processing on the shape of the pointing object 71 and detects the coordinates of the position P1 of the feature point of the pointing object 71, thereby enabling the pointing operation of the projection image V2.
- the coordinate of the detection position P2 on the image sensor 22 side is the position of the feature point position P1 of the indicator 71
- the indicator 71 may be two or more. For example, the coordinates of the tips of the fingers of both hands can be detected.
- the projection area 31 has a predetermined height h of several millimeters to several tens of millimeters from the projection surface 30A, the height direction is 2 to 3 mm, and the area direction. Applies a film-like near-infrared barrier to a range covering the projection area 31. Then, since the projection surface 30A is generally flat, the film of the emitted near-infrared light goes straight without being interrupted in the middle if there is no indicator 71 such as a shield, a finger, or a pointer. Therefore, the image is not captured on the image pickup device 22 that is monitoring the projection surface 30A.
- the position can be detected. Further, in the case of the ultra short focus type, the projection light passes through the vicinity of the projection surface 30A, and a part of the operating human body is difficult to block the projection light, so that there is an advantage that the screen is easy to see when operating.
- the detection light L2 incident on the wire grid 27 includes an S-polarized component Ls3 and a P-polarized component Lp3 as polarized components.
- the wire grid 27 reflects most of the P-polarized component Lp3 in the third direction Z3. If the polarizer 25S removes the S-polarized light component, almost all of the reflected P-polarized light component Lp3 reaches the image sensor 22. In addition, of the illumination light L1 incident on the wire grid 27, the S-polarized component Ls1 is emitted in the third direction Z3.
- the S-polarized component Ls1 becomes a noise component with respect to the detection light L2, and when incident on the image sensor 22, the S / N ratio at the time of detection is reduced, and the detection accuracy is deteriorated.
- the polarizer 25S and removing the S-polarized component Ls1 it is possible to increase the S / N ratio and increase the detection accuracy.
- the P-polarized component Lp1 in the illumination light L1 is reflected by the wire grid 27 in a direction different from that of the imaging device 22, and the S-polarized component Ls1 is removed by the polarizer 25S, so that the imaging device ideally. Only the detection light L ⁇ b> 2 can be incident on 22.
- the polarizing beam splitter 23 may have a configuration in which prisms coated with a multilayer film are bonded together, or may be a beam splitter similar to a prism in which an element having polarization characteristics is sanded.
- the wire grid 27 in the configuration of FIG. 4 uses the first polarization component as the P polarization component, the second polarization component as the S polarization component, reflects the P polarization component, and transmits the S polarization component.
- the beam splitter 23 has the opposite characteristic.
- the polarizing beam splitter 23 has four optical surfaces. Here, the two surfaces facing in the horizontal direction in FIG. 6 are described as the first optical surface and the third optical surface, and the two surfaces facing in the vertical direction are described as the second optical surface and the fourth optical surface. To do. As shown in FIG. 6, the illumination light L1 is incident on the first optical surface of the polarization beam splitter 23 from the first direction Z1. A light valve 21 is arranged in the second direction Z2 with respect to the second optical surface of the polarization beam splitter 23. With respect to the third optical surface of the polarization beam splitter 23, the polarizer 25 and the imaging device 22 are arranged in the third direction Z3. A projection lens 24 is disposed in the fourth direction Z4 with respect to the fourth optical surface of the polarization beam splitter 23.
- the polarization beam splitter 23 is a polarization separation element that separates incident light into a first polarization component (for example, an S polarization component) and a second polarization component (for example, a P polarization component) and emits them in different directions. is there.
- the polarization beam splitter 23 selectively reflects the specific first polarization component and selectively transmits the specific second polarization component.
- the polarization beam splitter 23 emits (reflects) almost all of the S-polarized component Ls1 included in the illumination light L1 incident from the first direction Z1 in the second direction Z2.
- Almost all of the P-polarized light component Lp1 is emitted (transmitted) in the third direction Z3.
- the polarization beam splitter 23 also emits (reflects) almost all of the S-polarized component Ls3 included in the detection light L2 incident from the direction opposite to the fourth direction Z4 in the third direction Z3. ).
- the polarization beam splitter 23 reflects most of the S polarization component and transmits most of the P polarization component. . For this reason, for example, 99.5% of the illumination light L1 can be made dominant by using the S-polarized component Ls1, and the remaining 0.5% can be made the P-polarized component Lp1. As shown in FIG. 6, the polarization beam splitter 23 reflects almost all the dominant S-polarized component Ls ⁇ b> 1 and emits it to the light valve 21.
- the S-polarized component Ls1 incident on the light valve 21 is modulated (rotated) by the light valve 21 to become modulated light of the P-polarized component Lp2, and then enters the projection lens 24 via the polarizing beam splitter 23.
- the P-polarized light component Lp2 that is the modulated light is projected as a projection image V2 onto the projection surface 30A of the screen 30 via the projection lens 24 as shown in FIG.
- the detection light L2 incident on the polarization beam splitter 23 includes an S polarization component Ls3 and a P polarization component Lp3 as polarization components.
- the polarization beam splitter 23 reflects almost all of the S-polarized component Ls3 in the third direction Z3. If the polarizer 25 removes the P-polarized component, almost all of the S-polarized component Ls3 reaches the image sensor 22.
- the P-polarized component Lp1 is emitted in the third direction Z3.
- the P-polarized component Lp1 becomes a noise component with respect to the detection light L2, and when incident on the image sensor 22, the S / N ratio at the time of detection is reduced, and the detection accuracy is deteriorated.
- the S / N ratio can be increased and the detection accuracy can be increased.
- the S-polarized component Ls1 in the illumination light L1 is reflected by the polarizing beam splitter 23 in a direction different from that of the imaging element 22, and the P-polarized component Lp1 is removed by the polarizer 25, so that ideal imaging is performed.
- Only the detection light L2 can be incident on the element 22, but depending on the incident angle of the light to the polarization beam splitter 23 and the optical performance of the polarization beam splitter 23 and the polarizer 25, unnecessary light included in the illumination light L1 is necessary.
- a noise component enters the image sensor 22. Therefore, as shown in FIG. 6, it is desirable to have a configuration in which the noise component with respect to the detection light is reduced inside the illumination unit 1.
- either the first fly-eye lens 151 or the second fly-eye lens 152 may be one.
- the first condenser lens 161 and the second condenser lens 162 are not necessary.
- the third condenser lens 163 and the fourth condenser lens 164 are not necessary.
- the polarizer 25S in the configuration of FIG. 1 may be omitted.
- This technology is also applicable as a digital mirror device type projector.
- infrared band light is taken as an example of the detection light L2
- the ultraviolet light may be the detection light L2.
- FIG. 7 shows a configuration example of a control system in the projection display apparatus of the present embodiment.
- the projection display device projects a portion (projection display unit 110) that projects an image on the projection surface 30A with the illumination light L1, and an object on or near the projection surface 30A.
- a detection portion detection unit 120.
- the projection display unit 110 includes a light valve 21 and a blue laser 11B, a green laser 11G, and a red laser 11R as display light sources that emit light serving as illumination light L1.
- the projection display unit 110 also includes a display LD driver 111, a control unit 112, a display element controller 113, and a video processing unit 114.
- the detection unit 120 includes an image sensor 22 and an infrared laser 121 as a light source for detection that emits light for detection.
- the detection unit 120 also includes a detection LD driver 122 and an image processing / touch detection unit 123.
- the video processing unit 114 generates a video signal for display based on the input video data.
- the display element controller 113 generates a display control signal based on the video signal from the video processing unit 114, and performs drive control of the light valve 21 as a display element.
- the image processing / touch detection unit 123 analyzes the detection signal from the image sensor 22 as described above and outputs position data (coordinate data) of the detected object.
- the image processing / touch detection unit 123 may have a function of analyzing not only the position of the object but also the movement of the object such as a gesture operation.
- FIG. The image processing / touch detection unit 123 also, when an object is detected after shifting to the power saving mode, a cancellation instruction that instructs to cancel the power saving mode to shift to the normal operation mode (wakeup mode).
- the detection LD driver 122 controls lighting of the detection light source in accordance with a lighting control signal (LD enable (IR) signal) from the control unit 112.
- the display LD driver 111 controls lighting of the display light source according to a lighting control signal (LD enable (RGB) signal) from the control unit 112.
- the control unit 112 is composed of, for example, a microcomputer.
- the control unit 112 performs on / off control of the detection light source via the detection LD driver 122.
- the control unit 112 also performs on (lighting) / off (non-lighting) control of the display light source via the display LD driver 111.
- the detection light source and the display light source can be separately turned on (lighted) / off (not lighted) by the control unit 112.
- FIG. 8 shows an example of an operation flow of power saving control by the image processing / touch detection unit 123 of the detection unit 120.
- the image processing / touch detection unit 123 always counts the time since the object is detected.
- the threshold value of the time for shifting to the power saving mode is “t_sleep”.
- a sleep signal is output from the image processing / touch detection unit 123 to the control unit 112.
- the sleep signal is H (High)
- the sleep signal is L (Low)
- the sleep signal is switched from L to H when no object is detected for a predetermined period (time equal to or longer than t_sleep).
- the sleep signal is immediately set to L.
- the threshold value (t_sleep) for the time to shift to the power saving mode may be changeable according to the user's preference.
- the image processing / touch detection unit 123 first determines whether or not the sleep signal is L (step S11). Thereby, it is determined whether or not the status of the current operation mode is the power saving mode.
- the image processing / touch detection unit 123 determines that the current operation mode is not the power saving mode but the normal operation mode, and then the previous object It is determined whether or not the time since the detection is equal to or greater than a predetermined threshold (t_sleep) (step S12). If it is equal to or greater than the predetermined threshold (step S12; Y), the image processing / touch detection unit 123 switches the sleep signal from L to H so as to shift to the power saving mode and outputs it (step S14). If the predetermined threshold value has not been reached (step S12; N), the image processing / touch detection unit 123 keeps the sleep signal at L and continues the normal operation mode.
- t_sleep a predetermined threshold
- the image processing / touch detection unit 123 determines that the power saving mode is currently set, and then detects the previous object. It is determined whether or not the time from is less than a predetermined threshold (t_sleep) (step S13). If it is less than the predetermined threshold (step S13; Y), the image processing / touch detection unit 123 switches the sleep signal from H to L so as to cancel the power saving mode and shift to the normal operation mode. Output (step S15). If it is equal to or greater than the predetermined threshold (step S13; N), the image processing / touch detection unit 123 keeps the sleep signal at H and continues the power saving mode.
- t_sleep a predetermined threshold
- FIG. 9 shows an example of an operation flow of power saving control by the control unit 112 of the projection display unit 110.
- the control unit 112 performs on / off control of the display light source based on the sleep signal from the image processing / touch detection unit 123.
- the on / off control of the display light source is performed by outputting a lighting control signal (LD enable (RGB) signal) to the display LD driver 111.
- LD enable (RGB) signal when the LD enable (RGB) signal is H (High), it means lighting.
- the LD enable (RGB) signal is L (Low), it means that the light is turned off.
- control unit 112 first determines whether or not the LD enable (RGB) signal is H (step S21). Thereby, it is determined whether or not the display light source is currently lit.
- the control unit 112 determines that the display light source is currently in a normal operation mode, and then displays an image. It is determined whether or not the sleep signal from the processing / touch detection unit 123 is H (step S22). When the sleep signal is H (step S22; Y), the control unit 112 switches the LD enable (RGB) signal from H to L so as to shift to the power saving mode (step S24). As a result, the light source for display is turned off. When the sleep signal is not H (step S22; N), the control unit 112 keeps the LD enable (RGB) signal at H and continues the normal operation mode. Thereby, the lighting of the display light source continues.
- the control unit 112 determines that the display light source is currently in the power saving mode, and then It is determined whether or not the sleep signal from the image processing / touch detection unit 123 is L (step S23).
- the sleep signal is L (step S23; Y)
- the control unit 112 switches the LD enable (RGB) signal from L to H so as to shift to the normal operation mode and outputs it (step S25).
- the light source for display is turned on.
- the control unit 112 keeps the LD enable (RGB) signal at L and continues the power saving mode. Thereby, the extinguishing state of the light source for display continues.
- the projection display unit 110 uses the display light source. Since the light is turned off, power saving can be achieved. As a result, in a projection display device having a touch detection function, for example, the display light source can be turned on / off with an intuitive operation of touching the wall surface, which is the projection surface of the image, without touching the power switch or remote control of the main body, for example. Can control.
- FIG. 10 shows an example of an operation flow of power saving control by the image processing / touch detection unit 123 of the detection unit 120 in the projection display device of the present embodiment.
- the overall configuration of the optical system of the projection display apparatus may be substantially the same as that shown in FIG.
- the configuration of the control system may be substantially the same as in FIG.
- the operation flow of power saving control by the control unit 112 of the projection display unit 110 may be substantially the same as that in FIG.
- substantially the same operation as the operation flow of FIG. 8 may be performed.
- the transition between the power saving mode (sleep mode) and the normal operation mode (wakeup mode) is simply performed according to the presence or absence of object detection. You may make it perform electric power control.
- the image processing / touch detection unit 123 may have an advanced image processing function that analyzes not only the coordinates of the detected object but also the movement locus of the object.
- the gesture operation recognized by the image processing / touch detection unit 123 may be, for example, a double tap or a swipe.
- the image processing / touch detection unit 123 detects a power-saving gesture operation (sleep gesture) when the sleep signal is L in step S11 (step S11; Y). It may be determined whether or not (step S16). When the power saving gesture operation is detected (step S16; Y), the sleep signal may be switched from L to H so as to shift to the power saving mode (step S14). When the power saving gesture operation is not detected (step S16; N), the sleep signal may remain L and the normal operation mode may be continued.
- a power-saving gesture operation swipe gesture
- step S11 the image processing / touch detection unit 123 determines whether or not a release gesture operation (wakeup gesture) has been detected when the sleep signal is not L (step S11; N). It is also possible (step S17).
- the sleep signal may be switched from H to L so as to cancel the power saving mode and shift to the normal operation mode (step S15).
- the image processing / touch detection unit 123 may keep the sleep signal at H and continue the power saving mode.
- power saving control may be performed depending on whether or not an object is detected at a specific coordinate position.
- FIG. 11 shows an example of the configuration of a control system in the projection display apparatus of the present embodiment.
- FIG. 12 shows an example of an operation flow of power saving control by the control unit 112 of the projection display unit 110 in the present embodiment.
- the overall configuration of the optical system of the projection display apparatus may be substantially the same as that shown in FIG.
- the operation flow of power saving control by the image processing / touch detection unit 123 of the detection unit 120 may be substantially the same as that in FIG. 8 or FIG.
- the operation similar to the operation flow of FIG. 9 may be performed.
- the case where the power saving control target is only the display light source is taken as an example, but other portions may be the power saving control target.
- the image sensor 22 of the detection unit 120 and the infrared laser 121 that is a light source for detection may be targeted for power saving control.
- the control unit 112 may perform control so that the frame rate of the image sensor 22 in the power saving mode is lower than the frame rate in the normal operation mode.
- the frame rate in the normal operation mode may be 60 fps (frame per second), for example, and the frame rate in the power saving mode may be 15 fps, for example.
- the detection light source may be intermittently turned on in synchronization with the frame rate of the image sensor 22.
- the control unit 112 outputs an LD enable (RGB) signal by switching from H to L so as to shift to the power saving mode in step S24, and further controls the frame rate.
- the signal (Frame Rate signal) may be set to L and output (step S26). Thereby, the frame rate of the image sensor 22 in the power saving mode is lowered.
- control unit 112 may switch the LD enable (RGB) signal from L to H so as to shift to the normal operation mode, and may further output the frame rate signal by setting the frame rate signal to H ( Step S27). Thereby, the frame rate of the image sensor 22 in the normal operation mode becomes a normal state.
- LD enable RGB
- Step S27 the frame rate of the image sensor 22 in the normal operation mode becomes a normal state.
- FIG. 13 shows a configuration example of a control system in the projection display apparatus of the present embodiment.
- the overall configuration of the optical system of the projection display apparatus may be substantially the same as that shown in FIG.
- the operation flow of power saving control by the image processing / touch detection unit 123 of the detection unit 120 may be substantially the same as that in FIG. 8 or FIG.
- the operation flow of power saving control by the control unit 112 of the projection display unit 110 may be substantially the same as that in FIG. 9 or FIG.
- the detection of the object and the output of the sleep signal are performed by the detection unit 120.
- An image processing / touch detection unit 125 may be added to detect an object and output a sleep signal.
- the camera 124 may be installed, for example, so as to photograph the vicinity of the image projection plane.
- the image processing / touch detection unit 125 may perform object detection at a detection level that is necessary for power saving control.
- the operation of power saving control by the image processing / touch detection unit 125 may be substantially the same as in FIG.
- the present technology can take the following configurations.
- a projection display unit that has a display light source that emits light to be illumination light, and projects an image on a projection surface by the illumination light;
- a detection light source that emits detection light, and a detection unit that detects an object on or near the projection surface,
- the detection unit outputs, to the projection display unit, a power saving instruction signal instructing transition to a power saving mode according to a detection state of the object.
- the projection display unit turns off the display light source based on the power saving instruction signal.
- the detection unit When the object is detected after shifting to the power saving mode, the detection unit outputs a release instruction signal for instructing to release the power saving mode according to a detection state of the object. Output to The projection display unit turns on the light source for display based on the release instruction signal.
- the detection unit outputs the release instruction signal to the projection display unit when detecting a release gesture operation for instructing release of the power saving mode as the detection state of the object after transition to the power saving mode.
- the detection unit outputs the power saving instruction signal to the projection display unit at least when the detection of the object is not performed for a predetermined period. Any one of (1) to (4) Projection type display device with detection function.
- the detection unit outputs the power saving instruction signal to the projection display unit when detecting a power saving gesture operation that instructs to shift to the power saving mode as at least the detection state of the object. Thru
- the detection unit further includes an image sensor that detects the object, and lowers a frame rate of the image sensor in the power saving mode as compared with a frame rate in a case where the power saving mode is not set.
- the projection display device with a detection function according to any one of (6).
- the projection display unit modulates the illumination light based on video data, emits the modulated light, projects the modulated light from the light valve onto the projection plane, and transmits the modulated light.
- the detection unit further includes an image sensor arranged at a position optically conjugate with the light valve, The light that is diffused when the detection light hits the object is incident on the imaging element as detection light of the object through the projection lens.
- a projection display device with a detection function described in 1. (9) The detection light according to any one of (1) to (8), wherein the detection light is emitted so as to cover at least an area corresponding to a projection area of an image at a predetermined height from the projection plane.
- Projection type display device with function (10) The projection display device with a detection function according to any one of (1) to (9), wherein the detection light is infrared light.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Toxicology (AREA)
- Optics & Photonics (AREA)
- Projection Apparatus (AREA)
- Transforming Electric Information Into Light Information (AREA)
Abstract
Description
なお、ここに記載された効果は必ずしも限定されるものではなく、本開示中に記載されたいずれかの効果であってもよい。
1.第1の実施の形態(タッチ検出機能を有する投射型表示装置)
1.1 投射型表示装置の光学系の全体構成および動作(図1~図6)
1.1.1 投射型表示装置の光学系の全体構成例
1.1.2 投射型表示装置の基本動作
1.1.3 偏光子の作用
1.1.4 光学系の変形例(図6)
1.1.5 光学系のその他の変形例
1.2 制御系の構成および動作(図7~図9)
1.2.1 制御系の構成例
1.2.2 省電力制御の動作
1.3 効果
2.第2の実施の形態(ジェスチャー動作に応じた省電力機能を有する投射型表示装置)(図10)
2.1 制御系の構成および動作
3.第3の実施の形態(検出部のフレームレートを制御する省電力機能を有する投射型表示装置)(図11~図12)
3.1 制御系の構成および動作
4.第4の実施の形態(カメラを用いた省電力機能を有する投射型表示装置)(図13)
4.1 制御系の構成および動作
5.その他の実施の形態
[1.1 投射型表示装置の光学系の全体構成および動作]
(1.1.1 投射型表示装置の光学系の全体構成例)
図1は、本開示の第1の実施の形態に係る投射型表示装置(プロジェクタ)の光学系の全体構成の一例を示している。この投射型表示装置は、映像表示と共に、近赤外光を用いて、アクティブに物体検出を行う機能(タッチ検出機能)を有している。図2は、投射型表示装置において映像表示および物体検出をしている状態の一例を示している。図3は、図2に示した投射型表示装置を側面方向から見た状態の一例を示している。図4は、図1に示した投射型表示装置において、ライトバルブ21と撮像素子22とに入射する光の一例を示している。図5は、この投射型表示装置による映像表示および物体検出の概念を模式的に示している。
この投射型表示装置では、図1および図5に示したように、ライトバルブ21に形成された映像情報V1を投射レンズ24によってスクリーン30の投影面30A上に投影し、投影画像V2として拡大表示する。この投射型表示装置はまた、投影面30A上における物体の位置、例えば人の指やポインタ等の指示物(物体)71の特徴点の位置P1を撮像素子22を用いて検出する。撮像素子22は、投影面30A上の投影エリア31と略同一のエリアを撮影エリア32とした撮影を行う。
次に、図4を参照して、偏光子25Sの作用について説明する。ワイヤグリッド27に入射される検出光L2には偏光成分としてS偏光成分Ls3とP偏光成分Lp3とが含まれている。ワイヤグリッド27は、P偏光成分Lp3の多くを第3の方向Z3に反射する。偏光子25Sを、S偏光成分を除去するものとすれば、反射されたP偏光成分Lp3のほぼすべてが撮像素子22に到達する。また、ワイヤグリッド27に入射される照明光L1のうち、S偏光成分Ls1が第3の方向Z3に出射される。このS偏光成分Ls1は、検出光L2に対してノイズ成分となり、撮像素子22に入射すると、検出の際のS/N比が小さくなり、検出精度を悪化させる。偏光子25Sを配置してS偏光成分Ls1を除去することで、S/N比を大きくして検出精度を上げることができる。
図1および図4には、偏光分離素子としてワイヤグリッド27を用いた構成例を示したが、図6に示したようにワイヤグリッド27に代えて偏光ビームスプリッタ23を用いた構成であっても良い。また、この変形例では、S偏光成分を除去する偏光子25Sに代えてP偏光成分を除去する偏光子25を備えている。
図1の構成における照明部1において、第1のフライアイレンズ151および第2のフライアイレンズ152は、どちらか一方であってもよい。第2のフライアイレンズ152のみにする場合、第1のコンデンサレンズ161および第2のコンデンサレンズ162は不要となる。第1のフライアイレンズ151のみにする場合、第3のコンデンサレンズ163および第4のコンデンサレンズ164は不要となる。
次に、本実施の形態の投射型表示装置における制御系の構成および動作を説明する。なお、以下では、上述の光学系と同様の構成および作用を有する部分については、同一の符号を付し、適宜説明を省略する。
図7は、本実施の形態の投射型表示装置における制御系の一構成例を示している。
本実施の形態に係る投射型表示装置は、上述したように、照明光L1によって映像を投影面30Aに投影する部分(投射表示部110)と、投影面30A上もしくは投影面30A近傍における物体を検出する部分(検出部120)とを備えている。
次に、省電力制御に関する動作例を説明する。
本実施の形態に係る投射型表示装置では、所定の期間(例えば10分程度)、タッチ入力(物体の検出)がなされなかった場合に、省電力モード(sleepモード)へ移行して表示用光源を消灯することで省電力化を図る。その一方で、検出用光源は省電力モードにおいても点灯させ続けることで、タッチ入力の監視動作(物体の検出動作)は続行できるようにする。これにより、表示用光源が消灯して投影面30Aに映像が表示されていない状態であっても、タッチ入力があった場合には、再び表示用光源を点灯して投影面30Aに映像を表示できる。以下、図8および図9を参照して、このような省電力制御を行う場合の動作例を説明する。
以上のように、本実施の形態によれば、物体の検出状態に応じて検出部120から出力された省電力指示信号(sleep信号=H)に基づいて、投射表示部110が表示用光源を消灯させるようにしたので、省電力化を図ることができる。これにより、タッチ検出機能を有する投射型表示装置において、例えば本体の電源スイッチやリモコン等を触ることなく、映像の投射面である壁面等を触るという直感的な動作で表示用光源のオン/オフ制御ができる。
次に、本開示の第2の実施の形態について説明する。以下では、上記第1の実施の形態と同様の構成および作用を有する部分については、適宜説明を省略する。
図10は、本実施の形態の投射型表示装置における検出部120の画像処理/タッチ検出部123による省電力制御の動作フローの一例を示している。なお、本実施の形態において、投射型表示装置の光学系の全体構成は図1と略同様であっても良い。また、制御系の構成は図7と略同様であっても良い。また、投射表示部110の制御部112による省電力制御の動作フローは、図9と略同様であっても良い。また、図10の動作フローにおいて、図8と同一のステップ番号を付した部分では、図8の動作フローと略同様の動作を行ってもよい。
次に、本開示の第3の実施の形態について説明する。以下では、上記第1または第2の実施の形態と同様の構成および作用を有する部分については、適宜説明を省略する。
図11は、本実施の形態の投射型表示装置における制御系の一構成例を示している。図12は、本実施の形態における投射表示部110の制御部112による省電力制御の動作フローの一例を示している。なお、本実施の形態において、投射型表示装置の光学系の全体構成は図1と略同様であっても良い。また、検出部120の画像処理/タッチ検出部123による省電力制御の動作フローは、図8または図10と略同様であっても良い。また、図12の動作フローにおいて、図9と同一のステップ番号を付した部分では、図9の動作フローと略同様の動作を行ってもよい。
次に、本開示の第4の実施の形態について説明する。以下では、上記第1ないし第3の実施の形態と同様の構成および作用を有する部分については、適宜説明を省略する。
図13は、本実施の形態の投射型表示装置における制御系の一構成例を示している。なお、本実施の形態において、投射型表示装置の光学系の全体構成は図1と略同様であっても良い。また、検出部120の画像処理/タッチ検出部123による省電力制御の動作フローは、図8または図10と略同様であっても良い。また、投射表示部110の制御部112による省電力制御の動作フローは、図9または図12と略同様であっても良い。
本開示による技術は、上記各実施の形態および変形例の説明に限定されず種々の変形実施が可能である。
(1)
照明光となる光を発する表示用光源を有し、前記照明光によって映像を投影面に投影する投射表示部と、
検出用の光を発する検出用光源を有し、前記投影面上もしくは前記投影面近傍における物体を検出する検出部と
を備え、
前記検出部は、前記物体の検出状態に応じて、省電力モードへの移行を指示する省電力指示信号を前記投射表示部に出力し、
前記投射表示部は、前記省電力指示信号に基づいて、前記表示用光源を消灯させる
検出機能付き投射型表示装置。
(2)
前記検出部は、前記省電力モードへの移行後においても、前記検出用光源を点灯させ、前記物体の検出を続行する
上記(1)に記載の検出機能付き投射型表示装置。
(3)
前記検出部は、前記省電力モードへの移行後に前記物体の検出がなされた場合には、前記物体の検出状態に応じて、前記省電力モードの解除を指示する解除指示信号を前記投射表示部に出力し、
前記投射表示部は、前記解除指示信号に基づいて、前記表示用光源を点灯させる
上記(2)に記載の検出機能付き投射型表示装置。
(4)
前記検出部は、前記省電力モードへの移行後に、前記物体の検出状態として前記省電力モードの解除を指示する解除ジェスチャー動作を検出した場合に、前記解除指示信号を前記投射表示部に出力する
上記(3)に記載の検出機能付き投射型表示装置。
(5)
前記検出部は、少なくとも、所定の期間、前記物体の検出がなされなかった場合に、前記省電力指示信号を前記投射表示部に出力する
上記(1)ないし(4)のいずれか1つに記載の検出機能付き投射型表示装置。
(6)
前記検出部は、少なくとも、前記物体の検出状態として前記省電力モードへの移行を指示する省電力ジェスチャー動作を検出した場合に、前記省電力指示信号を前記投射表示部に出力する
上記(1)ないし(5)のいずれか1つに記載の検出機能付き投射型表示装置。
(7)
前記検出部は、前記物体を検出する撮像素子をさらに有し、前記省電力モードにおける前記撮像素子のフレームレートを、前記省電力モードではない場合のフレームレートに比べて低くする
上記(1)ないし(6)のいずれか1つに記載の検出機能付き投射型表示装置。
(8)
前記投射表示部は、前記照明光を映像データに基づいて変調し、その変調光を出射するライトバルブと、前記ライトバルブからの前記変調光を前記投影面上に投影すると共に、前記変調光の進行方向とは逆方向から前記物体の検出光が入射される投射レンズとをさらに有し、
前記検出部は、前記ライトバルブと光学的に共役な位置に配置された撮像素子をさらに有し、
前記撮像素子には、前記検出用の光が前記物体に当たって拡散された光が、前記物体の検出光として、前記投射レンズを介して入射される
上記(1)ないし(7)のいずれか1つに記載の検出機能付き投射型表示装置。
(9)
前記検出用の光は、前記投影面から所定の高さにおいて、少なくとも映像の投影エリアに対応するエリアを覆うように出射される
上記(1)ないし(8)のいずれか1つに記載の検出機能付き投射型表示装置。
(10)
前記検出用の光は、赤外光である
上記(1)ないし(9)のいずれか1つに記載の検出機能付き投射型表示装置。
Claims (10)
- 照明光となる光を発する表示用光源を有し、前記照明光によって映像を投影面に投影する投射表示部と、
検出用の光を発する検出用光源を有し、前記投影面上もしくは前記投影面近傍における物体を検出する検出部と
を備え、
前記検出部は、前記物体の検出状態に応じて、省電力モードへの移行を指示する省電力指示信号を前記投射表示部に出力し、
前記投射表示部は、前記省電力指示信号に基づいて、前記表示用光源を消灯させる
検出機能付き投射型表示装置。 - 前記検出部は、前記省電力モードへの移行後においても、前記検出用光源を点灯させ、前記物体の検出を続行する
請求項1に記載の検出機能付き投射型表示装置。 - 前記検出部は、前記省電力モードへの移行後に前記物体の検出がなされた場合には、前記物体の検出状態に応じて、前記省電力モードの解除を指示する解除指示信号を前記投射表示部に出力し、
前記投射表示部は、前記解除指示信号に基づいて、前記表示用光源を点灯させる
請求項2に記載の検出機能付き投射型表示装置。 - 前記検出部は、前記省電力モードへの移行後に、前記物体の検出状態として前記省電力モードの解除を指示する解除ジェスチャー動作を検出した場合に、前記解除指示信号を前記投射表示部に出力する
請求項3に記載の検出機能付き投射型表示装置。 - 前記検出部は、少なくとも、所定の期間、前記物体の検出がなされなかった場合に、前記省電力指示信号を前記投射表示部に出力する
請求項1に記載の検出機能付き投射型表示装置。 - 前記検出部は、少なくとも、前記物体の検出状態として前記省電力モードへの移行を指示する省電力ジェスチャー動作を検出した場合に、前記省電力指示信号を前記投射表示部に出力する
請求項1に記載の検出機能付き投射型表示装置。 - 前記検出部は、前記物体を検出する撮像素子をさらに有し、前記省電力モードにおける前記撮像素子のフレームレートを、前記省電力モードではない場合のフレームレートに比べて低くする
請求項1に記載の検出機能付き投射型表示装置。 - 前記投射表示部は、前記照明光を映像データに基づいて変調し、その変調光を出射するライトバルブと、前記ライトバルブからの前記変調光を前記投影面上に投影すると共に、前記変調光の進行方向とは逆方向から前記物体の検出光が入射される投射レンズとをさらに有し、
前記検出部は、前記ライトバルブと光学的に共役な位置に配置された撮像素子をさらに有し、
前記撮像素子には、前記検出用の光が前記物体に当たって拡散された光が、前記物体の検出光として、前記投射レンズを介して入射される
請求項1に記載の検出機能付き投射型表示装置。 - 前記検出用の光は、前記投影面から所定の高さにおいて、少なくとも映像の投影エリアに対応するエリアを覆うように出射される
請求項1に記載の検出機能付き投射型表示装置。 - 前記検出用の光は、赤外光である
請求項1に記載の検出機能付き投射型表示装置。
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2016546389A JP6586956B2 (ja) | 2014-09-03 | 2015-08-05 | 検出機能付き投射型表示装置 |
| US15/505,692 US10805586B2 (en) | 2014-09-03 | 2015-08-05 | Projection display unit with detection function |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2014-178963 | 2014-09-03 | ||
| JP2014178963 | 2014-09-03 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2016035500A1 true WO2016035500A1 (ja) | 2016-03-10 |
Family
ID=55439568
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2015/072258 Ceased WO2016035500A1 (ja) | 2014-09-03 | 2015-08-05 | 検出機能付き投射型表示装置 |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US10805586B2 (ja) |
| JP (1) | JP6586956B2 (ja) |
| WO (1) | WO2016035500A1 (ja) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN107508976A (zh) * | 2017-08-15 | 2017-12-22 | 上海青橙实业有限公司 | 操作方法及可投影移动终端 |
| CN108471525A (zh) * | 2018-03-27 | 2018-08-31 | 百度在线网络技术(北京)有限公司 | 用于投影仪的控制方法和装置 |
| JPWO2018173584A1 (ja) * | 2017-03-23 | 2020-01-23 | ソニー株式会社 | ビーム照射装置、および検出機能付きプロジェクタ |
Families Citing this family (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2017130698A1 (ja) * | 2016-01-28 | 2017-08-03 | 日立マクセル株式会社 | 撮像装置 |
| JP2019174513A (ja) * | 2018-03-27 | 2019-10-10 | セイコーエプソン株式会社 | 表示装置、及び、表示装置の制御方法 |
| CN110223619A (zh) * | 2019-06-11 | 2019-09-10 | 上海易视计算机科技股份有限公司 | 投影控制方法、装置、滤光片及投影系统 |
| WO2024082885A1 (zh) * | 2022-10-17 | 2024-04-25 | 青岛海信激光显示股份有限公司 | 投影系统及其控制方法 |
| CN118678032A (zh) * | 2023-03-17 | 2024-09-20 | 昆山扬皓光电有限公司 | 投影设备及其控制方法 |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH06236236A (ja) * | 1992-11-19 | 1994-08-23 | Sextant Avionique | スクリーン表示、およびスクリーンに対する位置マーカの位置の検出を兼用している装置 |
| JP2010197440A (ja) * | 2009-02-23 | 2010-09-09 | Seiko Epson Corp | プロジェクター、その制御方法及びプログラム |
| JP2011197645A (ja) * | 2010-02-24 | 2011-10-06 | Sanyo Electric Co Ltd | 投写型映像表示装置 |
| JP2012093767A (ja) * | 2011-11-28 | 2012-05-17 | Olympus Corp | 情報表示装置 |
| JP2013068813A (ja) * | 2011-09-22 | 2013-04-18 | Casio Comput Co Ltd | 投影装置、投影制御方法及びプログラム |
| JP2013115649A (ja) * | 2011-11-29 | 2013-06-10 | Toshiba Corp | 制御装置、電力量制御方法 |
| JP2013182061A (ja) * | 2012-02-29 | 2013-09-12 | Nikon Corp | 投影システム |
Family Cites Families (32)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP4090939B2 (ja) * | 2002-05-29 | 2008-05-28 | ニッタ株式会社 | 静電容量式センサおよびその製造方法 |
| WO2003104892A1 (ja) * | 2002-06-10 | 2003-12-18 | ソニー株式会社 | 画像投射装置及び画像投射方法 |
| US7567366B2 (en) * | 2003-05-29 | 2009-07-28 | Seiko Epson Corporation | Image scanner provided with power saving mode and a system having a power saving mode |
| TWI232346B (en) * | 2004-01-09 | 2005-05-11 | Coretronic Corp | A method and apparatus for an anti-melt projection lens cap |
| US20070165409A1 (en) * | 2004-01-28 | 2007-07-19 | Yusaku Shimaoka | Projection display and image display method |
| US9229540B2 (en) * | 2004-01-30 | 2016-01-05 | Electronic Scripting Products, Inc. | Deriving input from six degrees of freedom interfaces |
| JP4180003B2 (ja) * | 2004-03-24 | 2008-11-12 | 三洋電機株式会社 | 投射型映像表示装置 |
| US20080055318A1 (en) * | 2006-08-31 | 2008-03-06 | Glen David I J | Dynamic frame rate adjustment |
| TWI332647B (en) * | 2007-11-20 | 2010-11-01 | Au Optronics Corp | Liquid crystal display device with dynamically switching driving method to reduce power consumption |
| JP5553022B2 (ja) * | 2008-08-22 | 2014-07-16 | ソニー株式会社 | 画像表示装置、制御方法およびコンピュータプログラム |
| US9046935B2 (en) * | 2008-10-29 | 2015-06-02 | Pixart Imaging Inc. | Motion estimation device and power saving method thereof |
| US20110103643A1 (en) * | 2009-11-02 | 2011-05-05 | Kenneth Edward Salsman | Imaging system with integrated image preprocessing capabilities |
| WO2011058528A1 (en) * | 2009-11-15 | 2011-05-19 | Ram Friedlander | An enhanced pointing interface |
| TW201133121A (en) * | 2010-03-31 | 2011-10-01 | Acer Inc | Projection device and projection light source control method thereof |
| US8586925B2 (en) * | 2010-06-03 | 2013-11-19 | Jeremy P. Willden | Ultra-low-power occupancy sensor |
| JP2012256000A (ja) * | 2011-06-10 | 2012-12-27 | Sanyo Electric Co Ltd | 投写型映像表示装置 |
| JP2013003859A (ja) * | 2011-06-16 | 2013-01-07 | Sony Corp | 投影装置、投影方法、及びプログラム |
| US9498231B2 (en) * | 2011-06-27 | 2016-11-22 | Board Of Regents Of The University Of Nebraska | On-board tool tracking system and methods of computer assisted surgery |
| JP2013122501A (ja) * | 2011-12-09 | 2013-06-20 | Sanyo Electric Co Ltd | 投写型映像表示装置、投写型映像表示システムおよび電子機器 |
| US10191363B2 (en) * | 2012-02-09 | 2019-01-29 | Maxell, Ltd. | Projection-type image display device |
| US20140208957A1 (en) * | 2012-02-14 | 2014-07-31 | Panasonic Corporation | Electronic device |
| US8666361B2 (en) * | 2012-07-17 | 2014-03-04 | Blackberry Limited | System and method of providing information access on a portable device |
| US10105149B2 (en) * | 2013-03-15 | 2018-10-23 | Board Of Regents Of The University Of Nebraska | On-board tool tracking system and methods of computer assisted surgery |
| JP2014186610A (ja) * | 2013-03-25 | 2014-10-02 | Sony Corp | 情報処理装置および記憶媒体 |
| JP5813690B2 (ja) * | 2013-04-08 | 2015-11-17 | シャープ株式会社 | 座標位置検出装置及び表示入力装置 |
| US9658688B2 (en) * | 2013-10-15 | 2017-05-23 | Microsoft Technology Licensing, Llc | Automatic view adjustment |
| US9791708B2 (en) * | 2014-06-10 | 2017-10-17 | Seiko Epson Corporation | Display apparatus |
| US20160086574A1 (en) * | 2014-09-19 | 2016-03-24 | Pixtronix, Inc. | Adaptive flicker control |
| US9753136B2 (en) * | 2015-02-11 | 2017-09-05 | Motorola Mobility Llc | Portable electronic device with proximity sensors for gesture control and contact detection |
| US9983717B2 (en) * | 2015-04-21 | 2018-05-29 | Dell Products L.P. | Disambiguation of false touch inputs at an information handling system projected user interface |
| US10156901B2 (en) * | 2015-10-16 | 2018-12-18 | Wisconsin Alumni Research Foundation | Touch surface for mobile devices using near field light sensing |
| US10827053B2 (en) * | 2018-05-21 | 2020-11-03 | Metropolitan Industries, Inc. | Message projector |
-
2015
- 2015-08-05 US US15/505,692 patent/US10805586B2/en active Active
- 2015-08-05 JP JP2016546389A patent/JP6586956B2/ja active Active
- 2015-08-05 WO PCT/JP2015/072258 patent/WO2016035500A1/ja not_active Ceased
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH06236236A (ja) * | 1992-11-19 | 1994-08-23 | Sextant Avionique | スクリーン表示、およびスクリーンに対する位置マーカの位置の検出を兼用している装置 |
| JP2010197440A (ja) * | 2009-02-23 | 2010-09-09 | Seiko Epson Corp | プロジェクター、その制御方法及びプログラム |
| JP2011197645A (ja) * | 2010-02-24 | 2011-10-06 | Sanyo Electric Co Ltd | 投写型映像表示装置 |
| JP2013068813A (ja) * | 2011-09-22 | 2013-04-18 | Casio Comput Co Ltd | 投影装置、投影制御方法及びプログラム |
| JP2012093767A (ja) * | 2011-11-28 | 2012-05-17 | Olympus Corp | 情報表示装置 |
| JP2013115649A (ja) * | 2011-11-29 | 2013-06-10 | Toshiba Corp | 制御装置、電力量制御方法 |
| JP2013182061A (ja) * | 2012-02-29 | 2013-09-12 | Nikon Corp | 投影システム |
Cited By (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPWO2018173584A1 (ja) * | 2017-03-23 | 2020-01-23 | ソニー株式会社 | ビーム照射装置、および検出機能付きプロジェクタ |
| EP3605294A4 (en) * | 2017-03-23 | 2020-02-26 | Sony Corporation | RADIATION RADIATION DEVICE AND PROJECTOR WITH DETECTION FUNCTION |
| JP7143840B2 (ja) | 2017-03-23 | 2022-09-29 | ソニーグループ株式会社 | ビーム照射装置、および検出機能付きプロジェクタ |
| CN107508976A (zh) * | 2017-08-15 | 2017-12-22 | 上海青橙实业有限公司 | 操作方法及可投影移动终端 |
| CN108471525A (zh) * | 2018-03-27 | 2018-08-31 | 百度在线网络技术(北京)有限公司 | 用于投影仪的控制方法和装置 |
| CN108471525B (zh) * | 2018-03-27 | 2020-07-17 | 百度在线网络技术(北京)有限公司 | 用于投影仪的控制方法和装置以及实现该方法的投影仪 |
Also Published As
| Publication number | Publication date |
|---|---|
| US20170251187A1 (en) | 2017-08-31 |
| JP6586956B2 (ja) | 2019-10-09 |
| US10805586B2 (en) | 2020-10-13 |
| JPWO2016035500A1 (ja) | 2017-06-22 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP6586956B2 (ja) | 検出機能付き投射型表示装置 | |
| JP6128008B2 (ja) | 投射型表示装置 | |
| US10228611B2 (en) | Projector, projection system, and control method of projector | |
| JP6696425B2 (ja) | 投射型表示装置 | |
| JP6372266B2 (ja) | 投射型表示装置および機能制御方法 | |
| CN106537248B (zh) | 投影型显示装置 | |
| JP2012256000A (ja) | 投写型映像表示装置 | |
| JP6669068B2 (ja) | 投射型表示装置 | |
| JP7014009B2 (ja) | 操作デバイス、位置検出システム及び操作デバイスの制御方法 | |
| JP6806220B2 (ja) | 電子デバイス、及び電子デバイスの制御方法 | |
| JP6634904B2 (ja) | 電子デバイス、及び電子デバイスの制御方法 | |
| US20120176304A1 (en) | Projection display apparatus | |
| JP6287432B2 (ja) | 操作デバイス、位置検出システム、及び、操作デバイスの制御方法 | |
| JP7010076B2 (ja) | 画像投写システム、プロジェクター及び画像投写システムの制御方法 | |
| JP2012078641A (ja) | プロジェクタ | |
| JP6988985B2 (ja) | 電子デバイス、及び電子デバイスの制御方法 | |
| JP2012108232A (ja) | 電子機器 | |
| JP2013041214A (ja) | 画像投影装置 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15838124 Country of ref document: EP Kind code of ref document: A1 |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 15505692 Country of ref document: US |
|
| ENP | Entry into the national phase |
Ref document number: 2016546389 Country of ref document: JP Kind code of ref document: A |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 15838124 Country of ref document: EP Kind code of ref document: A1 |