WO2025207596A1 - Activation préemptive de caméra - Google Patents
Activation préemptive de caméraInfo
- Publication number
- WO2025207596A1 WO2025207596A1 PCT/US2025/021290 US2025021290W WO2025207596A1 WO 2025207596 A1 WO2025207596 A1 WO 2025207596A1 US 2025021290 W US2025021290 W US 2025021290W WO 2025207596 A1 WO2025207596 A1 WO 2025207596A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- camera
- screen
- computing system
- input
- area
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/69—Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/45—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/62—Control of parameters via user interfaces
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
Definitions
- image capturing devices Many modern computing devices, including mobile phones, personal computers, and tablets, include image capturing devices. Some image capturing devices are configured with telephoto capabilities.
- a method includes displaying on a screen a representation of an environment based on primary sensor data captured by a primary camera.
- the method also includes detecting at least one input associated with an area of the screen.
- the method additionally includes, based on the at least one input associated with the screen area, determining a secondary camera to preemptively activate in anticipation of zooming in or out.
- the method further includes preemptively activating the secondary camera.
- the method also includes, based on further associated with the screen area, zooming the displayed representation of the environment in or out based on secondary sensor data captured by the secondary camera.
- a computing system in another embodiment, includes a control system.
- the control system is configured to display on a screen a representation of an environment based on primary sensor data captured by a primary camera.
- the control system is also configured to detect at least one input associated with an area of the screen.
- the control system is additionally configured to, based on the at least one input associated with the screen area, determining a secondary camera to preemptively activate in anticipation of zooming in or out.
- the control system is further configured to preemptively activate the secondary camera.
- the control system is additionally configured to, based on further input associated with the screen area, zooming the displayed representation of the environment in or out based on secondary sensor data captured by the secondary camera.
- a system in another embodiment, includes means for displaying on a screen a representation of an environment based on primary sensor data captured by a primary camera.
- the system also includes means for detecting at least one input associated with an area of the screen.
- the system additionally includes means for, based on the at least one input associated with the screen area, determining a secondary camera to preemptively activate in anticipation of zooming in or out.
- the system further includes means for preemptively activating the secondary camera.
- the system also includes means for, based on further input associated with the screen area, zooming the displayed representation of the environment in or out based on secondary sensor data captured by the secondary camera.
- Figure 1 illustrates an example computing device, in accordance with example embodiments.
- Figure 4 depicts an image, in accordance with example embodiments.
- Figure 7 depicts a graph, in accordance with example embodiments.
- the computing system may use various other inputs as an indication to zoom in and/or out (e.g., detecting one or more faces that are in the distance as an indication that an input to zoom in may follow, determining that one or more objects take up a large portion of the frame as an indication that an input to zoom out may follow, etc.).
- various other inputs e.g., detecting one or more faces that are in the distance as an indication that an input to zoom in may follow, determining that one or more objects take up a large portion of the frame as an indication that an input to zoom out may follow, etc.
- a user may use one finger to tap an area of the screen, and the user may subsequently move another finger or a plurality of fingers over the screen to zoom in and/or out.
- the computing system may infer that the user intends to zoom in and/or out based on the initial one finger tap and subsequently select a secondary camera to activate.
- the computing system may detect the screen area associated with the one finger tap. Based on the associated screen area, the computing system may infer whether the user intends to zoom in or out and select an appropriate secondary camera to activate.
- the computing system may store the location of an element on a graphical user interface, where tapping on the element of the graphical user interface causes the appearance of a slider through which the user may indicate how much to zoom in and/or out.
- the computing system may determine whether the screen area associated with the initial user input corresponds with the location of the element to determine whether to activate a secondary camera.
- the computing system may use further user input to confirm that the secondary camera was accurately selected.
- the computing system may have activated a zoom lens, and after activating the zoom lens, the computing system may detect a user dragging both fingers such that the distance between the areas of the screen associated with the further user input increases.
- the further user input may thus confirm that the computing system accurately selected a secondary camera.
- the further user input may indicate that the computing system did not accurately select a secondary camera.
- the computing system may have activated a wide angle lens, and after activating the wide angle lens, the computing system may detect a user dragging both fingers as to zoom. The computing system may then deactivate the wide angle lens and activate the zoom lens.
- the rear facing cameras may include a wide angle camera, a main camera, and a telephoto camera.
- the wide angle camera may capture a larger portion of the environment compared to the main camera and the telephoto camera, and the telephoto camera may capture more detailed images of a smaller portion of the environment compared to the main camera and the wide angle camera.
- Front-facing camera 104 may be positioned on a side of body 102 typically facing a user while in operation (e.g., on the same side as display 106).
- Rear-facing camera 112 may be positioned on a side of body 102 opposite front-facing camera 104 Referring to the cameras as front and rear facing is arbitrary, and computing device 100 may include multiple cameras positioned on various sides of body 102.
- Display 106 could represent a cathode ray tube (CRT) display, a light emitting diode (LED) display, a liquid crystal (LCD) display, a plasma display, an organic light emitting diode (OLED) display, or any other type of display known in the art.
- display 106 may display a digital representation of the current image being captured by front-facing camera 104 and/or rear-facing camera 112, an image that could be captured by one or more of these cameras, an image that was recently captured by one or more of these cameras, and/or a modified version of one or more of these images.
- display 106 may serve as a viewfinder for the cameras.
- Display 106 may also support touchscreen functions that may be able to adjust the settings and/or configuration of one or more aspects of computing device 100.
- Front-facing camera 104 may include an image sensor and associated optical elements such as lenses. Front-facing camera 104 may offer zoom capabilities or could have a fixed focal length. In other examples, interchangeable lenses could be used with front-facing camera 104. Front-facing camera 104 may have a variable mechanical aperture and a mechanical and/or electronic shutter. Front-facing camera 104 also could be configured to capture still images, video images, or both. Further, front-facing camera 104 could represent, for example, a monoscopic, stereoscopic, or multiscopic camera. Rear-facing camera 112 may be similarly or differently arranged. Additionally, one or more of front-facing camera 104 and/or rear-facing camera 112 may be an array of one or more cameras.
- One or more of front-facing camera 104 and/or rear-facing camera 112 may include or be associated with an illumination component that provides a light field to illuminate a target object.
- an illumination component could provide flash or constant illumination of the target object.
- An illumination component could also be configured to provide a light field that includes one or more of structured light, polarized light, and light with specific spectral content. Other types of light fields known and used to recover three-dimensional (3D) models from an object are possible within the context of the examples herein.
- Computing device 100 could be configured to use display 106 and front-facing camera 104 and/or rear-facing camera 112 to capture images of a target object.
- the captured images could be a plurality of still images or a video stream.
- the image capture could be triggered by activating button 108, pressing a softkey on display 106, or by some other mechanism.
- the images could be captured automatically at a specific time interval, for example, upon pressing button 108, upon appropriate lighting conditions of the target object, upon moving computing device 100 a predetermined distance, or according to a predetermined capture schedule.
- FIG. 2 is a simplified block diagram showing some of the components of an example computing system 200, such as an image capturing device and/or a video capturing device.
- computing system 200 may be a cellular mobile telephone (e g., a smartphone), a computer (such as a desktop, notebook, tablet, server, or handheld computer), a home automation component, a digital video recorder (DVR), a digital television, a remote control, a wearable computing device, a gaming console, a robotic device, a vehicle, or some other type of device.
- Computing system 200 may represent, for example, aspects of computing device 100.
- Communication interface 202 may allow computing system 200 to communicate, using analog or digital modulation, with other devices, access networks, and/or transport networks.
- communication interface 202 may facilitate circuit-switched and/or packet-switched communication, such as plain old telephone service (POTS) communication and/or Internet protocol (IP) or other packetized communication.
- POTS plain old telephone service
- IP Internet protocol
- communication interface 202 may include a chipset and antenna arranged for wireless communication with a radio access network or an access point.
- communication interface 202 may take the form of or include a wireline interface, such as an Ethernet, Universal Serial Bus (USB), or High-Definition Multimedia Interface (HDMI) port, among other possibilities.
- USB Universal Serial Bus
- HDMI High-Definition Multimedia Interface
- User interface 204 may function to allow computing system 200 to interact with a human or non-human user, such as to receive input from a user and to provide output to the user.
- user interface 204 may include input components such as a keypad, keyboard, touch-sensitive panel, computer mouse, trackball, joystick, microphone, and so on.
- User interface 204 may also include one or more output components such as a display screen, which, for example, may be combined with a touch-sensitive panel.
- the display screen may be based on CRT, LCD, LED, and/or OLED technologies, or other technologies now known or later developed.
- User interface 204 may also be configured to generate audible output(s), via a speaker, speaker jack, audio output port, audio output device, earphones, and/or other similar devices. User interface 204 may also be configured to receive and/or capture audible utterance(s), noise(s), and/or signal(s) by way of a microphone and/or other similar devices.
- user interface 204 may include a display that serves as a viewfinder for still camera and/or video camera functions supported by computing system 200. Additionally, user interface 204 may include one or more buttons, switches, knobs, and/or dials that facilitate the configuration and focusing of a camera function and the capturing of images. It may be possible that some or all of these buttons, switches, knobs, and/or dials are implemented by way of a touch-sensitive panel.
- Processor 206 may comprise one or more general purpose processors - e.g., microprocessors - and/or one or more special purpose processors - e.g., digital signal processors (DSPs), graphics processing units (GPUs), floating point units (FPUs), network processors, or application-specific integrated circuits (ASICs).
- DSPs digital signal processors
- GPUs graphics processing units
- FPUs floating point units
- ASICs application-specific integrated circuits
- special purpose processors may be capable of image processing, image alignment, and merging images, among other possibilities.
- Data storage 208 may include one or more volatile and/or non-volatile storage components, such as magnetic, optical, flash, or organic storage, and may be integrated in whole or in part with processor 206.
- Data storage 208 may include removable and/or non-removable components.
- program instructions 218 may include an operating system 222 (e.g., an operating system kernel, device driver(s), and/or other modules) and one or more application programs 220 (e.g., camera functions, address book, email, web browsing, social networking, audio-to-text functions, text translation functions, and/or gaming applications) installed on computing system 200.
- data 212 may include operating system data 216 and application data 214.
- Operating system data 216 may be accessible primarily to operating system 222
- application data 214 may be accessible primarily to one or more of application programs 220.
- Application data 214 may be arranged in a file system that is visible to or hidden from a user of computing system 200.
- Application programs 220 may communicate with operating system 222 through one or more application programming interfaces (APIs). These APIs may facilitate, for instance, application programs 220 reading and/or writing application data 214, transmitting or receiving information via communication interface 202, receiving and/or displaying information on user interface 204, and so on.
- APIs application programming interfaces
- Histogram processing algorithm(s) 226 may include one or more stored algorithms programmed to process histogram information to facilitate autofocus as described herein. In some examples, histogram processing algorithm(s) 226 may include one or more trained machine learning models. In other examples, histogram processing algorithm(s) 226 may be based on heuristics without the use of machine learning. In further examples, a combination of different types of histogram processing algorithm(s) 226 may be used as well.
- one or more remote cameras 230 may be controlled by computing system 200.
- computing system 200 may transmit control signals to the one or more remote cameras 230 through a wireless or wired connection. Such signals may be transmitted as part of an ambient computing environment.
- inputs received at the computing system 200 for instance, physical movements of a wearable device
- images captured by the one or more remote cameras 230 may be transmitted to the computing system 200 for further processing. Such images may be treated as images captured by cameras physically located on the computing system 200.
- method 300 may be executed by a system that includes a mobile device.
- the mobile device may include one or more cameras and/or one or more computing systems. Additionally and/or alternatively, the mobile device may communicate with one or more sensors and/or one or more cameras remote from the device. The mobile device may receive information transmitted by the sensors and/or cameras, and the mobile device may then use the information for the process described herein.
- method 300 includes displaying on a screen a representation of an environment based on primary sensor data captured by a primary camera.
- a computing system may include a screen, such as display 106 discussed in the context of Figure 1 or a display that is wirelessly connected to a processor of the computing system.
- the computing system may also include a camera, such as camera 104 or camera 112 discussed in the context of Figure 1 or a camera that is wirelessly connected to the processor of the computing system.
- the camera of the computing system may collect sensor data of the environment and transmit the collected sensor data to the computing system.
- the computing system may receive the sensor data and transmit the sensor data to a screen to display.
- method 300 includes detecting at least one input associated with an area of the screen.
- the input may be a user input, such as tapping or touching the screen with one or more fingers.
- the input may also be clicking an area of the screen or otherwise indicating an area of the screen.
- the computing system may determine whether the input is likely indicative of future input based on a distance of inputs. For example, if the computing system receives a plurality of inputs, the computing system may calculate the distance between the inputs and determine whether the input is indicative of further input based on the distance being greater or lesser than a threshold value.
- a threshold high value may indicate a threshold distance over which a calculated distance is indicated to be disregarded and a threshold low value may indicate a threshold distance over which a calculated distance is also indicated to be disregarded.
- the input may be on a screen of the computing system while the screen is displaying an image as described above.
- Figure 5 depicts image 500, in accordance with example embodiments.
- the computing system may display image 500 on a screen.
- the computing system may display a stream of images as a preview, and as such, image 500 may be a different or the same image as image 400.
- image 500 While image 500 is being displayed on a screen of the computing system, the computing system may detect input 502 and additional input 504, which correspond to the indicated areas on image 500.
- Input 502 and additional input 504 may be indicated by a user tapping a screen using two fingers, clicking two areas of a screen, programmatically through two sets of dynamic or static coordinates, among other examples.
- the computing system may also detect one or more further inputs. For example, in addition to input 502 and additional input 504, the computing system may detect a further input, perhaps in an example where a computing system uses a three touch zoom process (e.g., a process where the user moves their fingers inwards and/or outwards to zoom out and/or in, respectively).
- a three touch zoom process e.g., a process where the user moves their fingers inwards and/or outwards to zoom out and/or in, respectively.
- Each of the inputs may be detected simultaneously, which may be defined as inputs within a threshold time period of each other.
- the computing system may detect input 502 and subsequently, after 1 ms, the computing system may detect input 504.
- the computing system may determine that input 502 and input 504 are simultaneous inputs and likely indicative of further input.
- the computing system may use historical data to determine whether the input is indicative of likely future input.
- the historical data may be stored on the computing system, and the computing system may retrieve the data and determine whether the detected areas of the screen are similar to areas associated with historical inputs that were followed by a zoom in and/or out process.
- the computing system may take an average of the coordinates indicating more areas of an image or of a screen that were frequently inputted before a zoom process, and the computing system may determine whether the indicated one or more areas (e g., input 502 and/or additional input 504) are within a threshold distance of the average historical areas.
- the historical data may also include one or more additional fields, including, for example, which camera was then activated and used.
- method 300 includes, based on the at least one input associated with the screen area, determining a secondary camera to preemptively activate in anticipation of zooming in or out.
- the computing system may be connected to a plurality of cameras, and the computing system may select a camera from the plurality of cameras to activate based on the at least one input associated with the screen area.
- the computing system may be connected to a wide angle camera, a default camera, and a zoom camera.
- the wide camera may have a larger field of view than the default camera and the zoom camera, while the default camera may have a narrower field of view than the wide angle camera but a larger field of view than the zoom camera.
- the zoom camera may have the narrowest field of view and may be able to capture zoomed-in scenes in more detail.
- the computing system may also include one or more other cameras with distinct characteristics as well.
- the computing system may select the camera with the most likelihood of subsequently being used based on the inputs. Selecting the camera with the most likelihood of being used may be based on the camera that is activated and being used at the time of receiving the input. For example, the computing system may receive an input indicative of zooming in and/or out while receiving and/or displaying data from a wide angle camera, and the wide angle camera has the largest field of view of any camera in the computing system. As such, the computing system may select to activate the camera with a narrower field of view than the wide angle camera. In the example of the three cameras above, the computing system may select to activate the default camera with a wider field of view than the zoom lens but a narrower field of view than the wide angle lens.
- the computing system may determine the distance between the plurality of inputs to select a camera.
- Figure 6 depicts graph 600 of distances, in accordance with example embodiments.
- the computing system may detect two inputs and the computing system may determine a distance between the two inputs.
- Distance 604 is plotted against time 602 in graph 600.
- the computing system may then compare the distances to a threshold distance, as indicated by threshold distance 606.
- the computing system may select a default camera with a wider field of view than the zoom camera to activate at point 622. If the computing system is displaying data from the default camera, the computing system may select a wide angle camera to activate at point 622. After activating the wide angle camera at point 622, the computing system may receive further input validating the input.
- inputs may be user inputs as well as other inputs, such as inputs from analyzing sensor data from the primary camera and other sensors.
- the computing system may perform one or more image analysis processes to determine if a subject in an image received from the camera takes up more than a threshold high amount of the area within the image (which may indicate an intention to zoom out) or if the image received from the camera takes up less than a threshold low amount of the area within the image (which may indicate an intention to zoom in).
- the computing system may use dimensions of the display area and/or physical dimensions of the screen in addition to the distance between the detected screen areas.
- the display area may be an area in which an image is displayed.
- the computing system may determine a ratio between the distance between the detected screen areas and the dimensions of the display to select a camera to activate, perhaps by comparing the determined ratio to one or more threshold ratios. For example, the computing system may detect a plurality of inputs and may determine that the distance between the inputs is 2 cm. The computing system may then compare that distance of 2 cm to dimensions of the screen which may be 15 cm by 20 cm to obtain a ratio of 2/20. Since the ratio of 2/20 is relatively small, the computing system may determine that it is likely for the user to zoom in, and as such, the computing system may determine to select a camera with a narrower field of view to preemptively activate.
- method 300 includes preemptively activating the secondary camera.
- preemptively activating the secondary camera may be advantageous in terms of battery consumption and user experience.
- the computing system may be able to activate the secondary camera in a targeted manner, where the computing system activates the secondary camera after detecting an input indicating further input to zoom in and/or out.
- preemptively activating the secondary camera may cause the camera to converge the settings of the selected secondary camera with those of the primary camera.
- Settings may include an exposure setting, a white balance setting, a focus setting, among other settings.
- the computing system may determine settings for the secondary camera that result in sensor data with similar white balance, focus, and exposure as to the primary camera.
- the converged settings of the secondary camera may be different than the settings of the primary camera, despite the sensor data of the secondary camera having similar properties when compared with the primary camera.
- converging the one or more settings of the secondary camera with one or more settings of the primary camera may involve adjusting the settings based on analyzing data from the secondary camera and/or one or more additional sensors.
- Figure 7 depicts graph 700, in accordance with example embodiments.
- Graph 700 depicts battery consumption as the input is detected and the secondary camera is activated.
- Graph 700 plots power 704 against time 702.
- the computing system may detect an input.
- the computing system may activate the secondary camera. After activating the secondary camera, and as discussed below, the computing system may then detect further inputs that may validate selecting the secondary camera and cause the computing system to switch to the secondary camera at point 714.
- the computing system may receive further user input that may be used to verify that the secondary camera was correctly activated For example, if the computing system detects two inputs that are above a threshold distance apart, the computing system may preemptively activate a wide angle camera. When the computing system detects further inputs that are decreasing distances apart, the computing system may verify that the further user input indicates that activating the wide angle camera was correct. If the computing system determines that selecting that particular secondary camera to activate was inaccurate, the computing system may deactivate that camera and select an appropriate secondary camera to activate instead.
- method 300 includes, based on further input associated with the screen area, zooming the displayed representation of the environment in or out based on secondary sensor data captured by the secondary camera.
- the further input may be a user input of dragging the fingers of the initial user input apart.
- the further input may indicate an amount to zoom in and/or out. For example, if the distance between the detected inputs are increasing (e.g., the fingers are being dragged apart), then the computing system may determine that the further input is indicating to zoom in. If the distance between the detected inputs are decreasing (e g., the fingers are being dragged together), then the computing system may determine that the further input is indicating to zoom out.
- the further input may also indicate an amount to zoom in and/or out. For example, if the computing system detects further user inputs with increasing distances between a plurality of inputs (e.g., that fingers are being dragged apart), the computing system may determine how much to zoom in based on how far apart the inputs are. And conversely, if the computing system detects further user inputs with decreasing distances between a plurality of inputs (e g., that fingers are being dragged together), the computing system may determine how much to zoom out based on how far apart the inputs are.
- the computing system may determine to gradually crop sensor data from the primary camera as if to zoom in while the distances are changing from 100 pixels apart to 500 pixels apart.
- the computing system may deactivate the primary camera. If the computing system receives input indicating to zoom in and/or out, the computing system may repeat the process described herein with the selected secondary camera as the primary camera. Based on the selected secondary camera as the primary camera, the computing system may then determine another secondary camera to preemptively activate.
- the computing system may detect further input associated with another area of the screen.
- the computing system may determine a third camera to preemptively activate based on the further user input and based on the disclosure herein.
- the computing system may include a wide angle camera, a default camera, and a zoom camera.
- the wide angle camera may be the primary camera, and the computing system may preemptively activate the default camera as the secondary camera based on input indicating to zoom in.
- the computing system may then switch to the default camera, and while the computing system is receiving and displaying sensor data from the default camera, the computing system may receive additional input indicating to zoom in further.
- the computing system may then preemptively activate the zoom camera as the tertiary camera.
- the primary camera and the tertiary camera may be the same.
- Process 900 may be triggered by detecting an input at block 902. Based on detecting an input at block 902, the computing system may then simultaneously zoom in and/or out at block 904 and estimate zoom at block 910.
- the computing system may estimate how much the user may want to zoom at block 910.
- the computing system may determine whether zooming in or zooming out is more likely, and based on this determination, the computing system may determine a secondary camera to activate.
- method 300 further includes detecting an additional input associated with an additional area of the screen, wherein determining the secondary camera to preemptively activate in anticipation of zooming in or out is based on the additional input associated with the additional area of the screen.
- the method further includes determining a distance between the area of the screen and the additional area of the screen, wherein determining the secondary camera to preemptively activate is based on the determined distance.
- determining the secondary camera to preemptively activate comprises comparing the distance between the area of the screen and the additional area of the screen to a threshold distance and, based on the distance being above the threshold distance, selecting a camera having a larger field of view than the primary camera, wherein the selected camera is the secondary camera.
- determining the secondary camera to preemptively activate comprises comparing the distance between the area of the screen and the additional area of the screen to a threshold distance and, based on the distance being below the threshold distance, selecting a camera having a smaller field of view than the primary camera, wherein the selected camera is the secondary camera.
- method 300 further comprises receiving historical input data, wherein determining the secondary camera is further based on the historical input data.
- determining the secondary camera comprises based on the historical input data, determining a location associated with past inputs, wherein each of the past inputs is associated with activating a camera. Determining the secondary camera may also comprise determining a distance between the area of the screen associated with the at least one input and the location associated with the past inputs and determining the secondary camera to activate based on the determined distance.
- the one or more settings includes an exposure setting, a white balance setting, and a focus setting, wherein converging the one or more settings of the secondary camera with the one or more settings of the primary camera results in the one or more settings of the secondary camera being different from the one or more settings of the primary camera.
- the screen is a touchscreen, wherein detecting the at least one input associated with the area of the screen comprises detecting a user touch input on the screen.
- method 300 further includes after zooming the displayed representation of the environment in or out based on the secondary sensor data captured by the secondary camera, deactivating the primary camera.
- a system may comprise a processor and a non-transitory computer- readable medium having stored thereon instructions that, when executed by the processor, cause the processor to perform operations of method 300.
- a non-transitory computer readable medium comprising program instructions executable by one or more processors to perform operations, the operations comprising the method of Figure 3.
- each step, block, and/or communication can represent a processing of information and/or a transmission of information in accordance with example embodiments.
- Alternative embodiments are included within the scope of these example embodiments.
- operations described as steps, blocks, transmissions, communications, requests, responses, and/or messages can be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved.
- blocks and/or operations can be used with any of the message flow diagrams, scenarios, and flow charts discussed herein, and these message flow diagrams, scenarios, and flow charts can be combined with one another, in part or in whole.
- a step or block that represents a processing of information may correspond to circuitry that can be configured to perform the specific logical functions of a herein-described method or technique.
- a block that represents a processing of information may correspond to a module, a segment, or a portion of program code (including related data).
- the program code may include one or more instructions executable by a processor for implementing specific logical operations or actions in the method or technique.
- the program code and/or related data may be stored on any type of computer readable medium such as a storage device including random access memory (RAM), a disk drive, a solid state drive, or another storage medium.
- the computer readable medium may also include non-transitory computer readable media such as computer readable media that store data for short periods of time like register memory, processor cache, and RAM.
- the computer readable media may also include non- transitory computer readable media that store program code and/or data for longer periods of time.
- the computer readable media may include secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, solid state drives, compactdisc read only memory (CD-ROM), for example.
- the computer readable media may also be any other volatile or non-volatile storage systems.
- a computer readable medium may be considered a computer readable storage medium, for example, or a tangible storage device.
- a step or block that represents one or more information transmissions may correspond to information transmissions between software and/or hardware modules in the same physical device. However, other information transmissions may be between software modules and/or hardware modules in different physical devices.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Studio Devices (AREA)
Abstract
Un procédé consiste à afficher sur un écran une représentation d'un environnement d'après des données primaires de capteur capturées par une caméra primaire. Le procédé comprend également la détection d'au moins une entrée associée à une zone de l'écran. Le procédé comprend en outre, sur la base de l'entrée ou des entrées associées à la zone d'écran, la détermination d'une caméra secondaire à activer de manière préemptive en prévision d'un zoom avant ou arrière. Le procédé consiste en outre à activer de manière préemptive la caméra secondaire. Le procédé comprend également, sur la base d'une autre entrée associée à la zone d'écran, un zoom avant ou arrière sur la représentation affichée de l'environnement d'après des données secondaires de capteur capturées par la caméra secondaire.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202463569336P | 2024-03-25 | 2024-03-25 | |
| US63/569,336 | 2024-03-25 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2025207596A1 true WO2025207596A1 (fr) | 2025-10-02 |
Family
ID=95446494
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/US2025/021290 Pending WO2025207596A1 (fr) | 2024-03-25 | 2025-03-25 | Activation préemptive de caméra |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2025207596A1 (fr) |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106303258A (zh) * | 2016-09-19 | 2017-01-04 | 深圳市金立通信设备有限公司 | 一种基于双摄像头的拍摄方法和终端 |
| US20190037189A1 (en) * | 2017-07-25 | 2019-01-31 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method and device for adjusting white balance and storage medium |
| US20210400206A1 (en) * | 2019-02-19 | 2021-12-23 | Samsung Electronics Co., Ltd. | Electronic device and method for changing magnification of image using multiple cameras |
| US20220053133A1 (en) * | 2020-07-29 | 2022-02-17 | Google Llc | Multi-Camera Video Stabilization |
| CN111614891B (zh) * | 2020-04-24 | 2022-09-16 | 深圳传音控股股份有限公司 | 一种相机应用功能切换方法、终端及计算机可读存储介质 |
-
2025
- 2025-03-25 WO PCT/US2025/021290 patent/WO2025207596A1/fr active Pending
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106303258A (zh) * | 2016-09-19 | 2017-01-04 | 深圳市金立通信设备有限公司 | 一种基于双摄像头的拍摄方法和终端 |
| US20190037189A1 (en) * | 2017-07-25 | 2019-01-31 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method and device for adjusting white balance and storage medium |
| US20210400206A1 (en) * | 2019-02-19 | 2021-12-23 | Samsung Electronics Co., Ltd. | Electronic device and method for changing magnification of image using multiple cameras |
| CN111614891B (zh) * | 2020-04-24 | 2022-09-16 | 深圳传音控股股份有限公司 | 一种相机应用功能切换方法、终端及计算机可读存储介质 |
| US20220053133A1 (en) * | 2020-07-29 | 2022-02-17 | Google Llc | Multi-Camera Video Stabilization |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9998651B2 (en) | Image processing apparatus and image processing method | |
| EP2975838B1 (fr) | Procédé et dispositif de réglage de paramètre de capture d'image | |
| US10230901B2 (en) | Realtime capture exposure adjust gestures | |
| US10594925B2 (en) | Camera zoom level and image frame capture control | |
| US20190174054A1 (en) | Camera zoom level and image frame capture control | |
| EP3697079B1 (fr) | Procédé et appareil de capture d'image et terminal | |
| US9535604B2 (en) | Display device, method for controlling display, and recording medium | |
| US11856293B2 (en) | Camera autofocus using time-of-flight assistance | |
| US20120306786A1 (en) | Display apparatus and method | |
| KR102121528B1 (ko) | 촬영 장치 및 이의 제어 방법 | |
| CN108830219A (zh) | 基于人机交互的目标跟踪方法、装置及存储介质 | |
| US20150029381A1 (en) | Electronic device and method of photographing image using the same | |
| WO2025207596A1 (fr) | Activation préemptive de caméra | |
| US20150363004A1 (en) | Improved tracking of an object for controlling a touchless user interface | |
| US20240430567A1 (en) | Depth-Histogram Based Autofocus | |
| WO2024232859A1 (fr) | Cadrage intelligent basé sur la saillance d'image avec prise en compte de la position de tapotement et d'un réglage continu | |
| WO2024263717A1 (fr) | Prévisualisation de l'autofocus en une seule fois pour les scènes en contre-jour | |
| EP4581559A2 (fr) | Cadrage intelligent basé sur le relief d'image | |
| WO2024129131A1 (fr) | Réduction de bruit d'image basée sur la perception de la vision humaine | |
| KR20250069960A (ko) | 카메라 자동 초점을 위한 관심 영역 결정 방법 | |
| WO2024263152A1 (fr) | Prédiction de mise au point automatique de caméra basée sur l'apprentissage pour capteurs d'imagerie à détection de phase clairsemée | |
| WO2025144842A1 (fr) | Commutation automatique de mode de caméra basée sur l'utilisation |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 25719511 Country of ref document: EP Kind code of ref document: A1 |