US20190124294A1 - Remote collaboration system with projector-camera based robot device and head mounted display, and remote interaction method using the same - Google Patents
Remote collaboration system with projector-camera based robot device and head mounted display, and remote interaction method using the same Download PDFInfo
- Publication number
- US20190124294A1 US20190124294A1 US16/039,436 US201816039436A US2019124294A1 US 20190124294 A1 US20190124294 A1 US 20190124294A1 US 201816039436 A US201816039436 A US 201816039436A US 2019124294 A1 US2019124294 A1 US 2019124294A1
- Authority
- US
- United States
- Prior art keywords
- mounted display
- head mounted
- image
- remote
- space
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
- H04N7/141—Systems for two-way working between two video terminals, e.g. videophone
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B21/00—Projectors or projection-type viewers; Accessories therefor
- G03B21/12—Projectors or projection-type viewers; Accessories therefor adapted for projection of either still pictures or motion pictures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/45—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/66—Remote control of cameras or camera parts, e.g. by remote control devices
- H04N23/661—Transmitting camera control signals through networks, e.g. control via the Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/695—Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
-
- H04N5/23238—
-
- H04N5/247—
Definitions
- Example embodiments relate generally to remote collaboration technique, and more particularly to remote collaboration systems with projector-camera based robot devices and head mounted displays, and remote interaction methods performed by the remote collaboration systems.
- a remote collaboration system is a system in which a remote user and a local guide (or a local helper) communicate with each other to perform necessary tasks or educations.
- the remote user is a person who is located in a remote space and needs assistance.
- the local guide is a person who is located in a local space, provides assistance to the remote user and cooperates with the remote user.
- Various remote collaboration systems have been implemented based on various devices.
- the remote collaboration system When a remote collaboration system is implemented based on a mobile device, the remote collaboration system has high accessibility because the mobile device is friendly and convenient for the remote user, however, the remote user has a problem of low task efficiency because the remote user performs a task with holding the mobile device and watching a display on the mobile device to obtain explanations from the local guide and the local guide also has a problem of difficulty in observing the remote space.
- the remote collaboration system is implemented based on a wearable device, the remote user does not need to hold the wearable device, however, the remote user may feel uncomfortable with the wearable device and the local guide still has a problem of difficulty in observing the remote space.
- the local guide may easily observe the remote space by controlling the robot device as desired and the remote user may feel more comfortable in comparison with the wearable device and the mobile device, however, the remote user has a problem of low task efficiency because the remote user performs a task with watching a monitor on the robot device to obtain explanations from the local guide and the local guide still has a problem of difficulty in observing the remote space if a camera on the robot device has narrow angle of view.
- the present disclosure is provided to substantially obviate one or more problems due to limitations and disadvantages of the related art.
- At least one example embodiment of the present disclosure provides a remote interaction method that is performed by a remote collaboration system with a head mounted display and a robot device including a projector and at least one camera capable of efficiently cooperating a remote user and a local guide with each other.
- At least one example embodiment of the present disclosure provides a remote collaboration system with a head mounted display and a robot device including a projector and at least one camera capable of efficiently cooperating a remote user and a local guide with each other.
- the robot device in a remote interaction method performed by a remote collaboration system comprising a robot device and a head mounted display, the robot device is located in a remote space and comprising a projector, a panoramic camera and a high resolution camera.
- the head mounted display is located in a local space apart from the remote space.
- a communication between the robot device and the head mounted display is established based on a communication request.
- the remote space is observed by the head mounted display based on first image information and second image information.
- the first image information is collected by the panoramic camera and received from the robot device.
- the second image information is collected by the high resolution camera and received from the robot device.
- Necessary information to be provided to the remote space is searched by the head mounted display based on a result of observing the remote space.
- the necessary information received from the head mounted display is provided by the robot device to the remote space using the projector.
- a panoramic image of the remote space may be displayed by the head mounted display based on the first image information.
- the panoramic image and a magnified image of a first region in the remote space may be displayed by the head mounted display together based on the first image information and the second image information when a screen magnifying interaction is performed while the head mounted display displays the panoramic image.
- the magnified image may be displayed by the head mounted display based on the second image information when the screen magnifying interaction is performed while the head mounted display displays the panoramic image and the magnified image together.
- a current screen image displayed on the head mounted display is stored when a screen saving interaction is performed.
- the panoramic image in observing the remote space, may be displayed by the head mounted display based on the first image information when a screen demagnifying interaction is performed while the head mounted display displays the panoramic image and the magnified image together.
- the panoramic image and the magnified image may be displayed by the head mounted display together based on the first image information and the second image information when the screen demagnifying interaction is performed while the head mounted display displays the magnified image.
- the magnified image displayed on the head mounted display may be removed when a removing interaction is performed while the head mounted display displays the panoramic image and the magnified image together, or a user interface object displayed on the head mounted display may be removed when the removing interaction is performed while the head mounted display displays the panoramic image or the magnified image with the user interface object.
- the user interface object may be used for controlling an operation of the head mounted display.
- the local space in searching the necessary information, may be observed by the head mounted display based on third image information collected by a first camera included in the head mounted display when a screen demagnifying interaction is performed while the head mounted display displays the panoramic image.
- a first searching operation may be performed on a plurality of first data stored in the head mounted display.
- the necessary information may be obtained based on at least one of the result of observing the remote space, a result of observing the local space and a result of the first searching operation.
- a second searching operation may be performed on a plurality of second data stored in the robot device.
- a third searching operation may be performed on a plurality of third data associated with the necessary information by connecting an Internet network via the head mounted display.
- the necessary information may be obtained further based on at least one of a result of the second searching operation and a result of the third searching operation.
- the necessary information in providing the necessary information to the remote space, may be displayed by the projector on the remote space when the necessary information is provided as an image.
- a pointer may be displayed by the projector on the first region in the remote space when the necessary information is associated with the first region in the remote space.
- both the high resolution camera and the projector may be controlled toward the first region in the remote space when the head mounted display displays the panoramic image and the magnified image together or only the magnified image and when the necessary information is associated with the first region in the remote space and the projector displays the pointer on the first region in the remote space.
- the remote collaboration system may further comprise a controller that is located in the local space and controls an operation of the head mounted display.
- Direction controls of the high resolution camera and the projector may be performed based on head tracking information obtained by the head mounted display.
- a movement of the pointer displayed on the first region in the remote space may be controlled by the controller.
- a remote collaboration system comprises a robot device and a head mounted display.
- the robot device is located in a remote space and comprises a projector, a panoramic camera and a high resolution camera.
- the projector provides necessary information to the remote space.
- the panoramic camera collects first image information of the remote space.
- the high resolution camera collects second image information of the remote space.
- the head mounted display is located in a local space apart from the remote space.
- a communication between the robot device and the head mounted display is established based on a communication request.
- the head mounted display observes the remote space based on the first image information and the second image information.
- the head mounted display searches the necessary information based on a result of observing the remote space.
- the robot device provides the necessary information received from the head mounted display to the remote space using the projector.
- the robot device including the projector and the at least one camera may be used and may be controlled by the head mounted display.
- the collaboration between the remote user and the local guide may be efficiently performed.
- the remote user may feel more comfortable without holding or wearing the device, may obtain the explanations from the local guide immediately displayed on the region having problems via the projector, and thus may have improved task efficiency.
- the local guide may control the robot device to watch the remote space and provide the explanations as desired, and may wear the head mounted display to observe the remote space with heightened immersion.
- the remote collaboration may be performed with high task efficiency and without inconvenience based on the screen change, the control of the robot device, etc. according to example embodiments even if the local guide wears the head mounted display.
- FIG. 1 is a block diagram illustrating a remote collaboration system according to example embodiments.
- FIG. 2 is a diagram illustrating a prototype of a robot device included in a remote collaboration system according to example embodiments.
- FIG. 3 is a flow chart illustrating a remote interaction method performed by a remote collaboration system according to example embodiments.
- FIGS. 4A and 4B are diagrams for describing an operation associated with a communication request by a head mounted display included in a remote collaboration system according to example embodiments.
- FIGS. 5, 6A, 6B and 6C are diagrams for describing an operation of performing an event by a head mounted display included in a remote collaboration system according to example embodiments.
- FIGS. 7A, 7B, 7C and 8 are diagrams for describing an operation of observing a remote space by a head mounted display included in a remote collaboration system according to example embodiments.
- FIGS. 9, 10 and 11 are flow charts for describing an operation of changing a screen on a head mounted display included in a remote collaboration system according to example embodiments.
- FIG. 12 is a flow chart for describing an operation of controlling a robot device performed by a head mounted display included in a remote collaboration system according to example embodiments.
- FIG. 13 is a flow chart for describing an operation of searching information performed by a head mounted display included in a remote collaboration system according to example embodiments.
- FIGS. 14A, 14B and 14C are diagrams for describing an operation of displaying necessary information performed by a robot device included in a remote collaboration system according to example embodiments.
- FIG. 1 is a block diagram illustrating a remote collaboration system according to example embodiments.
- a remote collaboration system 10 includes a robot device 100 and a head mounted display (HMD) 200 .
- the remote collaboration system 10 may further include a controller 250 .
- the robot device 100 is located or disposed in a remote space and is a device which provides help, assistance, support or guidance to a remote user in the remote space.
- the head mounted display or head mounted display device 200 is located or disposed in a local space that is apart from the remote space and freely controls the robot device 100 regardless of any motions or movements of the remote user.
- a local guide or a local helper in the local space may wear the head mounted display 200 .
- the head mounted display 200 may be controlled by the local guide.
- the robot device 100 and the head mounted display 200 may communicate with each other via a network 300 .
- the robot device 100 and the head mounted display 200 may exchange image data IMG and/or voice data VOC with each other via the network 300 , and the head mounted display 200 may transmit a command signal CS, an annotation signal AS, necessary information RI, etc. to the robot device 100 via the network 300 .
- the network 300 may include at least one of various wire/wireless communication networks.
- the network 300 may include a WiFi communication network and/or other wireless communication networks, such as a third generation (3G) communication network, a fourth generation (4G) communication network, a long term evolution (LTE) communication network, etc.
- 3G third generation
- 4G fourth generation
- LTE long term evolution
- the robot device 100 includes a projector 110 , a panoramic camera 120 and a high resolution camera 130 .
- the robot device 100 may further include a computer 140 , motors 150 and 160 , and a transportation unit 170 .
- the robot device 100 may be referred to as a remote collaboration device.
- the projector 110 provides the necessary information RI to the remote space (e.g., to the remote user).
- the necessary information RI may be obtained by the head mounted display 200 and may be transmitted from the head mounted display 200 to the projector 110 via the network 300 and the computer 140 .
- the panoramic camera 120 collects first image information IMG 1 of the remote space.
- the first image information IMG 1 may represent or correspond to a panoramic image or a 360 degree image of the remote space.
- the high resolution camera 130 collects second image information IMG 2 of the remote space.
- the second image information IMG 2 may represent or correspond to a high resolution image or a magnified image of a specific region in the remote space.
- an angle of view of the high resolution camera 130 may be narrower than an angle of view of the panoramic camera 120
- a resolution of the high resolution camera 130 may be higher than a resolution of the panoramic camera 120 .
- the high resolution camera 130 may be omitted.
- the computer 140 may control overall operations of the robot device 100 .
- the computer 140 may communicate with the head mounted display 200 via the network 300 , may transmit the first image information IMG 1 collected by the panoramic camera 120 and the second image information IMG 2 collected by the high resolution camera 130 to the head mounted display 200 , may receive the necessary information RI from the head mounted display 200 to transmit the necessary information RI to the projector 110 , may control the projector 110 , the panoramic camera 120 and the high resolution camera 130 based on the command signal CS and/or the annotation signal AS received from the head mounted display 200 , and may generate motor control signals MS 1 and MS 2 for controlling the motors 150 and 160 .
- the motor 150 may control movements of at least one of the projector 110 , the panoramic camera 120 and the high resolution camera 130 based on the motor control signal MS 1 .
- the motor 150 may control a projecting direction of the projector 110 or a lens direction of the high resolution camera 130 .
- FIG. 1 illustrates a single motor 150 for controlling all of the projector 110 , the panoramic camera 120 and the high resolution camera 130
- the remote collaboration system 10 may include a plurality of motors each of which controls a respective one of the projector 110 , the panoramic camera 120 and the high resolution camera 130 .
- the motor 160 may control the transportation unit 170 based on the motor control signal MS 2 .
- a location of the robot device 100 in the remote space may be controlled based on an operation of the transportation unit 170 .
- the head mounted display 200 displays images such that the local guide observes the remote space based on the images displayed on the head mounted display 200 .
- the head mounted display 200 may display the panoramic image of the remote space based on the first image information IMG 1 , may display the high resolution image of the specific region in the remote space based on the second image information IMG 2 , or may display the panoramic image and the high resolution image together.
- the head mounted display 200 may include a camera 210 .
- the head mounted display 200 may display an image of the local space based on third image information collected by the camera 210 .
- the head mounted display 200 searches the necessary information RI that is to be provided to the remote space.
- the head mounted display 200 may further include a processing device (e.g., a central processing unit (CPU)), a storage device (e.g., a memory), etc. for performing the searching operation.
- a processing device e.g., a central processing unit (CPU)
- a storage device e.g., a memory
- the controller 250 may be located or disposed in the local space and may control an operation of the head mounted display 200 .
- the annotation signal AS may be transmitted to the head mounted display 200 based on a control of the controller 250 .
- the controller 250 may include a touch pad, a touch screen or a dedicated control unit for the head mounted display 200 .
- the command signal CS may be transmitted to the head mounted display 200 based on a control of a user interface (UI) 220 .
- UI user interface
- the user interface 220 may include a graphic user interface (GUI) displayed on the head mounted display 200 .
- GUI graphic user interface
- FIG. 1 illustrates an example where the command signal CS is generated by the user interface 220 and the annotation signal AS is generated by the controller 250
- inventive concepts are not limited thereto.
- the user interface 220 may generate the annotation signal AS
- the controller 250 may generate the command signal CS
- the command signal CS and/or the annotation signal AS may be generated based on head tracking information obtained by motions or movements of the local guide, or the command signal CS and/or the annotation signal AS may be generated while the searching operation for the necessary information RI is performed.
- the local guide may observe the remote space (e.g., a remote environment) using the robot device 100 (e.g., the remote collaboration device) located at a remote user side.
- the panoramic image or the 360 degree image of the remote space may be obtained by the panoramic camera 120 to observe the whole remote space
- the high resolution image or the magnified image of the specific region in the remote space may be obtained by the high resolution camera 130 to observe the specific region in the remote space, if necessary.
- Explanations of the local guide may be provided from the projector 110 to the remote space.
- the movements of the projector 110 , the panoramic camera 120 , the high resolution camera 130 and the whole robot device 100 may be controlled by the motors 150 and 160 .
- the local guide may wear the head mounted display 200 to feel or realize the remote space vividly, finely or with details.
- the local guide may observe the local space (e.g., a local environment) using the camera 210 included in the head mounted display 200 .
- a screen and/or the graphic user interface may be implemented in the head mounted display 200 , and the head tracking information and/or the controller 250 (e.g., the touch pad or the dedicated control unit) may be used for controlling the screen and/or the graphic user interface implemented in the head mounted display 200 .
- a remote collaboration interaction may be performed based on the head mounted display 200 .
- the remote collaboration interaction may include a control interaction for the head mounted display 200 (e.g., the screen and/or the graphic user interface implemented in the head mounted display 200 ) and a control interaction for the robot device 100 .
- FIG. 2 is a diagram illustrating a prototype of a robot device included in a remote collaboration system according to example embodiments.
- a body including a plurality of receiving spaces is prepared, a projector, a panoramic camera, a high resolution camera, a mini PC (personal computer) and a motor control board are mounted on the plurality of receiving spaces, the body is equipped with motors, wheels and a battery, and thus a prototype of a robot device is manufactured.
- the mini PC and the wheels in FIG. 2 may correspond to the computer 140 and the transportation unit 170 in FIG. 1 , respectively, and the motor control board and the battery are added for controlling the motors and supplying power to the motors, respectively.
- FIG. 3 is a flow chart illustrating a remote interaction method performed by a remote collaboration system according to example embodiments.
- a communication between the robot device 100 located in the remote space and the head mounted display 200 located in the local space is established based on a communication request (step S 100 ).
- the communication request may be generated by one of the robot device 100 and the head mounted display 200 .
- the head mounted display 200 may generate the communication request based on a control of the local guide.
- the head mounted display 200 receives the first image information IMG 1 collected by the panoramic camera 120 and the second image information IMG 2 collected by the high resolution camera 130 , and the remote space is observed by the head mounted display 200 based on the first image information IMG 1 and the second image information IMG 2 (step S 200 ).
- the head mounted display 200 may display the panoramic image of the remote space based on the first image information IMG 1 , may display the high resolution image of the specific region in the remote space based on the second image information IMG 2 , or may display the panoramic image and the high resolution image together.
- the necessary information RI that is to be provided to the remote space is searched by the head mounted display 200 based on a result of observing the remote space (step S 300 ).
- the remote space may be observed to check, realize or detect problems in the remote space, and the necessary information RI may be searched to resolve the problems in the remote space.
- the head mounted display 200 may observe the remote space based on a control of the local guide or may search data stored in the head mounted display 200 , data stored in the robot device 100 or data in an Internet network to obtain the necessary information RI.
- the robot device 100 receives the necessary information RI from the head mounted display 200 , and the necessary information RI is provided by the robot device 100 to the remote space using the projector 110 (step S 400 ).
- the necessary information RI may be provided as an image such as a picture, a sketch, etc., or may be provided as a pointer that is indicated on the specific region in the remote space.
- FIGS. 4A and 4B are diagrams for describing an operation associated with a communication request by a head mounted display included in a remote collaboration system according to example embodiments.
- a configuration of screen GUIs of the head mounted display 200 may include a first screen GUI for the communication request and a second screen GUI after the communication between the robot device 100 and the head mounted display 200 is established.
- FIGS. 4A and 4B illustrate the first screen GUI for the communication request.
- an initial screen of the head mounted display 200 may include the image of the local space displayed as a background.
- the image of the local space may be displayed based on third image information collected by the camera 210 .
- the initial screen of the head mounted display 200 may further include a phone number input menu, a contact menu, a recent call menu and a bookmark list menu displayed as GUI forms (e.g., buttons, etc.).
- each button may be clicked using a HMD pointer to perform a corresponding function.
- a recent call button is clicked, a recent call list may be displayed on a screen of the head mounted display 200 .
- a contact button is clicked, a contact list may be displayed on a screen of the head mounted display 200 .
- the contact list may include a plurality of contacts each of which includes a photo, a name, a phone number, an address, etc.
- the contact list may be displayed.
- a specific contact e.g., a contact of the remote user having the robot device 100 in the remote space
- a communication request screen with a message e.g., “call?”
- the communication request may be transmitted from the head mounted display 200 to the robot device 100 .
- the robot device 100 accepts the communication request the communication between the robot device 100 and the head mounted display 200 may be established. In other words, step S 100 in FIG. 3 may be performed by the communication request from the head mounted display 200 .
- a screen for adding a new contact may be displayed.
- an image list for selecting an image to be added to the new contact may be displayed on a screen of the head mounted display 200 .
- Bidirectional arrows illustrated in FIGS. 4A and 4B may indicate a scrolling function.
- the bookmark list and the recent call list in FIG. 4A and the contact list and the image list in FIG. 4B may be scrolled.
- FIGS. 5, 6A, 6B and 6C are diagrams for describing an operation of performing an event by a head mounted display included in a remote collaboration system according to example embodiments.
- An operation of performing an event by the head mounted display 200 may indicate an operation of clicking or selecting a button or a menu in both the first screen GUI for the communication request and the second screen GUI after the communication between the robot device 100 and the head mounted display 200 is established.
- an operation of FIG. 5 may be performed in step S 100 and may also be performed in steps S 200 , S 300 and S 400 .
- a HMD pointer may be controlled to perform a specific event such that the HMD pointer is located or disposed on a specific UI object on a screen of the head mounted display 200 (step S 110 ).
- the controller 250 may move the HMD pointer on the screen of the head mounted display 200 until the HMD pointer is located or disposed on the contact button.
- the controller 250 may be a touch pad or a touch screen as illustrated in FIG. 6B , or a dedicated control unit for the head mounted display 200 as illustrated in FIG. 6C .
- a click operation may be performed, and a virtual beam may be emitted from a current location of the HMD pointer toward the specific UI object (step S 120 ).
- the click operation may be performed by the controller 250 .
- step S 130 NO
- the location of the HMD pointer is not appropriate (e.g., the location of the HMD pointer does not correspond to the specific UI object), and thus steps S 110 and S 120 may be repeated.
- an object event corresponding to the specific UI object may be performed (step S 140 ). For example, when the contact button in FIG. 4A is clicked, a screen of the head mounted display 200 may be changed from the screen of FIG. 4A to the screen of FIG. 4B .
- a keyboard input may be performed on a screen of the head mounted display 200 .
- a virtual keyboard may be displayed on a screen of the head mounted display 200 , any button on the virtual keyboard may be clicked using the HMD pointer, and thus the keyboard input may be performed.
- the controller 250 may be a touch pad or a touch screen
- a keyboard may be displayed on the touch screen
- the local guide may touch any button on the keyboard, and thus the keyboard input may be performed.
- the image of the local space may be displayed on a screen of the head mounted display 200 to watch the keyboard on the touch screen via the head mounted display 200 .
- FIGS. 7A, 7B, 7C and 8 are diagrams for describing an operation of observing a remote space by a head mounted display included in a remote collaboration system according to example embodiments.
- the remote space may be observed via a screen of the head mounted display 200 .
- the head mounted display 200 may display the panoramic image of the remote space based on the first image information IMG 1 collected by the panoramic camera 120 .
- the head mounted display 200 may display the magnified image of the specific region in the remote space based on the second image information IMG 2 collected by the high resolution camera 130 .
- the second image information IMG 2 collected by the high resolution camera 130 .
- the head mounted display 200 may display the panoramic image and the magnified image together based on the first image information IMG 1 and the second image information IMG 2 .
- a screen of the head mounted display 200 may be changed in response to one of various interactions.
- FIG. 8 illustrates an example where an image obtained from the prototype of the robot device of FIG. 2 is displayed on the head mounted display 200 .
- the panoramic image and the magnified image e.g., the high resolution image
- a menu in FIG. 8 may correspond to UI objects.
- FIGS. 9, 10 and 11 are flow charts for describing an operation of changing a screen on a head mounted display included in a remote collaboration system according to example embodiments.
- the head mounted display 200 may display the panoramic image as a default image or a default screen based on the first image information IMG 1 , as illustrated in FIG. 7A .
- a screen magnifying interaction may be performed (step S 210 ), and a screen of the head mounted display 200 may be changed in response to the screen magnifying interaction (step S 220 ).
- the head mounted display 200 may display the panoramic image and the magnified image together based on the first and second image information IMG 1 and IMG 2 , as illustrated in FIG. 7B .
- the head mounted display 200 may display the magnified image based on the second image information IMG 2 , as illustrated in FIG. 7C .
- the screen magnifying interaction may be performed by the head mounted display 200 .
- head tracking information corresponding to the motion of sticking the head out may be recognized as the screen magnifying interaction.
- the screen magnifying interaction may be performed by the controller 250 .
- a pinch-to-zoom motion on the touch screen or a motion of sweeping on the dedicated controller may be recognized as the screen magnifying interaction.
- a screen demagnifying interaction may be performed (step S 230 ), and the screen of the head mounted display 200 may be changed in response to the screen demagnifying interaction (step S 240 ).
- the head mounted display 200 may display the panoramic image based on the first image information IMG 1 .
- the head mounted display 200 may display the panoramic image and the magnified image together based on the first and second image information IMG 1 and IMG 2 .
- the head mounted display 200 may display the image of the local space based on the third image information collected by the camera 210 .
- the screen of the head mounted display 200 may be changed in response to the screen magnifying interaction in the order of the image of the local space, the panoramic image of the remote space, the mixed image (e.g., the panoramic image and the magnified image together) of the remote space and the magnified image of the remote space.
- the screen of the head mounted display 200 may be changed in response to the screen demagnifying interaction in the order of the magnified image of the remote space, the mixed image of the remote space, the panoramic image of the remote space and the image of the local space.
- a removing interaction may be performed (step S 250 ), and a screen of the head mounted display 200 may be changed in response to the removing interaction (step S 260 ).
- the head mounted display 200 may only display the panoramic image based on the first image information IMG 1 .
- the magnified image may be instantaneously removed in response to the removing interaction, and the screen of the head mounted display 200 may be changed from a screen including the panoramic image and the magnified image together to a screen including the panoramic image only.
- the removing interaction may be performed by the head mounted display 200 .
- the head mounted display 200 e.g., performed by the local guide
- head tracking information corresponding to the motion of tilting the head to the left or right may be recognized as the removing interaction.
- the removing interaction may be performed by the controller 250 .
- the removing interaction may be used for removing a user interface object displayed on the head mounted display 200 .
- the user interface object for controlling an operation of the head mounted display 200 may be displayed on the screen of the head mounted display 200 .
- the menu may be displayed with the panoramic image and the magnified image on the screen of the head mounted display 200 , as illustrated in FIG. 8 .
- the removing interaction may be performed (step S 250 ), and the user interface object may be removed from the screen of the head mounted display 200 in response to the removing interaction (step S 260 ).
- a restoring interaction may be performed, and the removed magnified image and/or the removed user interface object may be restored on the screen of the head mounted display 200 .
- a screen saving interaction may be performed to store information associated with the remote space, and a current screen image displayed on the head mounted display 200 may be stored in response to the screen saving interaction.
- the current screen image may be stored in the head mounted display 200 as a static image (e.g., a still image, a stopped image, etc.) or a dynamic image (e.g., a video, a moving image, etc.).
- FIG. 12 is a flow chart for describing an operation of controlling a robot device performed by a head mounted display included in a remote collaboration system according to example embodiments.
- the head mounted display 200 may check the specific region to be magnified in the panoramic image (step S 270 ). To obtain the magnified image of the specific region, head tracking information collected by the head mounted display 200 may be transmitted to the robot device 100 (step S 275 ).
- step S 280 a direction of the high resolution camera 130 may be controlled toward the specific region in the remote space based on the head tracking information and the screen magnifying interaction (step S 285 ).
- a direction of the projector 110 may be further controlled toward the specific region in the remote space.
- the high resolution camera 130 and the projector 110 may be controlled or driven independently (e.g., separately).
- FIG. 13 is a flow chart for describing an operation of searching information performed by a head mounted display included in a remote collaboration system according to example embodiments.
- the screen demagnifying interaction described with reference to FIG. 10 may be performed (step S 310 ) while the head mounted display 200 displays the panoramic image.
- the head mounted display 200 may display the image of the local space based on the third image information collected by the camera 210 in response to the screen demagnifying interaction to observe the local space (step S 320 ).
- At least one of a first searching operation performed on a plurality of first data stored in the head mounted display 200 , a second searching operation performed on a plurality of second data stored in the robot device 100 , and a third searching operation performed on a plurality of third data associated with the necessary information RI by connecting an Internet network via the head mounted display 200 may be additionally performed (step S 330 ).
- the necessary information RI may be obtained based on at least one of the result of observing the remote space, a result of observing the local space, a result of the first searching operation, a result of the second searching operation and a result of the third searching operation (step S 340 ).
- At least one of observing the local space, the first searching operation, the second searching operation and the third searching operation may be omitted.
- FIGS. 14A, 14B and 14C are diagrams for describing an operation of displaying necessary information performed by a robot device included in a remote collaboration system according to example embodiments.
- the projector 110 may display the necessary information RI on the remote space.
- the necessary information RI corresponds to a static image or a dynamic image as illustrated in FIG. 14A
- the necessary information RI corresponds to a sketch image (e.g., a drawing image) as illustrated in FIG. 14B
- the projector 110 may display the image on the remote space.
- the image may be selected and transmitted by pushing a send button or based on a drag-and-drop scheme. A position of the image may be controlled according to eyes of the remote user.
- the projector 110 may display a projector pointer on the specific region in the remote space.
- a movement of the projector pointer on the specific region in the remote space may be performed by moving the HMD pointer by the controller 250 .
- the sketch image in FIG. 14B may be drawn by moving the HMD pointer.
- the remote interaction method (e.g., the remote collaboration) performed by the remote collaboration system 10 according to example embodiments may include four steps of establishing the communication between the robot device 100 and the head mounted display 200 , observing the problems in the remote space, searching the information for resolving the problems, and explaining solution of the problems. Interactions in each step will be synthetically described with detailed examples.
- the local guide in the local space wears the head mounted display 200 and controls the robot device 100 disposed in the remote space.
- the interactions in each step are performed based on the controls of the HMD and the remote collaboration device as described above.
- the local guide clicks the contact list on the screen of the head mounted display 200 using the controller 250 to call the remote user.
- the local guide clicks the phone number input menu and inputs a phone number using a keyboard (e.g., a virtual keyboard) to call the remote user.
- a keyboard e.g., a virtual keyboard
- An alarm indicating a call from the local guide rings or sounds in the robot device 100 , and the remote user clicks a call button on the robot device 100 to answer the call.
- the local guide watches or observes the panoramic image of the remote space transmitted from the robot device 100 via the head mounted display 200 .
- the local guide performs the screen magnifying interaction to watch or observe the magnified image of the specific region in the remote space.
- the local guide captures the magnified image or records the image and/or voice to store information associated with the remote space.
- the remote space (or the remote environment) is observed by the local guide using the head mounted display 200 , and the screen magnifying interaction is performed to display the magnified image (or the high resolution image) and observe components of the home appliances in detail.
- the screen capture is performed to save the current screen image for reference before searching the necessary information RI.
- the information may exist in the remote space, the remote collaboration device (e.g., the robot device 100 ), the head mounted display 200 , the local space, the Internet network, etc.
- Types of the information may include images, videos, texts, documents, etc.
- Information stored in the device is displayed, selected and transmitted based on HMD interactions and GUIs associated with images and/or videos.
- Information in the remote space is searched and captured based on the panoramic image and the magnified image.
- the screen demagnifying interaction is performed to display the image of the local space, and information in the local space is searched and captured based on the image of the local space.
- Information in the Internet network is displayed, searched and transmitted based on the Internet browser, the web surfing, the website capture, etc.
- the manuals or the specifications in the local space are observed by the local guide using the head mounted display 200 and the screen demagnifying interaction.
- the screen saving interaction is performed to store or capture the current screen image.
- the captured image is transmitted to the robot device 100 .
- Types of explaining the solution of the problems may include pointing, sketching, transmitting of the images and/or the videos, etc.
- the pointing and the sketching are performed to display the projector pointer and the sketch image on the remote space based on an accurate portion seen by the local guide.
- the transmitting of the images and/or the videos is performed to display the images and/or the videos on a desired portion in the remote space based on the drag-and-drop scheme.
- the displayed solution is stored, the displayed solution disappears when the direction of the projector 110 is changed, and the disappeared solution appears again when the direction of the projector 110 is returned to the original direction.
- the captured image of the manuals or the specifications is displayed on the screen (e.g., the UI screen) of the head mounted display 200 as a thumbnail.
- the captured image is selected, dragged and dropped on the desired portion in the remote space by the local guide using the HMD pointer, and the captured image is displayed on the remote space using the projector 110 .
- the remote collaboration system 10 may include the robot device 100 with the projector and the at least one camera and the head mounted display 200 (e.g., HMD interface) controlling the robot device 100 .
- the camera functions may be performed to capture the 360 degree image (or the panoramic image) of the remote space and the high resolution image (or the magnified image) of the specific region in the remote space.
- the robot device 100 may include a single camera for capturing both the 360 degree image and the high resolution image or two cameras (e.g., the panoramic camera 120 and the high resolution camera 130 ) for capturing the 360 degree image and the high resolution image, respectively.
- the robot device 100 may further include the motor for controlling the high resolution camera, and the direction of the high resolution camera may be controlled when the screen is changed from the 360 degree image to the high resolution image.
- the robot device 100 may further include the motor for controlling the projector.
- the robot device 100 may further include the computer for communications and computations, and the wheels and the battery for a moveable robot device.
- the HMD interface may be implemented with a video see-through, an optical see-through, etc.
- the interaction interface may be implemented with a touch pad, a dedicated control unit, etc.
- the touch pad may be familiar with many users, may be used with various gestures, and may be utilized such as a keyboard.
- the dedicated control unit may include a sensor for tracking a movement of the dedicated control unit, buttons, a touch pad, etc., may provide an intuitive user experience, and may be useful for a 3D input.
- a keyboard input may be further implemented.
- the local guide may wear the head mounted display 200 and may communicate with the remote user using the screen GUIs of the head mounted display 200 and the interaction interface.
- the head mounted display 200 may exchange image, video and/or voice data with the robot device 100 , and may transmit information and data for controlling the motor to the robot device 100 .
- the robot device 100 may move by the control of the local guide and the head mounted display 200 , and may display the information using the projector 110 . All data may be transmitted via the computer.
- the inventive concept may be applied to various remote coaching services where the remote user in the remote space do not know how to handle some devices (e.g., electronic devices) or has some problems and the local guide in the local space provides assistance to the remote user to resolve the problems.
- the inventive concept may be applied to various examples such as teaching a computer in which a computer screen is captured by the robot device and explanations are provided via the projector, playing a Korean chess with the local guide in which the projector pointer is displayed on a Korean chessboard, and showing a way to destination in which the destination and a current location are displayed on a map.
- the inventive concept may be applied to various examples such as cooking, educating, gardening, repairing, styling, interior designing, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- User Interface Of Digital Computer (AREA)
- Controls And Circuits For Display Device (AREA)
- Manipulator (AREA)
- Studio Devices (AREA)
Abstract
Description
- This application claims priority under 35 USC § 119 to Korean Patent Application No. 10-2017-0135689, filed on Oct. 19, 2017 in the Korean Intellectual Property Office (KIPO), the contents of which are herein incorporated by reference in their entirety.
- Example embodiments relate generally to remote collaboration technique, and more particularly to remote collaboration systems with projector-camera based robot devices and head mounted displays, and remote interaction methods performed by the remote collaboration systems.
- A remote collaboration system is a system in which a remote user and a local guide (or a local helper) communicate with each other to perform necessary tasks or educations. The remote user is a person who is located in a remote space and needs assistance. The local guide is a person who is located in a local space, provides assistance to the remote user and cooperates with the remote user.
- Various remote collaboration systems have been implemented based on various devices. When a remote collaboration system is implemented based on a mobile device, the remote collaboration system has high accessibility because the mobile device is friendly and convenient for the remote user, however, the remote user has a problem of low task efficiency because the remote user performs a task with holding the mobile device and watching a display on the mobile device to obtain explanations from the local guide and the local guide also has a problem of difficulty in observing the remote space. When a remote collaboration system is implemented based on a wearable device, the remote user does not need to hold the wearable device, however, the remote user may feel uncomfortable with the wearable device and the local guide still has a problem of difficulty in observing the remote space. When a remote collaboration system is implemented based on a robot device, the local guide may easily observe the remote space by controlling the robot device as desired and the remote user may feel more comfortable in comparison with the wearable device and the mobile device, however, the remote user has a problem of low task efficiency because the remote user performs a task with watching a monitor on the robot device to obtain explanations from the local guide and the local guide still has a problem of difficulty in observing the remote space if a camera on the robot device has narrow angle of view.
- Accordingly, the present disclosure is provided to substantially obviate one or more problems due to limitations and disadvantages of the related art.
- At least one example embodiment of the present disclosure provides a remote interaction method that is performed by a remote collaboration system with a head mounted display and a robot device including a projector and at least one camera capable of efficiently cooperating a remote user and a local guide with each other.
- At least one example embodiment of the present disclosure provides a remote collaboration system with a head mounted display and a robot device including a projector and at least one camera capable of efficiently cooperating a remote user and a local guide with each other.
- According to example embodiments, in a remote interaction method performed by a remote collaboration system comprising a robot device and a head mounted display, the robot device is located in a remote space and comprising a projector, a panoramic camera and a high resolution camera. The head mounted display is located in a local space apart from the remote space. A communication between the robot device and the head mounted display is established based on a communication request. The remote space is observed by the head mounted display based on first image information and second image information. The first image information is collected by the panoramic camera and received from the robot device. The second image information is collected by the high resolution camera and received from the robot device. Necessary information to be provided to the remote space is searched by the head mounted display based on a result of observing the remote space. The necessary information received from the head mounted display is provided by the robot device to the remote space using the projector.
- In an example embodiment, in observing the remote space, a panoramic image of the remote space may be displayed by the head mounted display based on the first image information. The panoramic image and a magnified image of a first region in the remote space may be displayed by the head mounted display together based on the first image information and the second image information when a screen magnifying interaction is performed while the head mounted display displays the panoramic image. The magnified image may be displayed by the head mounted display based on the second image information when the screen magnifying interaction is performed while the head mounted display displays the panoramic image and the magnified image together. A current screen image displayed on the head mounted display is stored when a screen saving interaction is performed.
- In an example embodiment, in observing the remote space, the panoramic image may be displayed by the head mounted display based on the first image information when a screen demagnifying interaction is performed while the head mounted display displays the panoramic image and the magnified image together. The panoramic image and the magnified image may be displayed by the head mounted display together based on the first image information and the second image information when the screen demagnifying interaction is performed while the head mounted display displays the magnified image.
- In an example embodiment, in observing the remote space, the magnified image displayed on the head mounted display may be removed when a removing interaction is performed while the head mounted display displays the panoramic image and the magnified image together, or a user interface object displayed on the head mounted display may be removed when the removing interaction is performed while the head mounted display displays the panoramic image or the magnified image with the user interface object. The user interface object may be used for controlling an operation of the head mounted display.
- In an example embodiment, in searching the necessary information, the local space may be observed by the head mounted display based on third image information collected by a first camera included in the head mounted display when a screen demagnifying interaction is performed while the head mounted display displays the panoramic image. A first searching operation may be performed on a plurality of first data stored in the head mounted display. The necessary information may be obtained based on at least one of the result of observing the remote space, a result of observing the local space and a result of the first searching operation.
- In an example embodiment, in searching the necessary information, a second searching operation may be performed on a plurality of second data stored in the robot device. A third searching operation may be performed on a plurality of third data associated with the necessary information by connecting an Internet network via the head mounted display. The necessary information may be obtained further based on at least one of a result of the second searching operation and a result of the third searching operation.
- In an example embodiment, in providing the necessary information to the remote space, the necessary information may be displayed by the projector on the remote space when the necessary information is provided as an image. A pointer may be displayed by the projector on the first region in the remote space when the necessary information is associated with the first region in the remote space.
- In an example embodiment, both the high resolution camera and the projector may be controlled toward the first region in the remote space when the head mounted display displays the panoramic image and the magnified image together or only the magnified image and when the necessary information is associated with the first region in the remote space and the projector displays the pointer on the first region in the remote space.
- In an example embodiment, the remote collaboration system may further comprise a controller that is located in the local space and controls an operation of the head mounted display. Direction controls of the high resolution camera and the projector may be performed based on head tracking information obtained by the head mounted display. A movement of the pointer displayed on the first region in the remote space may be controlled by the controller.
- According to example embodiments, a remote collaboration system comprises a robot device and a head mounted display. The robot device is located in a remote space and comprises a projector, a panoramic camera and a high resolution camera. The projector provides necessary information to the remote space. The panoramic camera collects first image information of the remote space. The high resolution camera collects second image information of the remote space. The head mounted display is located in a local space apart from the remote space. A communication between the robot device and the head mounted display is established based on a communication request. The head mounted display observes the remote space based on the first image information and the second image information. The head mounted display searches the necessary information based on a result of observing the remote space. The robot device provides the necessary information received from the head mounted display to the remote space using the projector.
- In the remote interaction method and remote collaboration system according to example embodiments, the robot device including the projector and the at least one camera may be used and may be controlled by the head mounted display. Thus, the collaboration between the remote user and the local guide may be efficiently performed. For example, the remote user may feel more comfortable without holding or wearing the device, may obtain the explanations from the local guide immediately displayed on the region having problems via the projector, and thus may have improved task efficiency. The local guide may control the robot device to watch the remote space and provide the explanations as desired, and may wear the head mounted display to observe the remote space with heightened immersion. The remote collaboration may be performed with high task efficiency and without inconvenience based on the screen change, the control of the robot device, etc. according to example embodiments even if the local guide wears the head mounted display.
- Illustrative, non-limiting example embodiments will be more clearly understood from the following detailed description taken in conjunction with the accompanying drawings.
-
FIG. 1 is a block diagram illustrating a remote collaboration system according to example embodiments. -
FIG. 2 is a diagram illustrating a prototype of a robot device included in a remote collaboration system according to example embodiments. -
FIG. 3 is a flow chart illustrating a remote interaction method performed by a remote collaboration system according to example embodiments. -
FIGS. 4A and 4B are diagrams for describing an operation associated with a communication request by a head mounted display included in a remote collaboration system according to example embodiments. -
FIGS. 5, 6A, 6B and 6C are diagrams for describing an operation of performing an event by a head mounted display included in a remote collaboration system according to example embodiments. -
FIGS. 7A, 7B, 7C and 8 are diagrams for describing an operation of observing a remote space by a head mounted display included in a remote collaboration system according to example embodiments. -
FIGS. 9, 10 and 11 are flow charts for describing an operation of changing a screen on a head mounted display included in a remote collaboration system according to example embodiments. -
FIG. 12 is a flow chart for describing an operation of controlling a robot device performed by a head mounted display included in a remote collaboration system according to example embodiments. -
FIG. 13 is a flow chart for describing an operation of searching information performed by a head mounted display included in a remote collaboration system according to example embodiments. -
FIGS. 14A, 14B and 14C are diagrams for describing an operation of displaying necessary information performed by a robot device included in a remote collaboration system according to example embodiments. - Various example embodiments will be described more fully with reference to the accompanying drawings, in which embodiments are shown. The present disclosure may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the present disclosure to those skilled in the art. Like reference numerals refer to like elements throughout this application.
- It will be understood that, although the terms first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are used to distinguish one element from another. For example, a first element could be termed a second element, and, similarly, a second element could be termed a first element, without departing from the scope of the present disclosure. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
- It will be understood that when an element is referred to as being “connected” or “coupled” to another element, it can be directly connected or coupled to the other element or intervening elements may be present. In contrast, when an element is referred to as being “directly connected” or “directly coupled” to another element, there are no intervening elements present. Other words used to describe the relationship between elements should be interpreted in a like fashion (e.g., “between” versus “directly between,” “adjacent” versus “directly adjacent,” etc.).
- The terminology used herein is for the purpose of describing particular embodiments and is not intended to be limiting of the present disclosure. As used herein, the singular forms “a,” “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises,” “comprising,” “includes” and/or “including,” when used herein, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other.
- Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which the present disclosure belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
-
FIG. 1 is a block diagram illustrating a remote collaboration system according to example embodiments. - Referring to
FIG. 1 , aremote collaboration system 10 includes arobot device 100 and a head mounted display (HMD) 200. Theremote collaboration system 10 may further include acontroller 250. - The
robot device 100 is located or disposed in a remote space and is a device which provides help, assistance, support or guidance to a remote user in the remote space. The head mounted display or head mounteddisplay device 200 is located or disposed in a local space that is apart from the remote space and freely controls therobot device 100 regardless of any motions or movements of the remote user. A local guide or a local helper in the local space may wear the head mounteddisplay 200. In other words, the head mounteddisplay 200 may be controlled by the local guide. - The
robot device 100 and the head mounteddisplay 200 may communicate with each other via anetwork 300. For example, therobot device 100 and the head mounteddisplay 200 may exchange image data IMG and/or voice data VOC with each other via thenetwork 300, and the head mounteddisplay 200 may transmit a command signal CS, an annotation signal AS, necessary information RI, etc. to therobot device 100 via thenetwork 300. - In some example embodiments, the
network 300 may include at least one of various wire/wireless communication networks. For example, thenetwork 300 may include a WiFi communication network and/or other wireless communication networks, such as a third generation (3G) communication network, a fourth generation (4G) communication network, a long term evolution (LTE) communication network, etc. - The
robot device 100 includes aprojector 110, apanoramic camera 120 and ahigh resolution camera 130. Therobot device 100 may further include acomputer 140, 150 and 160, and amotors transportation unit 170. Therobot device 100 may be referred to as a remote collaboration device. - The
projector 110 provides the necessary information RI to the remote space (e.g., to the remote user). The necessary information RI may be obtained by the head mounteddisplay 200 and may be transmitted from the head mounteddisplay 200 to theprojector 110 via thenetwork 300 and thecomputer 140. - The
panoramic camera 120 collects first image information IMG1 of the remote space. The first image information IMG1 may represent or correspond to a panoramic image or a 360 degree image of the remote space. - The
high resolution camera 130 collects second image information IMG2 of the remote space. The second image information IMG2 may represent or correspond to a high resolution image or a magnified image of a specific region in the remote space. For example, an angle of view of thehigh resolution camera 130 may be narrower than an angle of view of thepanoramic camera 120, and a resolution of thehigh resolution camera 130 may be higher than a resolution of thepanoramic camera 120. - In some example embodiments, if the
panoramic camera 120 has sufficiently high resolution to obtain the high resolution image or the magnified image of the specific region in the remote space, thehigh resolution camera 130 may be omitted. - The
computer 140 may control overall operations of therobot device 100. For example, thecomputer 140 may communicate with the head mounteddisplay 200 via thenetwork 300, may transmit the first image information IMG1 collected by thepanoramic camera 120 and the second image information IMG2 collected by thehigh resolution camera 130 to the head mounteddisplay 200, may receive the necessary information RI from the head mounteddisplay 200 to transmit the necessary information RI to theprojector 110, may control theprojector 110, thepanoramic camera 120 and thehigh resolution camera 130 based on the command signal CS and/or the annotation signal AS received from the head mounteddisplay 200, and may generate motor control signals MS1 and MS2 for controlling the 150 and 160.motors - The
motor 150 may control movements of at least one of theprojector 110, thepanoramic camera 120 and thehigh resolution camera 130 based on the motor control signal MS1. For example, themotor 150 may control a projecting direction of theprojector 110 or a lens direction of thehigh resolution camera 130. AlthoughFIG. 1 illustrates asingle motor 150 for controlling all of theprojector 110, thepanoramic camera 120 and thehigh resolution camera 130, theremote collaboration system 10 according to example embodiments may include a plurality of motors each of which controls a respective one of theprojector 110, thepanoramic camera 120 and thehigh resolution camera 130. - The
motor 160 may control thetransportation unit 170 based on the motor control signal MS2. A location of therobot device 100 in the remote space may be controlled based on an operation of thetransportation unit 170. - The head mounted
display 200 displays images such that the local guide observes the remote space based on the images displayed on the head mounteddisplay 200. For example, the head mounteddisplay 200 may display the panoramic image of the remote space based on the first image information IMG1, may display the high resolution image of the specific region in the remote space based on the second image information IMG2, or may display the panoramic image and the high resolution image together. - The head mounted
display 200 may include acamera 210. The head mounteddisplay 200 may display an image of the local space based on third image information collected by thecamera 210. - In addition, the head mounted
display 200 searches the necessary information RI that is to be provided to the remote space. Although not illustrated inFIG. 1 , the head mounteddisplay 200 may further include a processing device (e.g., a central processing unit (CPU)), a storage device (e.g., a memory), etc. for performing the searching operation. - The
controller 250 may be located or disposed in the local space and may control an operation of the head mounteddisplay 200. For example, the annotation signal AS may be transmitted to the head mounteddisplay 200 based on a control of thecontroller 250. For example, thecontroller 250 may include a touch pad, a touch screen or a dedicated control unit for the head mounteddisplay 200. - The command signal CS may be transmitted to the head mounted
display 200 based on a control of a user interface (UI) 220. For example, theuser interface 220 may include a graphic user interface (GUI) displayed on the head mounteddisplay 200. - Although
FIG. 1 illustrates an example where the command signal CS is generated by theuser interface 220 and the annotation signal AS is generated by thecontroller 250, inventive concepts are not limited thereto. For example, theuser interface 220 may generate the annotation signal AS, thecontroller 250 may generate the command signal CS, the command signal CS and/or the annotation signal AS may be generated based on head tracking information obtained by motions or movements of the local guide, or the command signal CS and/or the annotation signal AS may be generated while the searching operation for the necessary information RI is performed. - In the
remote collaboration system 10 according to example embodiments, the local guide may observe the remote space (e.g., a remote environment) using the robot device 100 (e.g., the remote collaboration device) located at a remote user side. For example, the panoramic image or the 360 degree image of the remote space may be obtained by thepanoramic camera 120 to observe the whole remote space, and the high resolution image or the magnified image of the specific region in the remote space may be obtained by thehigh resolution camera 130 to observe the specific region in the remote space, if necessary. Explanations of the local guide may be provided from theprojector 110 to the remote space. The movements of theprojector 110, thepanoramic camera 120, thehigh resolution camera 130 and thewhole robot device 100 may be controlled by the 150 and 160.motors - In addition, the local guide may wear the head mounted
display 200 to feel or realize the remote space vividly, finely or with details. The local guide may observe the local space (e.g., a local environment) using thecamera 210 included in the head mounteddisplay 200. A screen and/or the graphic user interface may be implemented in the head mounteddisplay 200, and the head tracking information and/or the controller 250 (e.g., the touch pad or the dedicated control unit) may be used for controlling the screen and/or the graphic user interface implemented in the head mounteddisplay 200. A remote collaboration interaction may be performed based on the head mounteddisplay 200. As will be described later, the remote collaboration interaction may include a control interaction for the head mounted display 200 (e.g., the screen and/or the graphic user interface implemented in the head mounted display 200) and a control interaction for therobot device 100. -
FIG. 2 is a diagram illustrating a prototype of a robot device included in a remote collaboration system according to example embodiments. - Referring to
FIG. 2 , a body including a plurality of receiving spaces is prepared, a projector, a panoramic camera, a high resolution camera, a mini PC (personal computer) and a motor control board are mounted on the plurality of receiving spaces, the body is equipped with motors, wheels and a battery, and thus a prototype of a robot device is manufactured. The mini PC and the wheels inFIG. 2 may correspond to thecomputer 140 and thetransportation unit 170 inFIG. 1 , respectively, and the motor control board and the battery are added for controlling the motors and supplying power to the motors, respectively. -
FIG. 3 is a flow chart illustrating a remote interaction method performed by a remote collaboration system according to example embodiments. - Referring to
FIGS. 1 and 3 , in a remote interaction method performed by theremote collaboration system 10 according to example embodiments, a communication between therobot device 100 located in the remote space and the head mounteddisplay 200 located in the local space is established based on a communication request (step S100). The communication request may be generated by one of therobot device 100 and the head mounteddisplay 200. For example, as will be described with reference toFIGS. 4A and 4B , the head mounteddisplay 200 may generate the communication request based on a control of the local guide. - The head mounted
display 200 receives the first image information IMG1 collected by thepanoramic camera 120 and the second image information IMG2 collected by thehigh resolution camera 130, and the remote space is observed by the head mounteddisplay 200 based on the first image information IMG1 and the second image information IMG2 (step S200). For example, as will be described with reference toFIGS. 7A, 7B and 7C , the head mounteddisplay 200 may display the panoramic image of the remote space based on the first image information IMG1, may display the high resolution image of the specific region in the remote space based on the second image information IMG2, or may display the panoramic image and the high resolution image together. - The necessary information RI that is to be provided to the remote space is searched by the head mounted
display 200 based on a result of observing the remote space (step S300). For example, the remote space may be observed to check, realize or detect problems in the remote space, and the necessary information RI may be searched to resolve the problems in the remote space. For example, as will be described with reference toFIG. 13 , the head mounteddisplay 200 may observe the remote space based on a control of the local guide or may search data stored in the head mounteddisplay 200, data stored in therobot device 100 or data in an Internet network to obtain the necessary information RI. - The
robot device 100 receives the necessary information RI from the head mounteddisplay 200, and the necessary information RI is provided by therobot device 100 to the remote space using the projector 110 (step S400). For example, as will be described with reference toFIGS. 14A, 14B and 14C , the necessary information RI may be provided as an image such as a picture, a sketch, etc., or may be provided as a pointer that is indicated on the specific region in the remote space. -
FIGS. 4A and 4B are diagrams for describing an operation associated with a communication request by a head mounted display included in a remote collaboration system according to example embodiments. - A configuration of screen GUIs of the head mounted
display 200 may include a first screen GUI for the communication request and a second screen GUI after the communication between therobot device 100 and the head mounteddisplay 200 is established.FIGS. 4A and 4B illustrate the first screen GUI for the communication request. - Referring to
FIGS. 1, 4A and 4B , after the local guide wears the head mounteddisplay 200, the head mounteddisplay 200 may be turned on, and the remote interaction method according to example embodiments may be performed or executed. As illustrated inFIG. 4A , an initial screen of the head mounteddisplay 200 may include the image of the local space displayed as a background. The image of the local space may be displayed based on third image information collected by thecamera 210. The initial screen of the head mounteddisplay 200 may further include a phone number input menu, a contact menu, a recent call menu and a bookmark list menu displayed as GUI forms (e.g., buttons, etc.). - As will be described with reference to
FIG. 5 , each button may be clicked using a HMD pointer to perform a corresponding function. For example, when a recent call button is clicked, a recent call list may be displayed on a screen of the head mounteddisplay 200. When a contact button is clicked, a contact list may be displayed on a screen of the head mounteddisplay 200. The contact list may include a plurality of contacts each of which includes a photo, a name, a phone number, an address, etc. - As illustrated in
FIG. 4B , when the contact button is clicked, the contact list may be displayed. When a specific contact (e.g., a contact of the remote user having therobot device 100 in the remote space) is clicked in the contact list, a communication request screen with a message (e.g., “call?”) may be displayed on a screen of the head mounteddisplay 200. When an OK button is clicked, the communication request may be transmitted from the head mounteddisplay 200 to therobot device 100. When therobot device 100 accepts the communication request, the communication between therobot device 100 and the head mounteddisplay 200 may be established. In other words, step S100 inFIG. 3 may be performed by the communication request from the head mounteddisplay 200. - When an add contact button is clicked, a screen for adding a new contact may be displayed. When a photo button is clicked, an image list for selecting an image to be added to the new contact may be displayed on a screen of the head mounted
display 200. - Bidirectional arrows illustrated in
FIGS. 4A and 4B may indicate a scrolling function. In other words, the bookmark list and the recent call list inFIG. 4A and the contact list and the image list inFIG. 4B may be scrolled. -
FIGS. 5, 6A, 6B and 6C are diagrams for describing an operation of performing an event by a head mounted display included in a remote collaboration system according to example embodiments. - An operation of performing an event by the head mounted
display 200 may indicate an operation of clicking or selecting a button or a menu in both the first screen GUI for the communication request and the second screen GUI after the communication between therobot device 100 and the head mounteddisplay 200 is established. In other words, an operation ofFIG. 5 may be performed in step S100 and may also be performed in steps S200, S300 and S400. - Referring to
FIGS. 5, 6A, 6B and 6C , a HMD pointer may be controlled to perform a specific event such that the HMD pointer is located or disposed on a specific UI object on a screen of the head mounted display 200 (step S110). For example, when the contact button inFIG. 4A is to be clicked, thecontroller 250 may move the HMD pointer on the screen of the head mounteddisplay 200 until the HMD pointer is located or disposed on the contact button. For example, thecontroller 250 may be a touch pad or a touch screen as illustrated inFIG. 6B , or a dedicated control unit for the head mounteddisplay 200 as illustrated inFIG. 6C . - A click operation may be performed, and a virtual beam may be emitted from a current location of the HMD pointer toward the specific UI object (step S120). For example, the click operation may be performed by the
controller 250. - When the virtual beam is not collided with the specific UI object (step S130: NO), it means that the location of the HMD pointer is not appropriate (e.g., the location of the HMD pointer does not correspond to the specific UI object), and thus steps S110 and S120 may be repeated.
- When the virtual beam is collided with the specific UI object (step S130: YES), an object event corresponding to the specific UI object may be performed (step S140). For example, when the contact button in
FIG. 4A is clicked, a screen of the head mounteddisplay 200 may be changed from the screen ofFIG. 4A to the screen ofFIG. 4B . - Although not illustrated in
FIGS. 5, 6A, 6B and 6C , when a phone number input button inFIG. 4A or the add contact button inFIG. 4B is clicked, a keyboard input may be performed on a screen of the head mounteddisplay 200. In some example embodiments, a virtual keyboard may be displayed on a screen of the head mounteddisplay 200, any button on the virtual keyboard may be clicked using the HMD pointer, and thus the keyboard input may be performed. In other example embodiments, when thecontroller 250 may be a touch pad or a touch screen, a keyboard may be displayed on the touch screen, the local guide may touch any button on the keyboard, and thus the keyboard input may be performed. In this example, the image of the local space may be displayed on a screen of the head mounteddisplay 200 to watch the keyboard on the touch screen via the head mounteddisplay 200. -
FIGS. 7A, 7B, 7C and 8 are diagrams for describing an operation of observing a remote space by a head mounted display included in a remote collaboration system according to example embodiments. - Referring to
FIGS. 1, 3, 7A, 7B and 7C , after step S100 of establishing the communication between therobot device 100 and the head mounteddisplay 200, the remote space may be observed via a screen of the head mounteddisplay 200. For example, as illustrated inFIG. 7A , the head mounteddisplay 200 may display the panoramic image of the remote space based on the first image information IMG1 collected by thepanoramic camera 120. For another example, as illustrated inFIG. 7C , the head mounteddisplay 200 may display the magnified image of the specific region in the remote space based on the second image information IMG2 collected by thehigh resolution camera 130. For still another example, as illustrated inFIG. 7B , the head mounteddisplay 200 may display the panoramic image and the magnified image together based on the first image information IMG1 and the second image information IMG2. As will be described with reference toFIGS. 9, 10 and 11 , a screen of the head mounteddisplay 200 may be changed in response to one of various interactions. - Referring to
FIGS. 2 and 8 ,FIG. 8 illustrates an example where an image obtained from the prototype of the robot device ofFIG. 2 is displayed on the head mounteddisplay 200. InFIG. 8 , the panoramic image and the magnified image (e.g., the high resolution image) are displayed together. A menu inFIG. 8 may correspond to UI objects. -
FIGS. 9, 10 and 11 are flow charts for describing an operation of changing a screen on a head mounted display included in a remote collaboration system according to example embodiments. - Referring to
FIGS. 1, 3 and 9 , immediately after step S100 of establishing the communication between therobot device 100 and the head mounteddisplay 200, the head mounteddisplay 200 may display the panoramic image as a default image or a default screen based on the first image information IMG1, as illustrated inFIG. 7A . When it is required to observe the specific region in the remote space in detail, a screen magnifying interaction may be performed (step S210), and a screen of the head mounteddisplay 200 may be changed in response to the screen magnifying interaction (step S220). - For example, when the screen magnifying interaction is performed while the head mounted
display 200 displays the panoramic image, the head mounteddisplay 200 may display the panoramic image and the magnified image together based on the first and second image information IMG1 and IMG2, as illustrated inFIG. 7B . When the screen magnifying interaction is performed while the head mounteddisplay 200 displays the panoramic image and the magnified image together, the head mounteddisplay 200 may display the magnified image based on the second image information IMG2, as illustrated inFIG. 7C . - In some example embodiments, the screen magnifying interaction may be performed by the head mounted
display 200. For example, when a motion of sticking a head out is performed by the head mounted display 200 (e.g., performed by the local guide), head tracking information corresponding to the motion of sticking the head out may be recognized as the screen magnifying interaction. In other example embodiments, the screen magnifying interaction may be performed by thecontroller 250. For example, a pinch-to-zoom motion on the touch screen or a motion of sweeping on the dedicated controller (e.g., performed by the local guide) may be recognized as the screen magnifying interaction. - Referring to
FIGS. 1, 3 and 10 , after the screen magnifying interaction is performed and a screen of the head mounteddisplay 200 is changed, a screen demagnifying interaction may be performed (step S230), and the screen of the head mounteddisplay 200 may be changed in response to the screen demagnifying interaction (step S240). - For example, when the screen demagnifying interaction is performed while the head mounted
display 200 displays the panoramic image and the magnified image together, the head mounteddisplay 200 may display the panoramic image based on the first image information IMG1. When the screen demagnifying interaction is performed while the head mounteddisplay 200 displays the magnified image, the head mounteddisplay 200 may display the panoramic image and the magnified image together based on the first and second image information IMG1 and IMG2. - In addition, when the screen demagnifying interaction is performed while the head mounted
display 200 displays the panoramic image, the head mounteddisplay 200 may display the image of the local space based on the third image information collected by thecamera 210. - In other words, the screen of the head mounted
display 200 may be changed in response to the screen magnifying interaction in the order of the image of the local space, the panoramic image of the remote space, the mixed image (e.g., the panoramic image and the magnified image together) of the remote space and the magnified image of the remote space. The screen of the head mounteddisplay 200 may be changed in response to the screen demagnifying interaction in the order of the magnified image of the remote space, the mixed image of the remote space, the panoramic image of the remote space and the image of the local space. - Referring to
FIGS. 1, 3 and 11 , while the head mounteddisplay 200 displays the panoramic image and the magnified image together, a removing interaction may be performed (step S250), and a screen of the head mounteddisplay 200 may be changed in response to the removing interaction (step S260). - For example, when the removing interaction is performed while the head mounted
display 200 displays the panoramic image and the magnified image together, the head mounteddisplay 200 may only display the panoramic image based on the first image information IMG1. In other words, the magnified image may be instantaneously removed in response to the removing interaction, and the screen of the head mounteddisplay 200 may be changed from a screen including the panoramic image and the magnified image together to a screen including the panoramic image only. - In some example embodiments, the removing interaction may be performed by the head mounted
display 200. For example, when a motion of tilting a head to the left or right is performed by the head mounted display 200 (e.g., performed by the local guide), head tracking information corresponding to the motion of tilting the head to the left or right may be recognized as the removing interaction. In other example embodiments, the removing interaction may be performed by thecontroller 250. - In some example embodiments, the removing interaction may be used for removing a user interface object displayed on the head mounted
display 200. For example, while the panoramic image is displayed on the head mounteddisplay 200, while the magnified image is displayed on the head mounteddisplay 200, or while the panoramic image and the magnified image are is displayed on the head mounteddisplay 200 together, the user interface object for controlling an operation of the head mounteddisplay 200 may be displayed on the screen of the head mounteddisplay 200. For example, the menu may be displayed with the panoramic image and the magnified image on the screen of the head mounteddisplay 200, as illustrated inFIG. 8 . In this case, the removing interaction may be performed (step S250), and the user interface object may be removed from the screen of the head mounteddisplay 200 in response to the removing interaction (step S260). - Although not illustrated in
FIG. 11 , after the removing interaction is performed, a restoring interaction may be performed, and the removed magnified image and/or the removed user interface object may be restored on the screen of the head mounteddisplay 200. - In some example embodiments, while various interactions described with reference to
FIGS. 9, 10 and 11 are performed, a screen saving interaction may be performed to store information associated with the remote space, and a current screen image displayed on the head mounteddisplay 200 may be stored in response to the screen saving interaction. For example, the current screen image may be stored in the head mounteddisplay 200 as a static image (e.g., a still image, a stopped image, etc.) or a dynamic image (e.g., a video, a moving image, etc.). -
FIG. 12 is a flow chart for describing an operation of controlling a robot device performed by a head mounted display included in a remote collaboration system according to example embodiments. - Referring to
FIGS. 1, 3 and 12 , the head mounteddisplay 200 may check the specific region to be magnified in the panoramic image (step S270). To obtain the magnified image of the specific region, head tracking information collected by the head mounteddisplay 200 may be transmitted to the robot device 100 (step S275). When the screen magnifying interaction described with reference toFIG. 9 is performed (step S280), a direction of thehigh resolution camera 130 may be controlled toward the specific region in the remote space based on the head tracking information and the screen magnifying interaction (step S285). In addition, a direction of theprojector 110 may be further controlled toward the specific region in the remote space. Thehigh resolution camera 130 and theprojector 110 may be controlled or driven independently (e.g., separately). -
FIG. 13 is a flow chart for describing an operation of searching information performed by a head mounted display included in a remote collaboration system according to example embodiments. - Referring to
FIGS. 1, 3 and 13 , to search the necessary information RI, the screen demagnifying interaction described with reference toFIG. 10 may be performed (step S310) while the head mounteddisplay 200 displays the panoramic image. The head mounteddisplay 200 may display the image of the local space based on the third image information collected by thecamera 210 in response to the screen demagnifying interaction to observe the local space (step S320). - In addition, at least one of a first searching operation performed on a plurality of first data stored in the head mounted
display 200, a second searching operation performed on a plurality of second data stored in therobot device 100, and a third searching operation performed on a plurality of third data associated with the necessary information RI by connecting an Internet network via the head mounteddisplay 200 may be additionally performed (step S330). - The necessary information RI may be obtained based on at least one of the result of observing the remote space, a result of observing the local space, a result of the first searching operation, a result of the second searching operation and a result of the third searching operation (step S340).
- In some example embodiments, at least one of observing the local space, the first searching operation, the second searching operation and the third searching operation may be omitted.
-
FIGS. 14A, 14B and 14C are diagrams for describing an operation of displaying necessary information performed by a robot device included in a remote collaboration system according to example embodiments. - Referring to
FIGS. 14A and 14B , when the necessary information RI is provided as an image, theprojector 110 may display the necessary information RI on the remote space. For example, when the necessary information RI corresponds to a static image or a dynamic image as illustrated inFIG. 14A , or when the necessary information RI corresponds to a sketch image (e.g., a drawing image) as illustrated inFIG. 14B , theprojector 110 may display the image on the remote space. For example, the image may be selected and transmitted by pushing a send button or based on a drag-and-drop scheme. A position of the image may be controlled according to eyes of the remote user. - Referring to
FIG. 14C , when the necessary information RI is associated with the specific region in the remote space, theprojector 110 may display a projector pointer on the specific region in the remote space. For example, a movement of the projector pointer on the specific region in the remote space may be performed by moving the HMD pointer by thecontroller 250. For another example, the sketch image inFIG. 14B may be drawn by moving the HMD pointer. - As described above, the remote interaction method (e.g., the remote collaboration) performed by the
remote collaboration system 10 according to example embodiments may include four steps of establishing the communication between therobot device 100 and the head mounteddisplay 200, observing the problems in the remote space, searching the information for resolving the problems, and explaining solution of the problems. Interactions in each step will be synthetically described with detailed examples. - In an initial operation time, to perform the interactions according to example embodiments, the local guide in the local space wears the head mounted
display 200 and controls therobot device 100 disposed in the remote space. The interactions in each step are performed based on the controls of the HMD and the remote collaboration device as described above. - First, in the step of establishing the communication, the local guide clicks the contact list on the screen of the head mounted
display 200 using thecontroller 250 to call the remote user. Alternatively, the local guide clicks the phone number input menu and inputs a phone number using a keyboard (e.g., a virtual keyboard) to call the remote user. An alarm indicating a call from the local guide rings or sounds in therobot device 100, and the remote user clicks a call button on therobot device 100 to answer the call. - Next, in the step of observing the problems in the remote space, after the communication between the
robot device 100 and the head mounteddisplay 200 is established, the local guide watches or observes the panoramic image of the remote space transmitted from therobot device 100 via the head mounteddisplay 200. When it is required to observe the specific region in the remote space in detail, the local guide performs the screen magnifying interaction to watch or observe the magnified image of the specific region in the remote space. The local guide captures the magnified image or records the image and/or voice to store information associated with the remote space. - For example, in a situation where the local guide helps the remote user to assemble home appliances, the remote space (or the remote environment) is observed by the local guide using the head mounted
display 200, and the screen magnifying interaction is performed to display the magnified image (or the high resolution image) and observe components of the home appliances in detail. The screen capture is performed to save the current screen image for reference before searching the necessary information RI. - Next, in the step of searching the information for resolving the problems, it is required to obtain the information for resolving the problems after the problems in the remote space are checked. The information may exist in the remote space, the remote collaboration device (e.g., the robot device 100), the head mounted
display 200, the local space, the Internet network, etc. Types of the information may include images, videos, texts, documents, etc. Information stored in the device is displayed, selected and transmitted based on HMD interactions and GUIs associated with images and/or videos. Information in the remote space is searched and captured based on the panoramic image and the magnified image. The screen demagnifying interaction is performed to display the image of the local space, and information in the local space is searched and captured based on the image of the local space. Information in the Internet network is displayed, searched and transmitted based on the Internet browser, the web surfing, the website capture, etc. - For example, in a situation where manuals or specifications for assembling the home appliances are located in the local space, the manuals or the specifications in the local space are observed by the local guide using the head mounted
display 200 and the screen demagnifying interaction. When the information is found out in the manuals or the specifications, the screen saving interaction is performed to store or capture the current screen image. The captured image is transmitted to therobot device 100. - Finally, in the step of explaining the solution of the problems, it is performed to provide the solution of the problems to the remote space and the remote user using the
projector 110 after the remote space is observed and the information for resolving the problems is obtained. Types of explaining the solution of the problems may include pointing, sketching, transmitting of the images and/or the videos, etc. The pointing and the sketching are performed to display the projector pointer and the sketch image on the remote space based on an accurate portion seen by the local guide. The transmitting of the images and/or the videos is performed to display the images and/or the videos on a desired portion in the remote space based on the drag-and-drop scheme. In addition, the displayed solution is stored, the displayed solution disappears when the direction of theprojector 110 is changed, and the disappeared solution appears again when the direction of theprojector 110 is returned to the original direction. - For example, the captured image of the manuals or the specifications is displayed on the screen (e.g., the UI screen) of the head mounted
display 200 as a thumbnail. The captured image is selected, dragged and dropped on the desired portion in the remote space by the local guide using the HMD pointer, and the captured image is displayed on the remote space using theprojector 110. - The
remote collaboration system 10 according to example embodiments may include therobot device 100 with the projector and the at least one camera and the head mounted display 200 (e.g., HMD interface) controlling therobot device 100. The camera functions may be performed to capture the 360 degree image (or the panoramic image) of the remote space and the high resolution image (or the magnified image) of the specific region in the remote space. Therobot device 100 may include a single camera for capturing both the 360 degree image and the high resolution image or two cameras (e.g., thepanoramic camera 120 and the high resolution camera 130) for capturing the 360 degree image and the high resolution image, respectively. Therobot device 100 may further include the motor for controlling the high resolution camera, and the direction of the high resolution camera may be controlled when the screen is changed from the 360 degree image to the high resolution image. Therobot device 100 may further include the motor for controlling the projector. Therobot device 100 may further include the computer for communications and computations, and the wheels and the battery for a moveable robot device. - The HMD interface may be implemented with a video see-through, an optical see-through, etc. The interaction interface may be implemented with a touch pad, a dedicated control unit, etc. The touch pad may be familiar with many users, may be used with various gestures, and may be utilized such as a keyboard. The dedicated control unit may include a sensor for tracking a movement of the dedicated control unit, buttons, a touch pad, etc., may provide an intuitive user experience, and may be useful for a 3D input. A keyboard input may be further implemented.
- In the
remote collaboration system 10 according to example embodiments, the local guide may wear the head mounteddisplay 200 and may communicate with the remote user using the screen GUIs of the head mounteddisplay 200 and the interaction interface. The head mounteddisplay 200 may exchange image, video and/or voice data with therobot device 100, and may transmit information and data for controlling the motor to therobot device 100. Therobot device 100 may move by the control of the local guide and the head mounteddisplay 200, and may display the information using theprojector 110. All data may be transmitted via the computer. - The inventive concept may be applied to various remote coaching services where the remote user in the remote space do not know how to handle some devices (e.g., electronic devices) or has some problems and the local guide in the local space provides assistance to the remote user to resolve the problems. For example, the inventive concept may be applied to various examples such as teaching a computer in which a computer screen is captured by the robot device and explanations are provided via the projector, playing a Korean chess with the local guide in which the projector pointer is displayed on a Korean chessboard, and showing a way to destination in which the destination and a current location are displayed on a map. For another example, the inventive concept may be applied to various examples such as cooking, educating, gardening, repairing, styling, interior designing, etc.
- The foregoing is illustrative of example embodiments and is not to be construed as limiting thereof. Although a few example embodiments have been described, those skilled in the art will readily appreciate that many modifications are possible in the example embodiments without materially departing from the novel teachings and advantages of the present disclosure. Accordingly, all such modifications are intended to be included within the scope of the present disclosure as defined in the claims. Therefore, it is to be understood that the foregoing is illustrative of various example embodiments and is not to be construed as limited to the specific example embodiments disclosed, and that modifications to the disclosed example embodiments, as well as other example embodiments, are intended to be included within the scope of the appended claims.
Claims (10)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR1020170135689A KR102082433B1 (en) | 2017-10-19 | 2017-10-19 | Remote collaboration system with projector-camera based robot device and head mounted display and remote interaction method using the same |
| KR10-2017-0135689 | 2017-10-19 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US10250845B1 US10250845B1 (en) | 2019-04-02 |
| US20190124294A1 true US20190124294A1 (en) | 2019-04-25 |
Family
ID=65898736
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/039,436 Active US10250845B1 (en) | 2017-10-19 | 2018-07-19 | Remote collaboration system with projector-camera based robot device and head mounted display, and remote interaction method using the same |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US10250845B1 (en) |
| KR (1) | KR102082433B1 (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10893245B1 (en) * | 2019-06-20 | 2021-01-12 | Lg Electronics Inc. | Method for projecting image and robot implementing the same |
| US20240275925A1 (en) * | 2022-07-28 | 2024-08-15 | Altec Industries, Inc. | Reducing latency in head-mounted display for the remote operation of machinery |
| US12485543B2 (en) | 2022-07-28 | 2025-12-02 | Altec Industries, Inc. | Cooperative high-capacity and high-dexterity manipulators |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20020019107A1 (en) * | 2000-08-11 | 2002-02-14 | Fujitsu Limited | Semiconductor device and its manufacture |
| US20040018967A1 (en) * | 1997-03-27 | 2004-01-29 | Enright Frederick M. | Ligand/lytic peptide compositions and methods of use |
| US20050002822A1 (en) * | 2000-11-27 | 2005-01-06 | Shigeki Takaya | Method of manufacturing ceramic resistor |
| US20160013976A1 (en) * | 2014-07-14 | 2016-01-14 | Futurewei Technologies, Inc. | Wireless Through Link Traffic Reduction |
Family Cites Families (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6845297B2 (en) | 2000-05-01 | 2005-01-18 | Irobot Corporation | Method and system for remote control of mobile robot |
| US6937266B2 (en) * | 2001-06-14 | 2005-08-30 | Microsoft Corporation | Automated online broadcasting system and method using an omni-directional camera system for viewing meetings over a computer network |
| SE0203908D0 (en) * | 2002-12-30 | 2002-12-30 | Abb Research Ltd | An augmented reality system and method |
| US7995090B2 (en) * | 2003-07-28 | 2011-08-09 | Fuji Xerox Co., Ltd. | Video enabled tele-presence control host |
| CA2720886A1 (en) * | 2010-11-12 | 2012-05-12 | Crosswing Inc. | Customizable virtual presence system |
| KR101444858B1 (en) * | 2013-09-03 | 2014-11-03 | 재단법인 실감교류인체감응솔루션연구단 | Telepresence apparatus |
| KR101870067B1 (en) * | 2014-08-25 | 2018-06-22 | 엑스 디벨롭먼트 엘엘씨 | Methods and systems for augmented reality to display virtual representations of robotic device actions |
| KR101553849B1 (en) | 2014-08-28 | 2015-09-17 | 주식회사 케이엠씨로보틱스 | Remote control apparatus of robot arm camera for video teleconferencing |
| KR20160057783A (en) * | 2014-11-14 | 2016-05-24 | 엘지전자 주식회사 | Mobile terminal and method for controlling the same |
| US20160358383A1 (en) | 2015-06-05 | 2016-12-08 | Steffen Gauglitz | Systems and methods for augmented reality-based remote collaboration |
| KR101683130B1 (en) * | 2015-08-12 | 2016-12-06 | 재단법인 실감교류인체감응솔루션연구단 | Method for providing user interface capable of allowing a user at local area and a user at remote area to interact each other, terminal, and computer-readable recording medium using the same |
| KR101763636B1 (en) | 2015-10-15 | 2017-08-02 | 한국과학기술원 | Method for collaboration using head mounted display |
-
2017
- 2017-10-19 KR KR1020170135689A patent/KR102082433B1/en active Active
-
2018
- 2018-07-19 US US16/039,436 patent/US10250845B1/en active Active
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20040018967A1 (en) * | 1997-03-27 | 2004-01-29 | Enright Frederick M. | Ligand/lytic peptide compositions and methods of use |
| US20020019107A1 (en) * | 2000-08-11 | 2002-02-14 | Fujitsu Limited | Semiconductor device and its manufacture |
| US20050002822A1 (en) * | 2000-11-27 | 2005-01-06 | Shigeki Takaya | Method of manufacturing ceramic resistor |
| US20160013976A1 (en) * | 2014-07-14 | 2016-01-14 | Futurewei Technologies, Inc. | Wireless Through Link Traffic Reduction |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10893245B1 (en) * | 2019-06-20 | 2021-01-12 | Lg Electronics Inc. | Method for projecting image and robot implementing the same |
| US20240275925A1 (en) * | 2022-07-28 | 2024-08-15 | Altec Industries, Inc. | Reducing latency in head-mounted display for the remote operation of machinery |
| US12485543B2 (en) | 2022-07-28 | 2025-12-02 | Altec Industries, Inc. | Cooperative high-capacity and high-dexterity manipulators |
Also Published As
| Publication number | Publication date |
|---|---|
| KR102082433B1 (en) | 2020-02-27 |
| US10250845B1 (en) | 2019-04-02 |
| KR20190043749A (en) | 2019-04-29 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9762851B1 (en) | Shared experience with contextual augmentation | |
| US9992429B2 (en) | Video pinning | |
| KR20170018724A (en) | Mobile terminal and method for controlling the same | |
| KR20190008610A (en) | Mobile terminal and Control Method for the Same | |
| KR20180133743A (en) | Mobile terminal and method for controlling the same | |
| KR20170089662A (en) | Wearable device for providing augmented reality | |
| KR20160092820A (en) | Mobile terminal and method for controlling the same | |
| KR20180094340A (en) | Mobile terminal and method for controlling the same | |
| US10250845B1 (en) | Remote collaboration system with projector-camera based robot device and head mounted display, and remote interaction method using the same | |
| US20220070412A1 (en) | Communication terminal, image communication system, method of displaying image, and recording medium | |
| KR20160012009A (en) | Mobile terminal and method for controlling the same | |
| US20170177174A1 (en) | Display apparatus and display method | |
| KR101672268B1 (en) | Exhibition area control system and control method thereof | |
| KR101959507B1 (en) | Mobile terminal | |
| EP3190503B1 (en) | An apparatus and associated methods | |
| KR20160001229A (en) | Mobile terminal and method for controlling the same | |
| EP3907598A1 (en) | Apparatus for enlarging and displaying screen and relaying same in real time, and operation method therefor | |
| KR20180014616A (en) | Mobile terminal and method for controlling the same | |
| KR20150105131A (en) | System and method for augmented reality control | |
| KR20180057034A (en) | Mobile terminal and method for controlling the same | |
| US11132926B2 (en) | Display control method and apparatus | |
| KR20170039994A (en) | Mobile terminal and method for controlling the same | |
| KR20150137836A (en) | Mobile terminal and information display method thereof | |
| KR20180037375A (en) | Terminal and method for controlling the same | |
| KR20180031238A (en) | Mobile terminal and method for controlling the same |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: KOREA INSTITUTE OF SCIENCE AND TECHNOLOGY, KOREA, Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AHN, SANG CHUL;KWON, JIN UK;REEL/FRAME:046396/0704 Effective date: 20180628 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |