WO2019178276A1 - Virtual reality system and method - Google Patents

Virtual reality system and method Download PDF

Info

Publication number
WO2019178276A1
WO2019178276A1 PCT/US2019/022121 US2019022121W WO2019178276A1 WO 2019178276 A1 WO2019178276 A1 WO 2019178276A1 US 2019022121 W US2019022121 W US 2019022121W WO 2019178276 A1 WO2019178276 A1 WO 2019178276A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
headset
image
recited
wearing
Prior art date
Application number
PCT/US2019/022121
Other languages
French (fr)
Inventor
Ronald Winston
Michael TIGAR
Original Assignee
Ronald Winston
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ronald Winston filed Critical Ronald Winston
Priority to CA3089885A priority Critical patent/CA3089885A1/en
Priority to CN201980018658.6A priority patent/CN111936910A/en
Priority to JP2020546898A priority patent/JP7250809B2/en
Priority to CN202310667894.9A priority patent/CN116700489A/en
Priority to EP19766538.3A priority patent/EP3729177A4/en
Publication of WO2019178276A1 publication Critical patent/WO2019178276A1/en
Priority to JP2023044888A priority patent/JP2023078342A/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B27/0172Head mounted characterised by optical features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/77Retouching; Inpainting; Scratch removal
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/40Support for services or applications
    • H04L65/403Arrangements for multi-party communication, e.g. for conferences
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B2027/0178Eyeglass type
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Definitions

  • This description generally relates to methods and devices for capturing and processing two-dimensional (2D) and three-dimensional (3D) images.
  • VR Virtual Reality
  • VR headsets make it challenging to create a complete picture of the people participating in the experience.
  • MR Mixed Reality
  • VR a related medium that shares the virtual context of a VR user in a two dimensional video format allowing other viewers to get a feel for the user’ s virtual experience.
  • MR facilitates sharing, the headset continues to block facial expressions and eye gaze, presenting a significant hurdle to a fully engaging experience and complete view of the person in VR.
  • the general purpose of this project is to provide a live VR chat experience.
  • VR chat applications make use of animated avatars.
  • An object of the present invention is to replace the avatars with live video without the image of a user’s headsets which would be replaced visually by a 3D model of the users’ head (e.g., recorded previously without the headset being worn by a user).
  • Another object of the present invention is each user may adjust their level of realism (e.g., from avatar to live 3D video image).
  • FIG. 1 illustrates an example of a computer device and a mobile device that can he used to implement the techniques described herein;
  • FIG. 2 depicts users wearing headsets and the view a headset wearing user perceives of another headset wearing user
  • FIG. 3 depicts the process for removing the image of a headset from a user to create a VR of the user as if the headset was removed
  • FIGS. 4-7 illustrate portions of the process for creating a VR chat between users in accordance with the described illustrated embodiments.
  • the illustrated embodiments discussed below are preferably a software algorithm, program or code residing on computer useable medium having control logic for enabling execution on a machine having a computer processor.
  • the machine typically includes memory storage configured to provide output from execution of the computer algorithm or program.
  • the term“software” is meant to be synonymous with any code or program that can be in a processor of a host computer, regardless of whether the implementation is in hardware, firmware or as a software computer product available on a disc, a memory storage device, or for download from a remote machine.
  • the embodiments described herein include such software to implement the equations, relationships and algorithms described above.
  • One skilled in the art will appreciate further features and advantages of the illustrated embodiments based on the above- described embodiments. Accordingly, the illustrated embodiments are not to be limited by what has been particularly shown and described, except as indicated by the appended claims. All publications and references cited herein are expressly incorporated herein by reference in their entirety.
  • FIG. 1 depicts an exemplary general-purpose computing system 600 and a generic mobile computer device 650, which may be used with the techniques described here. It is to be appreciated computing system 600 and mobile device 650 are to be understood as an exemplary embodiments in which one or more components thereof may be used with the illustrated embodiments of the present invention as described herein.
  • Computing device 600 includes a processor 602, memory 604, a storage device 606, a high-speed interface 608 connecting to memory 604 and high-speed expansion ports 610, and a low speed interface 612 connecting to low speed bus 614 and storage device 606.
  • Each of the components 602, 604, 606, 608, 610, and 612 are interconnected using various busses, and may be mounted on a common motherboard or in other manners as appropriate.
  • the processor 602 can process instructions for execution within the computing device 600, including instructions stored in the memory 604 or on the storage device 606 to display graphical information for a GUI on an external input/output device, such as display 616 coupled to high speed interface 608.
  • multiple processors and/or multiple buses may be used, as appropriate, along with multiple memories and types of memory.
  • multiple computing devices 600 may be connected, with each device providing portions of the necessary operations (e.g., as a server bank, a group of blade servers, or a multi-processor system).
  • the memory 604 stores information within the computing device 600.
  • the memory 604 is a volatile memory unit or units.
  • the memory 604 is a non-volatile memory unit or units.
  • the memory 604 may also be another form of computer-readable medium, such as a magnetic or optical disk.
  • the storage device 606 is capable of providing mass storage for the computing device 600.
  • the storage device 606 may be or contain a computer-readable medium, such as a floppy disk device, a hard disk device, an optical disk device, or a tape device, a flash memory or other similar solid state memory device, or an array of devices, including devices in a storage area network or other configurations.
  • a computer program product can be tangibly embodied in an information carrier.
  • the computer program product may also contain instructions that, when executed, perform one or more methods, such as those described above.
  • the information carrier is a computer- or machine-readable medium, such as the memory 604, the storage device 606, or memory on processor 602.
  • the high speed controller 608 manages bandwidth-intensive operations for the computing device 600, while the low speed controller 612 manages lower bandwidth intensive operations. Such allocation of functions is exemplary only.
  • the high-speed controller 608 is coupled to memory 604, display 616 (e.g., through a graphics processor or accelerator), and to high-speed expansion ports 610, which may accept various expansion cards (not shown).
  • low-speed controller 612 is coupled to storage device 606 and low-speed expansion port 614.
  • the low-speed expansion port which may include various communication ports (e.g., USB, Bluetooth, Ethernet, wireless Ethernet) may be coupled to one or more input/output devices, such as a keyboard, a pointing device, a scanner, or a networking device such as a switch or router, e.g., through a network adapter.
  • input/output devices such as a keyboard, a pointing device, a scanner, or a networking device such as a switch or router, e.g., through a network adapter.
  • the computing device 600 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as a standard server 620, or multiple times in a group of such servers. It may also be implemented as part of a rack server system 624. In addition, it may be implemented in a personal computer such as a laptop computer 622. Alternatively, components from computing device 600 may be combined with other components in a mobile device (not shown), such as device 650. Each of such devices may contain one or more of computing device 600, 650, and an entire system may be made up of multiple computing devices 600, 650 communicating with each other.
  • Computing device 650 includes a processor 652, memory 664, an input/output device such as a display 654, a communication interface 666, and a transceiver 668, among other components.
  • the device 650 may also be provided with a storage device, such as a microdrive or other device, to provide additional storage.
  • a storage device such as a microdrive or other device, to provide additional storage.
  • Each of the components 650, 652, 664, 654, 666, and 668 are interconnected using various buses, and several of the components may be mounted on a common motherboard or in other manners as appropriate.
  • the processor 652 can execute instructions within the computing device 650, including instructions stored in the memory 664.
  • the processor may be implemented as a chipset of chips that include separate and multiple analog and digital processors.
  • the processor may provide, for example, for coordination of the other components of the device 650, such as control of user interfaces, applications ran by device 650, and wireless communication by device 650.
  • Processor 652 may communicate with a user through control interface 658 and display interface 656 coupled to a display 654.
  • the display 654 may be, for example, a TFT LCD (Thin-Film-Transistor Liquid Crystal Display) or an OLED (Organic Light Emitting Diode) display, or other appropriate display technology.
  • the display interface 656 may comprise appropriate circuitry for driving the display 654 to present graphical and other information to a user.
  • the control interface 658 may receive commands from a user and convert them for submission to the processor 652.
  • an external interface 662 may be provide in communication with processor 652, so as to enable near area communication of device 650 with other devices.
  • External interface 662 may provide, for example, for wired communication in some implementations, or for wireless communication in other implementations, and multiple interfaces may also be used.
  • the memory 664 stores information within the computing device 650.
  • the memory 664 can be implemented as one or more of a computer-readable medium or media, a volatile memory unit or units, or a non-volatile memory unit or units.
  • Expansion memory 674 may also be provided and connected to device 650 through expansion interface 672, which may include, for example, a SIMM (Single In Line Memory Module) card interface. Such expansion memory 674 may provide extra storage space for device 650, or may also store applications or other information for device 650. Specifically, expansion memory 674 may include instructions to carry out or supplement the processes described above, and may include secure information also. Thus, for example, expansion memory 674 may be provide as a security module for device 650, and may be programmed with instructions that permit secure use of device 650. In addition, secure applications may be provided via the SIMM cards, along with additional information, such as placing identifying information on the SIMM card in a non-hackable manner.
  • SIMM Single In Line Memory Module
  • the memory may include, for example, flash memory and/or NVRAM memory, as discussed below.
  • a computer program product is tangibly embodied in an information carrier.
  • the computer program product contains instructions that, when executed, perform one or more methods, such as those described above.
  • the information carrier is a computer- or machine-readable medium, such as the memory 664, expansion memory 674, or memory on processor 652, that may be received, for example, over transceiver 668 or external interface 662.
  • Device 650 may communicate wirelessly through communication interface 666, which may include digital signal processing circuitry where necessary.
  • Communication interface 666 may provide for communications under various modes or protocols, such as GSM voice calls, SMS, EMS, or MMS messaging, CDMA, TDMA, PDC, WCDMA, CDMA2000, or GPRS, among others. Such communication may occur, for example, through radio-frequency transceiver 668. In addition, short- range communication may occur, such as using a Bluetooth, Wi-Fi, or other such transceiver (not shown). In addition, GPS (Global Positioning System) receiver module 670 may provide additional navigation- and location-related wireless data to device 650, which may be used as appropriate by applications running on device 650.
  • GPS Global Positioning System
  • Device 650 may also communicate audibly using audio codec 660, which may receive spoken information from a user and convert it to usable digital information. Audio codec 660 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of device 650. Such sound may include sound from voice telephone calls, may include recorded sound (e.g., voice messages, music files, etc.) and may also include sound generated by applications operating on device 650.
  • Audio codec 660 may receive spoken information from a user and convert it to usable digital information. Audio codec 660 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of device 650. Such sound may include sound from voice telephone calls, may include recorded sound (e.g., voice messages, music files, etc.) and may also include sound generated by applications operating on device 650.
  • the computing device 650 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as a cellular telephone 680. It may also be implemented as part of a smart phone 682, personal digital assistant, or other similar mobile device.
  • a programmable system including at least one programmable processor, which may be special or general purpose, coupled to receive data and instructions from, and to transmit data and instructions to, a storage system, at least one input device, and at least one output device.
  • the systems and techniques described here can be implemented on a computer having a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) (and other suitable display technology) for displaying information to the user and a keyboard and a pointing device (e.g., a mouse or a trackball) by which the user can provide input to the computer.
  • a display device e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor
  • keyboard and a pointing device e.g., a mouse or a trackball
  • Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user can be received in any form, including acoustic, speech, or tactile input.
  • the systems and techniques described here can be implemented in a computing system that includes a back end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front end component (e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back end, middleware, or front end components.
  • the components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN”), a wide area network (“WAN”), and the Internet.
  • the computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
  • the computing devices depicted in FIG. 1 interface with a virtual reality (VR headset 690).
  • VR headset 690 virtual reality
  • one or more sensors are included in the VR headset to virtual remove the appearance of the headset 690 when one user interacts with another user.
  • the headset sensors 690 may include one or more LED’s 691 that illuminates an eye of a user while tracking the user’s pupil and one or more micro cameras 693 for recording the user’s eye/pupil.
  • an embodiment of the present invention creates a model of the user’s face 704 without a headset 690 using a camera device 400 as a proxy for the user’s hidden face when a headset 690 is worn.
  • This proxy is used to synthesize the face in a Mixed Reality (MR) video, thereby creating an impression of the headset 690 being removed when one using interacts with another user in a VR realm.
  • this initial calibration step may require the user 704 to sit in front of a color+depth camera 400 and a monitor, and then track a marker on the monitor with their eyes.
  • This calibration procedure acquires a 3D face model of the user 704, and create a database that maps appearance images (or textures) to different eye-gaze directions and blinks (via sensors 691 and 693).
  • the aforementioned eye-gaze database i.e. the face model preferably with textures indexed by eye-gaze
  • the construct of the face can also be altered by using suitable wire frame software technology.
  • one or more input devices in addition to the computing device can be rendered in a computer generated, 3D environment.
  • the rendered input devices e.g., the rendered mouse, the rendered keyboard
  • Computing device 600 is intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers.
  • Computing device 650 is intended to represent various forms of mobile devices, such as personal digital assistants, cellular telephones, smart phones, and other similar computing devices.
  • the components shown here, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the inventions described and/or claimed in this document.
  • FIG. 1 With the exemplary computing system environment 100 of FIG. 1 being generally shown and discussed above, and with reference now to FIGS. 2-7, described below is implementation of various exemplary embodiments.
  • an environment of use for an illustrated embodiment of the present invention relates to a system 600 that enables online chat in VR preferably with live video.
  • a system 600 that enables online chat in VR preferably with live video.
  • exemplary embodiments include one or more components of the above described system 600 including a VR system such as Oculus RiftTM, one or more webcams 400 per user and preferably a green screen.
  • a remote user 700 wearing headset 690 for example would view a live video of another user 702 (wearing headset 690), minus the headset 690 that the local user is wearing 704, in VR.
  • the remote user 702 wearing headset 690 would view a live video of the user 700 (wearing headset 690), minus the headset 690 that the local user is wearing 706, in VR.
  • the headset visual for a remote user is replaced by an overlay of pre-recorded static images/video of the local users’ head.
  • the live video effect is created by chromakey (“green screen”) technology preferably using a standard 2D webcam 400 (Fig. 4).
  • each user 704, 706 is preferably placed in front of a portable green screen (402). It is to be appreciated this may be negated by certain software techniques or alternatively it may be possible to determine the users outline without a green screen by using a 3D camera.
  • two or more users (704 and 706) may participate in a live online chat using a special webcam 400 at each user location.
  • Each user (704, 706) preferably wears a VR headset 690 (700, 702) whereby each user (704, 706) views a superimposed and prerecorded image of the other user’s head (706, 704).
  • the broadcast image is preferably a composite of a live video and prerecorded image.
  • the live video preferably includes the body up to just above a user’s mouth region (FIG. 3) (e.g., where the VR headset 690 generally resides).
  • the recorded image is preferably the nose, eyes and top of head of a user (704, 706). It is to be appreciated that it should be possible to achieve a live chat with more than two persons, such as a corporate teleconference.
  • This recorded image may also be created prior using a special form of photogrammetry.
  • This is the technique where multiple 2D images are stitched together and processed to create a 3D model and texture.
  • this process is expensive, time consuming, and requires special hardware and multiple cameras.
  • implementation of inexpensive stereoscopic cameras may be utilized consisting of two cameras facing the user, each with either stereoscopic or monoscopic vision. For instance, the user would rotate in their chair while the cameras record and capture the scene. Post processing of the stereoscopic or monoscopic camera output would be used to generate the 3D model.
  • each user 704 (FIG. 4), 706 (FIG. 5) without a headset is photographed by webcam 400.
  • each user (704, 706) then wears a headset 690 having the aforementioned LED sensors 691 and 693 for detecting eye gaze, movement and blinking.
  • the sensors 691 illuminates one eye of the user inside the headset 690 to track pupil movement while the micro video camera sensor 693 records a user’s pupil.
  • the processing chip 602 is preferably configured and operational to restore the moving pupil to the face by“stitching” it in preferably via software to complete the image of the face currently obscured by the headset 690.
  • both users 704 and 706 can view 3D images of each other (via their respective headset 690) in real time as if the face of the other user 704, 706 were in front of them and unobscured by a headset 690.
  • the frowns expression of each user 704, 706 may also be tracked via software processing on the processing chip 602.
  • the present invention enables virtual projection of a webcam video stream of a user in a video chat, minus the users’ headset. This is accomplished by modeling the users head in 3D using modelling techniques, such as photogrammetry. For instance, the top portion of a users’ head is replaced by this model in real time, so the other participating user only views the body, head, face and eyes of a participating user, minus the headset 690.
  • multiple users e.g., greater than two users
  • each user is modelled using one or more stationary stereoscopic 3D cameras (e.g., StereoLabs ZED camera, Microsoft Kinect, Intel RealSense line of 3D cameras, etc.).
  • the 3D camera techniques additional may calculate a user’s position, providing VR cooperative interaction. For instance, two users could play a game of catch on Mars or two people could take a virtual walk together in the Grand Canyon.
  • AR Augmented Reality
  • AR is to be understood to be a technology that layers computer-generated
  • An illustrated embodiment includes AR being developed for use as an app preferably a the headset 690 to blend digital components into the real world in such a way that they enhance one another.
  • two or more users, each in a different location, wearing the present invention headset 690 may use/operate the headsets 690 so as to appear next to one another in a chosen setting (e.g., skydiving) and/or location (e.g., Paris) whereby they not only in real-time visualize the same setting together (e.g., a street in Paris, France) but they view each other as if the headset 690 were removed, in accordance with the teachings herein.
  • a chosen setting e.g., skydiving
  • location e.g., Paris
  • additional embodiments of the computing system 600 and/or headset 690 may be configured to provide for real-time image stitching preferably using at least two separate video feeds and/or a single video feed preferably in conjunction with a predetermined still-frame Image Cache.
  • eye-tracking is performed via a camera mounted inside a VR display device enabling real-time selection of cached images and/or“blended” images to be derived from an initial“setup” phase (Image Cache) to be used in an image stitching step as the“eye region” of a user’ s face is typically occluded due to the VR display device thus providing real-time 3D face tracking and reconstruction using a final
  • the face/result is preferably composited onto either a real-time video and/or a depth-extraction of the user.
  • the user-body is preferably extracted from its background using either green screen technology and/or depth-extraction techniques.
  • a fully-reconstructed user may be composited into a multitude of applications such as (but not limited to): 360-degree video, VR Chat Rooms, and other applicable applications.
  • miscellaneous technology e.g., SDK’s, API’s, etc.
  • miscellaneous technology include (but are not to be limited to): EmguCV, OpenCV and MATLAB; ZED Camera with SDK; ARKit; Primary Programming Languages: C# and C++; Unity (game engine); and Oculus and/or Vive (VR Display Device).

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Computer Hardware Design (AREA)
  • Computer Graphics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Optics & Photonics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)
  • Controls And Circuits For Display Device (AREA)

Abstract

A computer apparatus and process for creating Virtual Reality (VR) interactions between two or more users wherein at least one user is wearing a VR headset. Provided is at least one VR headset device having one or more integrated sensor components and a camera device for at least capturing an image of a user wearing the VR headset device. A VR imaging computer device is communicatively coupled to each of the at least one VR headset device and camera device. The VR imaging computer creates a two-dimensional (2D) image of at least one user wearing a VR headset using a camera device wherein positioned in the headset is one or more sensors. A user's pupil movement is tracked using the one or more sensors such that a headset image from the created 2D image is removed from the user wearing the headset. The detected and recorded pupil image of the user rendered by the one or more sensors of the VR headset is then stitched in the captured image of the user so as to be superimposed in a region in which the headset image was removed from the created 2D image of the user such that another user can view a VR image of the user as if the user was not wearing the headset.

Description

IN THE UNITED STATES PATENT AND TRADEMARK OFFICE
VIRTUAL REALITY SYSTEM AND METHOD
CROSS REFERENCE TO RELATED APPLICATIONS
This application claims priority to U.S. Patent Application Serial No.
62/642,488 filed March 13, 2018 which is incorporated herein by reference in its entirety.
FIELD OF THE INVENTION
This description generally relates to methods and devices for capturing and processing two-dimensional (2D) and three-dimensional (3D) images.
BACKGROUND OF THE INVENTION
Virtual Reality (VR) enables remarkably immersive experiences, offering new ways to view the world and the ability to explore novel environments, both real and imaginary. However, compared to physical reality, sharing these experiences with others can be difficult, as VR headsets make it challenging to create a complete picture of the people participating in the experience.
Some of this disconnect is alleviated by Mixed Reality (MR), a related medium that shares the virtual context of a VR user in a two dimensional video format allowing other viewers to get a feel for the user’ s virtual experience. Even though MR facilitates sharing, the headset continues to block facial expressions and eye gaze, presenting a significant hurdle to a fully engaging experience and complete view of the person in VR.
SUMMARY OF THE INVENTION
The purpose and advantages of the below described illustrated embodiments will be set forth in and apparent from the description that follows. Additional advantages of the illustrated embodiments will be realized and attained by the devices, systems and methods particularly pointed out in the written description and claims hereof, as well as from the appended drawings.
The general purpose of this project is to provide a live VR chat experience. Currently VR chat applications make use of animated avatars. An object of the present invention is to replace the avatars with live video without the image of a user’s headsets which would be replaced visually by a 3D model of the users’ head (e.g., recorded previously without the headset being worn by a user). Another object of the present invention is each user may adjust their level of realism (e.g., from avatar to live 3D video image).
BRIEF DESCRIPTION OF THE DRAWINGS
The accompanying appendices and/or drawings illustrate various non-limiting, example, inventive aspects in accordance with the present disclosure:
FIG. 1 illustrates an example of a computer device and a mobile device that can he used to implement the techniques described herein;
FIG. 2 depicts users wearing headsets and the view a headset wearing user perceives of another headset wearing user; FIG. 3 depicts the process for removing the image of a headset from a user to create a VR of the user as if the headset was removed; and
FIGS. 4-7 illustrate portions of the process for creating a VR chat between users in accordance with the described illustrated embodiments.
DETAILED DESCRIPTION OF CERTAIN EMBODIMENTS
The illustrated embodiments are now described more fully with reference to the accompanying drawings wherein like reference numerals identify similar structural/functional features. The illustrated embodiments are not limited in any way to what is illustrated as the illustrated embodiments described below are merely exemplary, which can be embodied in various forms, as appreciated by one skilled in the art. Therefore, it is to be understood that any structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a basis for the claims and as a representation for teaching one skilled in the art to variously employ the discussed embodiments. Furthermore, the terms and phrases used herein are not intended to be limiting but rather to provide an understandable description of the illustrated embodiments.
Where a range of values is provided, it is understood that each intervening value, to the tenth of the unit of the lower limit unless the context clearly dictates otherwise, between the upper and lower limit of that range and any other stated or intervening value in that stated range is encompassed within the illustrated embodiments. The upper and lower limits of these smaller ranges may independently be included in the smaller ranges is also encompassed within the illustrated embodiments, subject to any specifically excluded limit in the stated range. Where the stated range includes one or both of the limits, ranges excluding either both of those included limits are also included in the illustrated embodiments.
Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. Although any methods and materials similar or equivalent to those described herein can also be used in the practice or testing of the illustrated embodiments, exemplary methods and materials are now described. All publications mentioned herein are incorporated herein by reference to disclose and describe the methods and/or materials in connection with which the publications are cited.
It must be noted that as used herein and in the appended claims, the singular forms“a”,“an,” and“the” include plural referents unless the context clearly dictates otherwise. Thus, for example, reference to“a stimulus” includes a plurality of such stimuli and reference to“the signal” includes reference to one or more signals and equivalents thereof known to those skilled in the art, and so forth.
The publications discussed herein are provided solely for their disclosure prior to the filing date of the present application. Nothing herein is to be construed as an admission that the illustrated embodiments are not entitled to antedate such publication by virtue of prior invention. Further, the dates of publication provided may differ from the actual publication dates which may need to be independently confirmed.
It is to be appreciated the illustrated embodiments discussed below are preferably a software algorithm, program or code residing on computer useable medium having control logic for enabling execution on a machine having a computer processor. The machine typically includes memory storage configured to provide output from execution of the computer algorithm or program.
As used herein, the term“software” is meant to be synonymous with any code or program that can be in a processor of a host computer, regardless of whether the implementation is in hardware, firmware or as a software computer product available on a disc, a memory storage device, or for download from a remote machine. The embodiments described herein include such software to implement the equations, relationships and algorithms described above. One skilled in the art will appreciate further features and advantages of the illustrated embodiments based on the above- described embodiments. Accordingly, the illustrated embodiments are not to be limited by what has been particularly shown and described, except as indicated by the appended claims. All publications and references cited herein are expressly incorporated herein by reference in their entirety.
Turning now descriptively to the drawings, in which similar reference characters denote similar elements throughout the several views, Fig. 1 depicts an exemplary general-purpose computing system 600 and a generic mobile computer device 650, which may be used with the techniques described here. It is to be appreciated computing system 600 and mobile device 650 are to be understood as an exemplary embodiments in which one or more components thereof may be used with the illustrated embodiments of the present invention as described herein.
Computing device 600 includes a processor 602, memory 604, a storage device 606, a high-speed interface 608 connecting to memory 604 and high-speed expansion ports 610, and a low speed interface 612 connecting to low speed bus 614 and storage device 606. Each of the components 602, 604, 606, 608, 610, and 612, are interconnected using various busses, and may be mounted on a common motherboard or in other manners as appropriate. The processor 602 can process instructions for execution within the computing device 600, including instructions stored in the memory 604 or on the storage device 606 to display graphical information for a GUI on an external input/output device, such as display 616 coupled to high speed interface 608. In other implementations, multiple processors and/or multiple buses may be used, as appropriate, along with multiple memories and types of memory. In addition, multiple computing devices 600 may be connected, with each device providing portions of the necessary operations (e.g., as a server bank, a group of blade servers, or a multi-processor system).
The memory 604 stores information within the computing device 600. In one implementation, the memory 604 is a volatile memory unit or units. In another implementation, the memory 604 is a non-volatile memory unit or units. The memory 604 may also be another form of computer-readable medium, such as a magnetic or optical disk.
The storage device 606 is capable of providing mass storage for the computing device 600. In one implementation, the storage device 606 may be or contain a computer-readable medium, such as a floppy disk device, a hard disk device, an optical disk device, or a tape device, a flash memory or other similar solid state memory device, or an array of devices, including devices in a storage area network or other configurations. A computer program product can be tangibly embodied in an information carrier. The computer program product may also contain instructions that, when executed, perform one or more methods, such as those described above. The information carrier is a computer- or machine-readable medium, such as the memory 604, the storage device 606, or memory on processor 602. The high speed controller 608 manages bandwidth-intensive operations for the computing device 600, while the low speed controller 612 manages lower bandwidth intensive operations. Such allocation of functions is exemplary only. In one implementation, the high-speed controller 608 is coupled to memory 604, display 616 (e.g., through a graphics processor or accelerator), and to high-speed expansion ports 610, which may accept various expansion cards (not shown). In the implementation, low-speed controller 612 is coupled to storage device 606 and low-speed expansion port 614. The low-speed expansion port, which may include various communication ports (e.g., USB, Bluetooth, Ethernet, wireless Ethernet) may be coupled to one or more input/output devices, such as a keyboard, a pointing device, a scanner, or a networking device such as a switch or router, e.g., through a network adapter.
The computing device 600 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as a standard server 620, or multiple times in a group of such servers. It may also be implemented as part of a rack server system 624. In addition, it may be implemented in a personal computer such as a laptop computer 622. Alternatively, components from computing device 600 may be combined with other components in a mobile device (not shown), such as device 650. Each of such devices may contain one or more of computing device 600, 650, and an entire system may be made up of multiple computing devices 600, 650 communicating with each other.
Computing device 650 includes a processor 652, memory 664, an input/output device such as a display 654, a communication interface 666, and a transceiver 668, among other components. The device 650 may also be provided with a storage device, such as a microdrive or other device, to provide additional storage. Each of the components 650, 652, 664, 654, 666, and 668, are interconnected using various buses, and several of the components may be mounted on a common motherboard or in other manners as appropriate.
The processor 652 can execute instructions within the computing device 650, including instructions stored in the memory 664. The processor may be implemented as a chipset of chips that include separate and multiple analog and digital processors. The processor may provide, for example, for coordination of the other components of the device 650, such as control of user interfaces, applications ran by device 650, and wireless communication by device 650.
Processor 652 may communicate with a user through control interface 658 and display interface 656 coupled to a display 654. The display 654 may be, for example, a TFT LCD (Thin-Film-Transistor Liquid Crystal Display) or an OLED (Organic Light Emitting Diode) display, or other appropriate display technology. The display interface 656 may comprise appropriate circuitry for driving the display 654 to present graphical and other information to a user. The control interface 658 may receive commands from a user and convert them for submission to the processor 652. In addition, an external interface 662 may be provide in communication with processor 652, so as to enable near area communication of device 650 with other devices. External interface 662 may provide, for example, for wired communication in some implementations, or for wireless communication in other implementations, and multiple interfaces may also be used.
The memory 664 stores information within the computing device 650. The memory 664 can be implemented as one or more of a computer-readable medium or media, a volatile memory unit or units, or a non-volatile memory unit or units.
Expansion memory 674 may also be provided and connected to device 650 through expansion interface 672, which may include, for example, a SIMM (Single In Line Memory Module) card interface. Such expansion memory 674 may provide extra storage space for device 650, or may also store applications or other information for device 650. Specifically, expansion memory 674 may include instructions to carry out or supplement the processes described above, and may include secure information also. Thus, for example, expansion memory 674 may be provide as a security module for device 650, and may be programmed with instructions that permit secure use of device 650. In addition, secure applications may be provided via the SIMM cards, along with additional information, such as placing identifying information on the SIMM card in a non-hackable manner.
The memory may include, for example, flash memory and/or NVRAM memory, as discussed below. In one implementation, a computer program product is tangibly embodied in an information carrier. The computer program product contains instructions that, when executed, perform one or more methods, such as those described above. The information carrier is a computer- or machine-readable medium, such as the memory 664, expansion memory 674, or memory on processor 652, that may be received, for example, over transceiver 668 or external interface 662.
Device 650 may communicate wirelessly through communication interface 666, which may include digital signal processing circuitry where necessary.
Communication interface 666 may provide for communications under various modes or protocols, such as GSM voice calls, SMS, EMS, or MMS messaging, CDMA, TDMA, PDC, WCDMA, CDMA2000, or GPRS, among others. Such communication may occur, for example, through radio-frequency transceiver 668. In addition, short- range communication may occur, such as using a Bluetooth, Wi-Fi, or other such transceiver (not shown). In addition, GPS (Global Positioning System) receiver module 670 may provide additional navigation- and location-related wireless data to device 650, which may be used as appropriate by applications running on device 650.
Device 650 may also communicate audibly using audio codec 660, which may receive spoken information from a user and convert it to usable digital information. Audio codec 660 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of device 650. Such sound may include sound from voice telephone calls, may include recorded sound (e.g., voice messages, music files, etc.) and may also include sound generated by applications operating on device 650.
The computing device 650 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as a cellular telephone 680. It may also be implemented as part of a smart phone 682, personal digital assistant, or other similar mobile device.
Various implementations of the systems and techniques described here can be realized in digital electronic circuitry, integrated circuitry, specially designed ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various implementations can include
implementation in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, coupled to receive data and instructions from, and to transmit data and instructions to, a storage system, at least one input device, and at least one output device.
These computer programs (also known as programs, software, software applications or code) include machine instructions for a programmable processor, and can be implemented in a high-level procedural and/or object-oriented programming language, and/or in assembly/machine language. As used herein, the terms "machine- readable medium" "computer-readable medium" refers to any computer program product, apparatus and/or device (e.g., magnetic discs, optical disks, memory, Programmable Logic Devices (PLDs)) used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal. The term "machine- readable signal" refers to any signal used to provide machine instructions and/or data to a programmable processor.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) (and other suitable display technology) for displaying information to the user and a keyboard and a pointing device (e.g., a mouse or a trackball) by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user can be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front end component (e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back end, middleware, or front end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), and the Internet. The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
The computing devices depicted in FIG. 1 interface with a virtual reality (VR headset 690). In an illustrated embodiment, one or more sensors are included in the VR headset to virtual remove the appearance of the headset 690 when one user interacts with another user. For instance, the headset sensors 690 may include one or more LED’s 691 that illuminates an eye of a user while tracking the user’s pupil and one or more micro cameras 693 for recording the user’s eye/pupil.
For instance, an embodiment of the present invention creates a model of the user’s face 704 without a headset 690 using a camera device 400 as a proxy for the user’s hidden face when a headset 690 is worn. This proxy is used to synthesize the face in a Mixed Reality (MR) video, thereby creating an impression of the headset 690 being removed when one using interacts with another user in a VR realm. For instance, this initial calibration step may require the user 704 to sit in front of a color+depth camera 400 and a monitor, and then track a marker on the monitor with their eyes. This calibration procedure acquires a 3D face model of the user 704, and create a database that maps appearance images (or textures) to different eye-gaze directions and blinks (via sensors 691 and 693). The aforementioned eye-gaze database (i.e. the face model preferably with textures indexed by eye-gaze) enables the appearance of a user’s face to be dynamically changed to generate desired (and real-time) eye-gaze, thus making the synthesized face look natural and alive. It is to be appreciated the construct of the face can also be altered by using suitable wire frame software technology.
In some implementations, one or more input devices in addition to the computing device (e.g., a mouse, a keyboard) can be rendered in a computer generated, 3D environment. The rendered input devices (e.g., the rendered mouse, the rendered keyboard) can be used as rendered in the VR space to control objects in the VR space.
Computing device 600 is intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. Computing device 650 is intended to represent various forms of mobile devices, such as personal digital assistants, cellular telephones, smart phones, and other similar computing devices. The components shown here, their connections and relationships, and their functions, are meant to be exemplary only, and are not meant to limit implementations of the inventions described and/or claimed in this document.
With the exemplary computing system environment 100 of FIG. 1 being generally shown and discussed above, and with reference now to FIGS. 2-7, described below is implementation of various exemplary embodiments.
It is to be appreciated an environment of use for an illustrated embodiment of the present invention relates to a system 600 that enables online chat in VR preferably with live video. In the example discussed below used to described an illustrated embodiment, it is to be understood it assumes a two person live chat with a local user and a remote user (which the present invention is not to be understood to be limited to). Exemplary embodiments include one or more components of the above described system 600 including a VR system such as Oculus Rift™, one or more webcams 400 per user and preferably a green screen. With reference to FIG. 2, during a VR chat session in accordance with an illustrated embodiment, a remote user 700 wearing headset 690 for example would view a live video of another user 702 (wearing headset 690), minus the headset 690 that the local user is wearing 704, in VR.
Likewise, during the same VR chat session, the remote user 702 wearing headset 690 would view a live video of the user 700 (wearing headset 690), minus the headset 690 that the local user is wearing 706, in VR.
In accordance with an aspect of the present invention (and with reference to FIG. 3), the headset visual for a remote user is replaced by an overlay of pre-recorded static images/video of the local users’ head. Preferably only a portion of the users head needs to be superimposed, as the mouth and lower portion of the face is still generally visible when wearing VR headsets. Preferably, the live video effect is created by chromakey (“green screen”) technology preferably using a standard 2D webcam 400 (Fig. 4). Thus, in accordance with an illustrated embodiment, each user 704, 706 is preferably placed in front of a portable green screen (402). It is to be appreciated this may be negated by certain software techniques or alternatively it may be possible to determine the users outline without a green screen by using a 3D camera.
As described below, in accordance with an illustrated embodiment, two or more users (704 and 706) may participate in a live online chat using a special webcam 400 at each user location. Each user (704, 706) preferably wears a VR headset 690 (700, 702) whereby each user (704, 706) views a superimposed and prerecorded image of the other user’s head (706, 704). The broadcast image is preferably a composite of a live video and prerecorded image. The live video preferably includes the body up to just above a user’s mouth region (FIG. 3) (e.g., where the VR headset 690 generally resides). Thus, the recorded image is preferably the nose, eyes and top of head of a user (704, 706). It is to be appreciated that it should be possible to achieve a live chat with more than two persons, such as a corporate teleconference.
This recorded image may also be created prior using a special form of photogrammetry. This is the technique where multiple 2D images are stitched together and processed to create a 3D model and texture. Currently this process is expensive, time consuming, and requires special hardware and multiple cameras. In contrast, in accordance with an embodiment of the present invention, implementation of inexpensive stereoscopic cameras may be utilized consisting of two cameras facing the user, each with either stereoscopic or monoscopic vision. For instance, the user would rotate in their chair while the cameras record and capture the scene. Post processing of the stereoscopic or monoscopic camera output would be used to generate the 3D model.
With reference now to FIGS. 4-7, an exemplary process for enabling a VR session in accordance with an illustrated embodiment is described. First the face of each user 704 (FIG. 4), 706 (FIG. 5) without a headset is photographed by webcam 400. As shown in FIGS 6 and 7, each user (704, 706) then wears a headset 690 having the aforementioned LED sensors 691 and 693 for detecting eye gaze, movement and blinking. Preferably the sensors 691 illuminates one eye of the user inside the headset 690 to track pupil movement while the micro video camera sensor 693 records a user’s pupil. The processing chip 602 is preferably configured and operational to restore the moving pupil to the face by“stitching” it in preferably via software to complete the image of the face currently obscured by the headset 690. Thus, both users 704 and 706 can view 3D images of each other (via their respective headset 690) in real time as if the face of the other user 704, 706 were in front of them and unobscured by a headset 690. It is to be further appreciated, that through the further use of electro- sensitive devices positioned preferably on the forehead and eyebrows of each user, the frowns expression of each user 704, 706 may also be tracked via software processing on the processing chip 602.
In summary, and in accordance with one or more illustrated embodiments, the present invention enables virtual projection of a webcam video stream of a user in a video chat, minus the users’ headset. This is accomplished by modeling the users head in 3D using modelling techniques, such as photogrammetry. For instance, the top portion of a users’ head is replaced by this model in real time, so the other participating user only views the body, head, face and eyes of a participating user, minus the headset 690.
It is to be appreciated multiple users (e.g., greater than two users) are to be encompassed wherein each user is modelled using one or more stationary stereoscopic 3D cameras (e.g., StereoLabs ZED camera, Microsoft Kinect, Intel RealSense line of 3D cameras, etc.). The 3D camera techniques additional may calculate a user’s position, providing VR cooperative interaction. For instance, two users could play a game of catch on Mars or two people could take a virtual walk together in the Grand Canyon.
It is to be appreciated the present invention computing system 600 and headset 690 may be configured and operative for use in an Augmented Reality (AR) setting. AR is to be understood to be a technology that layers computer-generated
enhancements atop an existing reality in order to make it more meaningful through the ability to interact with it. An illustrated embodiment includes AR being developed for use as an app preferably a the headset 690 to blend digital components into the real world in such a way that they enhance one another. For instance, two or more users, each in a different location, wearing the present invention headset 690 may use/operate the headsets 690 so as to appear next to one another in a chosen setting (e.g., skydiving) and/or location (e.g., Paris) whereby they not only in real-time visualize the same setting together (e.g., a street in Paris, France) but they view each other as if the headset 690 were removed, in accordance with the teachings herein.
It is to be further appreciated that additional embodiments of the computing system 600 and/or headset 690 may be configured to provide for real-time image stitching preferably using at least two separate video feeds and/or a single video feed preferably in conjunction with a predetermined still-frame Image Cache. Preferably, eye-tracking is performed via a camera mounted inside a VR display device enabling real-time selection of cached images and/or“blended” images to be derived from an initial“setup” phase (Image Cache) to be used in an image stitching step as the“eye region” of a user’ s face is typically occluded due to the VR display device thus providing real-time 3D face tracking and reconstruction using a final
(composited/stitched) video. It is to be appreciated that by employing a real-time digitally-created face (representing the User), the face/result is preferably composited onto either a real-time video and/or a depth-extraction of the user. It is to be appreciated that the user-body is preferably extracted from its background using either green screen technology and/or depth-extraction techniques. It is to be further appreciated a fully-reconstructed user may be composited into a multitude of applications such as (but not limited to): 360-degree video, VR Chat Rooms, and other applicable applications. It is to be further appreciated that that following miscellaneous technology (e.g., SDK’s, API’s, etc.) currently being used and/or explored include (but are not to be limited to): EmguCV, OpenCV and MATLAB; ZED Camera with SDK; ARKit; Primary Programming Languages: C# and C++; Unity (game engine); and Oculus and/or Vive (VR Display Device).
With certain illustrated embodiments described above, it is to be appreciated that various non-limiting embodiments described herein may be used separately, combined or selectively combined for specific applications. Further, some of the various features of the above non-limiting embodiments may be used without the corresponding use of other described features. The description provided herein should therefore be considered as merely illustrative of the principles, teachings and exemplary embodiments of this invention, and not in limitation thereof.

Claims

What is claimed is:
1. A for creating Virtual Reality (VR) interactions between two or more users wherein at least one user is wearing a VR headset, comprising: creating a two-dimensional (2D) image of at least one user wearing a VR headset using a camera device wherein positioned in the headset is one or more sensors; tracking the pupil movement of the user using the one or more sensors; remove an headset image from the created 2D image of the user wearing the headset; stitch in the detected and recorded pupil image of the user rendered by the one or more sensors so as to be superimposed where the headset image was removed from the created 2D image of the user such that another user can view a VR image of the user as if the user was not wearing the headset.
2. The computer method as recited in claim 1, wherein the 2D image is created using a video camera.
3. The computer method as recited in claim 1, wherein the camera device has a depth sensor.
4. The computer method as recited in claim 1, wherein the one or more sensors include infrared LED sensors.
5. The computer method as recited in claim 4, wherein the one or more sensors further include micro video camera for recording aspects of a user’ s pupil.
6. The method as recited in claim 1, wherein user’s may view each other in real time.
7. An apparatus for creating Virtual Reality (VR) interactions between two or more users wherein at least one user is wearing a VR headset, comprising:
at least one VR headset device having one or more integrated sensor components;
a camera device for at least capturing an image of a user wearing the VR
headset set device;
a VR imaging computer device communicatively coupled to each of the at least one VR headset device and camera device, the VR imaging computer device including:
a memory;
a processor disposed in communication with said memory, and
configured to issue a plurality of instructions stored in the memory’, wherein the instructions cause the processor to: create a two-dimensional (2D) image of at least one user
wearing a VR headset using a camera device wherein positioned in the headset is one or more sensors; track pupil movement of the user using the one or more sensors positioed in the VR headset; remove an headset image from the created 2D image of the user wearing the headset; stitch in the detected and recorded pupil image of the user rendered by the one or more sensors of the VR headset so as to be superimposed in a region in which the headset image was removed from the created 2D image of the user such that another user can view a VR image of the user as if the user was not wearing the headset.
8. The apparatus for creating Virtual Reality (VR) interactions between two or more users as recited in claim 9, wherein the camera device is a video camera.
9. The apparatus for creating Virtual Reality (VR) interactions between two or more users as recited in claim 9, wherein the camera device is intergrated with a smart phone device.
10. The apparatus for creating Virtual Reality (VR) interactions between two or more users as recited in claim 1, wherein the camera device has a depth sensor.
11. The apparatus for creating Virtual Reality (VR) interactions between two or more users as recited in claim 1, wherein the one or more sensor components intergrated in the VR headset include infrared LED sensors.
12. The apparatus for creating Virtual Reality (VR) interactions between two or more users as recited in claim 11, wherein the one or more sensor components intergrated in the VR headset include at least one micro video camera for recording aspects of a user’ s pupil.
13. The apparatus for creating Virtual Reality (VR) interactions between two or more users as recited in claim 7, wherein user’s may view each other in real-time.
14. The apparatus for creating Virtual Reality (VR) interactions between two or more users as recited in claim 7, wherein user’s may view against a predetermined backdrop image.
PCT/US2019/022121 2018-03-13 2019-03-13 Virtual reality system and method WO2019178276A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
CA3089885A CA3089885A1 (en) 2018-03-13 2019-03-13 Virtual reality system and method
CN201980018658.6A CN111936910A (en) 2018-03-13 2019-03-13 Virtual reality system and method
JP2020546898A JP7250809B2 (en) 2018-03-13 2019-03-13 Virtual reality system and method
CN202310667894.9A CN116700489A (en) 2018-03-13 2019-03-13 Virtual reality system and method
EP19766538.3A EP3729177A4 (en) 2018-03-13 2019-03-13 Virtual reality system and method
JP2023044888A JP2023078342A (en) 2018-03-13 2023-03-22 Virtual reality system and method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201862642488P 2018-03-13 2018-03-13
US62/642,488 2018-03-13

Publications (1)

Publication Number Publication Date
WO2019178276A1 true WO2019178276A1 (en) 2019-09-19

Family

ID=67908536

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2019/022121 WO2019178276A1 (en) 2018-03-13 2019-03-13 Virtual reality system and method

Country Status (5)

Country Link
EP (1) EP3729177A4 (en)
JP (2) JP7250809B2 (en)
CN (2) CN111936910A (en)
CA (1) CA3089885A1 (en)
WO (1) WO2019178276A1 (en)

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020080094A1 (en) 2000-12-22 2002-06-27 Frank Biocca Teleportal face-to-face system
US20120327196A1 (en) 2010-05-24 2012-12-27 Sony Computer Entertainment Inc. Image Processing Apparatus, Image Processing Method, and Image Communication System
US20150234455A1 (en) * 2013-05-30 2015-08-20 Oculus Vr, Llc Perception Based Predictive Tracking for Head Mounted Displays
US20160217621A1 (en) 2015-01-28 2016-07-28 Sony Computer Entertainment Europe Limited Image processing
US20160341959A1 (en) 2015-05-18 2016-11-24 Samsung Electronics Co., Ltd. Image processing for head mounted display devices
US20170178306A1 (en) 2015-12-21 2017-06-22 Thomson Licensing Method and device for synthesizing an image of a face partially occluded
US20170243334A1 (en) 2014-08-04 2017-08-24 Oculus Vr, Llc Method and System for Reconstructing Obstructed Face Portions for Virtual Reality Environment
US20170269685A1 (en) * 2016-03-17 2017-09-21 Sony Interactive Entertainment Inc. Spectating Virtual (VR) Environments Associated With VR User Interactivity
WO2017173457A1 (en) * 2016-04-01 2017-10-05 Incontext Solutions, Inc. Virtual reality platform for retail environment simulation
US20170287215A1 (en) * 2016-03-29 2017-10-05 Google Inc. Pass-through camera user interface elements for virtual reality
CN107305621A (en) 2016-04-17 2017-10-31 张翔宇 The image capture device and image synthesis system of a kind of virtual reality glasses
US20170326457A1 (en) * 2016-05-16 2017-11-16 Google Inc. Co-presence handling in virtual reality
WO2017212208A1 (en) * 2016-06-08 2017-12-14 Companion Limited System providing a shared environment
CN107491165A (en) 2016-06-12 2017-12-19 张翔宇 A kind of vr eyewear faces 3d images, plane picture capture and gesture capture systems
CN107680069A (en) 2017-08-30 2018-02-09 歌尔股份有限公司 A kind of image processing method, device and terminal device
US20180061133A1 (en) 2016-03-25 2018-03-01 Boe Technology Group Co., Ltd. Augmented reality apparatus and system, as well as image processing method and device

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3464754B2 (en) * 1997-09-19 2003-11-10 日本電信電話株式会社 Method and apparatus for synthesizing a face image of a person wearing a head-mounted display
US9265415B1 (en) * 2012-01-06 2016-02-23 Google Inc. Input detection
WO2015185537A1 (en) * 2014-06-03 2015-12-10 Thomson Licensing Method and device for reconstruction the face of a user wearing a head mounted display

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020080094A1 (en) 2000-12-22 2002-06-27 Frank Biocca Teleportal face-to-face system
US20120327196A1 (en) 2010-05-24 2012-12-27 Sony Computer Entertainment Inc. Image Processing Apparatus, Image Processing Method, and Image Communication System
US20150234455A1 (en) * 2013-05-30 2015-08-20 Oculus Vr, Llc Perception Based Predictive Tracking for Head Mounted Displays
US20170243334A1 (en) 2014-08-04 2017-08-24 Oculus Vr, Llc Method and System for Reconstructing Obstructed Face Portions for Virtual Reality Environment
US20160217621A1 (en) 2015-01-28 2016-07-28 Sony Computer Entertainment Europe Limited Image processing
US20160341959A1 (en) 2015-05-18 2016-11-24 Samsung Electronics Co., Ltd. Image processing for head mounted display devices
US20170178306A1 (en) 2015-12-21 2017-06-22 Thomson Licensing Method and device for synthesizing an image of a face partially occluded
US20170269685A1 (en) * 2016-03-17 2017-09-21 Sony Interactive Entertainment Inc. Spectating Virtual (VR) Environments Associated With VR User Interactivity
US20180061133A1 (en) 2016-03-25 2018-03-01 Boe Technology Group Co., Ltd. Augmented reality apparatus and system, as well as image processing method and device
US20170287215A1 (en) * 2016-03-29 2017-10-05 Google Inc. Pass-through camera user interface elements for virtual reality
WO2017173457A1 (en) * 2016-04-01 2017-10-05 Incontext Solutions, Inc. Virtual reality platform for retail environment simulation
CN107305621A (en) 2016-04-17 2017-10-31 张翔宇 The image capture device and image synthesis system of a kind of virtual reality glasses
US20170326457A1 (en) * 2016-05-16 2017-11-16 Google Inc. Co-presence handling in virtual reality
WO2017212208A1 (en) * 2016-06-08 2017-12-14 Companion Limited System providing a shared environment
CN107491165A (en) 2016-06-12 2017-12-19 张翔宇 A kind of vr eyewear faces 3d images, plane picture capture and gesture capture systems
CN107680069A (en) 2017-08-30 2018-02-09 歌尔股份有限公司 A kind of image processing method, device and terminal device

Also Published As

Publication number Publication date
EP3729177A1 (en) 2020-10-28
CN116700489A (en) 2023-09-05
JP2023078342A (en) 2023-06-06
JP7250809B2 (en) 2023-04-03
CN111936910A (en) 2020-11-13
JP2021517300A (en) 2021-07-15
EP3729177A4 (en) 2021-10-06
CA3089885A1 (en) 2019-09-19

Similar Documents

Publication Publication Date Title
US10527846B2 (en) Image processing for head mounted display devices
US11563779B2 (en) Multiuser asymmetric immersive teleconferencing
US12067686B2 (en) System and method for rendering virtual reality interactions
JP7504968B2 (en) Avatar display device, avatar generation device and program
US20240296636A1 (en) Systems and methods for generating and facilitating access to a personalized augmented rendering of a user
US10602121B2 (en) Method, system and apparatus for capture-based immersive telepresence in virtual environment
US8928659B2 (en) Telepresence systems with viewer perspective adjustment
KR20190112712A (en) Improved method and system for video conferencing with head mounted display (HMD)
KR20220047719A (en) Pew-Shot Composite for Talking Head
CN107274491A (en) A kind of spatial manipulation Virtual Realization method of three-dimensional scenic
US20230386147A1 (en) Systems and Methods for Providing Real-Time Composite Video from Multiple Source Devices Featuring Augmented Reality Elements
JP7250809B2 (en) Virtual reality system and method
KR102728245B1 (en) Avatar display device, avatar creation device and program
CN115079826B (en) Virtual reality implementation method, electronic equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19766538

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 3089885

Country of ref document: CA

ENP Entry into the national phase

Ref document number: 2019766538

Country of ref document: EP

Effective date: 20200724

ENP Entry into the national phase

Ref document number: 2020546898

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE