Eyewear having unsynchronized rolling shutter cameras

    公开(公告)号:US12242065B2

    公开(公告)日:2025-03-04

    申请号:US18223686

    申请日:2023-07-19

    Applicant: Snap Inc.

    Abstract: Eyewear having unsynchronized rolling shutter (RS) cameras such that images produced by each camera are unaligned, and a state that includes velocity and gravity orientation in the eyewear's reference system is calculated. The state is not limited to these two parameters as other parameters such as the acceleration bias, the gyroscope bias, or both may be included. Math solvers are used such that processing time to calculate the velocity and gravity orientation are acceptable. Arranging the RS cameras in an unsynchronized configuration allows estimating the motion of the eyewear from just one stereo image pair and removes the requirement of possessing more images.

    Revolving XR eyewear display
    2.
    发明授权

    公开(公告)号:US12092826B2

    公开(公告)日:2024-09-17

    申请号:US18128905

    申请日:2023-03-30

    Applicant: Snap Inc.

    Abstract: An eXtended Reality (XR) display system includes a Light Emitting Diode (LED) display controller, and a Light Emitting Diode (LED) near-eye display element operatively coupled to the LED display driver. The LED near-eye display element includes one or more motors and an LED array operably connected to the one or more motors. During operation, the LED display driver receives video data including a rendered virtual object of an XR experience and generates LED array control signals based on the video data, the LED array control signals causing one or more LEDs of the LED array to be energized in a sequence. The LED display driver also generates synchronized motor control signals and simultaneously communicates the LED array control signals to the LED array and the synchronized motor control signals to the one or more motors causing the LED near-eye display element to display the rendered virtual object.

    ENERGY-EFFICIENT ADAPTIVE 3D SENSING
    3.
    发明公开

    公开(公告)号:US20240288696A1

    公开(公告)日:2024-08-29

    申请号:US18653808

    申请日:2024-05-02

    Applicant: Snap Inc.

    CPC classification number: G02B27/0172 G06F3/013 G06T19/006 G02B2027/0138

    Abstract: An energy-efficient adaptive 3D sensing system. The adaptive 3D sensing system includes one or more cameras and one or more projectors. The adaptive 3D sensing system captures images of a real-world scene using the one or more cameras and computes depth estimates and depth estimate confidence values for pixels of the images. The adaptive 3D sensing system computes an attention mask based on the one or more depth estimate confidence values and commands the one or more projectors to send a distributed laser beam into one or more areas of the real-world scene based on the attention mask. The adaptive 3D sensing system captures 3D sensing image data of the one or more areas of the real-world scene and generates 3D sensing data for the real-world scene based on the 3D sensing image data.

    3D space carving using hands for object capture

    公开(公告)号:US12260567B2

    公开(公告)日:2025-03-25

    申请号:US17973167

    申请日:2022-10-25

    Applicant: Snap Inc.

    Abstract: A method for carving a 3D space using hands tracking is described. In one aspect, a method includes accessing a first frame from a camera of a display device, tracking, using a hand tracking algorithm operating at the display device, hand pixels corresponding to one or more user hands depicted in the first frame, detecting, using a sensor of the display device, depths of the hand pixels, identifying a 3D region based on the depths of the hand pixels, and applying a 3D reconstruction engine to the 3D region.

    Device pairing using machine-readable optical label

    公开(公告)号:US12243266B2

    公开(公告)日:2025-03-04

    申请号:US18124099

    申请日:2023-03-21

    Applicant: Snap Inc.

    Abstract: A method for aligning coordinate systems from separate augmented reality (AR) devices is described. In one aspect, a first device accesses first pose data from a first Visual Inertial Odometry (VIO) system of the first device. A camera of the first device captures an image of a machine-readable code that is displayed on a display of a second device. The second device encodes the machine-readable code with second pose data from a second VIO system of the second device. The first device decodes the second pose data from the machine-readable code, and determines a relative pose between the first device and the second device based on the first pose data and the second pose data.

    Augmented reality guided depth estimation

    公开(公告)号:US12229977B2

    公开(公告)日:2025-02-18

    申请号:US17529527

    申请日:2021-11-18

    Applicant: Snap Inc.

    Abstract: A method for AR-guided depth estimation is described. The method includes identifying a virtual object rendered in a first frame that is generated based on a first pose of an augmented reality (AR) device, determining a second pose of the AR device, the second pose following the first pose, identifying an augmentation area in the second frame based on the virtual object rendered in the first frame, and the second pose, determining depth information for the augmentation area in the second frame, and rendering the virtual object in the second frame based on the depth information.

    DEVICE PAIRING USING MACHINE-READABLE OPTICAL LABEL

    公开(公告)号:US20240221212A1

    公开(公告)日:2024-07-04

    申请号:US18124099

    申请日:2023-03-21

    Applicant: Snap Inc.

    Abstract: A method for aligning coordinate systems from separate augmented reality (AR) devices is described. In one aspect, a first device accesses first pose data from a first Visual Inertial Odometry (VIO) system of the first device. A camera of the first device captures an image of a machine-readable code that is displayed on a display of a second device. The second device encodes the machine-readable code with second pose data from a second VIO system of the second device. The first device decodes the second pose data from the machine-readable code, and determines a relative pose between the first device and the second device based on the first pose data and the second pose data.

    ENERGY-EFFICIENT ADAPTIVE 3D SENSING
    10.
    发明公开

    公开(公告)号:US20240126084A1

    公开(公告)日:2024-04-18

    申请号:US18299923

    申请日:2023-04-13

    Applicant: Snap Inc.

    CPC classification number: G02B27/0172 G06F3/013 G06T19/006 G02B2027/0138

    Abstract: An energy-efficient adaptive 3D sensing system. The adaptive 3D sensing system includes one or more cameras and one or more projectors. The adaptive 3D sensing system captures images of a real-world scene using the one or more cameras and computes depth estimates and depth estimate confidence values for pixels of the images. The adaptive 3D sensing system computes an attention mask based on the one or more depth estimate confidence values and commands the one or more projectors to send a distributed laser beam into one or more areas of the real-world scene based on the attention mask. The adaptive 3D sensing system captures 3D sensing image data of the one or more areas of the real-world scene and generates 3D sensing data for the real-world scene based on the 3D sensing image data.

Patent Agency Ranking