Abstract:
Method, apparatus, and computer program product for merging multiple maps for computer vision based tracking are disclosed. In one embodiment, a method of merging multiple maps for computer vision based tracking comprises receiving a plurality of maps of a scene in a venue from at least one mobile device, identifying multiple keyframes of the plurality of maps of the scene, and merging the multiple keyframes to generate a global map of the scene.
Abstract:
An apparatus and method for generating parameters for an application, such as an augmented reality application (AR app), using camera pose and gyroscope rotation is disclosed. The parameters are estimated based on pose from images and rotation from a gyroscope (e.g., using least-squares estimation with QR factorization or a Kalman filter). The parameters indicate rotation, scale and/or non-orthogonality parameters and optionally gyroscope bias errors. In addition, the scale and non-orthogonality parameters may be used for conditioning raw gyroscope measurements to compensate for scale and non-orthogonality.
Abstract:
Disclosed are a system, apparatus, and method for depth and color camera image synchronization. Depth and color camera input images are received or otherwise obtained unsynchronized and without associated creation timestamps. An image of one type is compared with an image of a different type to determine a match for synchronization. Matches may be determined according to edge detection or depth coordinate detection. When a match is determined a synchronized pair is formed for processing within an augmented reality output. Optionally the synchronized pair may be transformed to improve the match between the image pair.
Abstract:
Systems and methods are provided for alternating projection of content and capturing an image. The method includes steps of projecting, by a projection device, content at a first rate, projecting, by the projection device, a capture frame at a second rate, and capturing, by an image capture device, an image including the capture frame at the second rate, wherein capturing the image comprises capturing the image when the capture frame is projected. Systems and methods provided herein may provide increased tracking accuracy by a projecting a capture frame that does not obscure features in the image.
Abstract:
A master device images an object device and uses the image to identify the object device. The master device then automatically interfaces with the identified object device, for example, by pairing with the object device. The master device may receive broadcast data from the object device including information about the visual appearance of the object device and use the broadcast data in the identification of the object device. The master device may retrieve data related to the object device and display the related data, which may be display the data over the displayed image of the object device. The master device may provide an interface to control the object device or be used to pass data to the object device.
Abstract:
In one embodiment, the present invention comprises a vocoder having at least one input and at least one output, an encoder comprising a filter having at least one input operably connected to the input of the vocoder and at least one output, a decoder comprising a synthesizer having at least one input operably connected to the at least one output of the encoder, and at least one output operably connected to the at least one output of the vocoder, wherein the encoder comprises a memory and the encoder is adapted to execute instructions stored in the memory comprising classifying speech segments and encoding speech segments, and the decoder comprises a memory and the decoder is adapted to execute instructions stored in the memory comprising time-warping a residual speech signal to an expanded or compressed version of the residual speech signal.
Abstract:
An example system includes a first computing device comprising a first graphics processing unit (GPU) implemented in circuitry, and a second computing device comprising a second GPU implemented in circuitry. The first GPU is configured to perform a first portion of an image rendering process to generate intermediate graphics data and send the intermediate graphics data to the second computing device. The second GPU is configured to perform a second portion of the image rendering process to render an image from the intermediate graphics data. The first computing device may be a video game console, and the second computing device may be a virtual reality (VR) headset that warps the rendered image to produce a stereoscopic image pair.
Abstract:
A method of calculating a most likely map based on batch data includes gathering a corpus of sensor measurements indexed by a location of a sensor throughout an environment to be mapped. The method also includes determining, after gathering the corpus of sensor measurements, a most likely occupancy level of each voxel of multiple voxels of the environment in accordance with the corpus of sensor measurements and a stochastic sensor model. The method further includes calculating the most likely map based on the determined most likely occupancy level.