Abstract:
A method and apparatus for image stabilization takes an input image sequence including a plurality of frames (23), estimates frame-level motion vectors for each frame, and adaptively integrates the motion vectors to produce, for each frame, a motion vector to be used for image stabilization (7). A copy of the reference image of a frame is displaced by the corresponding adaptively integrated motion vector (21). In one embodiment, the perimeter of the image sensor is padded with margin to be used for image compensation. In another embodiment, vertical and horizontal components are treated independently (7). In still another embodiment, the motion estimation circuitry associated with an MPEG-4 encoder is used to calculate macroblock level vectors (11), and a histogram is used co compute a corresponding frame-level vector for that frame (7).
Abstract:
In various implementations, object tracking in a video content analysis system can be augmented with an image-based object re-identification system (e.g., for person re- identification or re-identification of other objects) to improve object tracking results for objects moving in a scene. The object re-identification system can use image recognition principles, which can be enhanced by considering data provided by object trackers that can be output by an object traffic system. In a testing stage, the object re-identification system can selectively test object trackers against object models. For most input video frames, not all object trackers need be tested against all object models. Additionally, different types of object trackers can be tested differently, so that a context provided by each object tracker can be considered. In a training stage, object models can also be selectively updated.
Abstract:
An apparatus includes a processor and a memory. The processor is configured to execute instructions stored at the memory to receive first characterization data and second characterization data. The first characterization data includes first values in a first order and corresponding to a first object. The second characterization data includes second values in a second order and corresponding to a second object. The processor is further configured to generate third characterization data and to generate fourth characterization. The third characterization data includes the first values in a third order. The fourth characterization data includes the second values in a fourth order. The processor is also configured to perform a first similarity operation using the first, second, third, and fourth characterization data to generate first result data and to determine whether the first object and the second object match based on the first result data.
Abstract:
An electronic device is described. The electronic device includes a processor. The processor is configured to obtain a plurality of images. The processor is also configured to obtain global motion information indicating global motion between at least two of the plurality of images. The processor is further configured to obtain object tracking information indicating motion of a tracked object between the at least two of the plurality of images. The processor is additionally configured to perform automatic zoom based on the global motion information and the object tracking information. Performing automatic zoom produces a zoom region including the tracked object. The processor is configured to determine a motion response speed for the zoom region based on a location of the tracked object within the zoom region.
Abstract:
A method for object classification by an electronic device is described. The method includes obtaining an image frame that includes an object. The method also includes determining samples from the image frame. Each of the samples represents a multidimensional feature vector. The method further includes adding the samples to a training set for the image frame. The method additionally includes pruning one or more samples from the training set to produce a pruned training set. One or more non-support vector negative samples are pruned first. One or more non-support vector positive samples are pruned second if necessary to avoid exceeding a sample number threshold. One or more support vector samples are pruned third if necessary to avoid exceeding the sample number threshold. The method also includes updating classifier model weights based on the pruned training set.
Abstract:
A three dimensional (3D) mixed reality system combines a real 3D image or video, captured by a 3D camera for example, with a virtual 3D image rendered by a computer or other machine to render a 3D mixed-reality image or video. A 3D camera can acquire two separate images (a left and a right) of a common scene, and superimpose the two separate images to create a real image with a 3D depth effect. The 3D mixed-reality system can determine a distance to a zero disparity plane for the real 3D image, determine one or more parameters for a projection matrix based on the distance to the zero disparity plane, render a virtual 3D object based on the projection matrix, combine the real image and the virtual 3D object to generate a mixed-reality 3D image.
Abstract:
Techniques for performing mesh-based video compression/decompression with domain transformation are described. A video encoder partitions an image into meshes of pixels, processes the meshes of pixels to obtain blocks of prediction errors, and codes the blocks of prediction errors to generate coded data for the image. The meshes may have arbitrary polygonal shapes and the blocks may have a predetermined shape, e.g., square. The video encoder may process the meshes of pixels to obtain meshes of prediction errors and may then transform the meshes of prediction errors to the blocks of prediction errors. Alternatively, the video encoder may transform the meshes of pixels to blocks of pixels and may then process the blocks of pixels to obtain the blocks of prediction errors. The video encoder may also perform mesh-based motion estimation to determine reference meshes used to generate the prediction errors.
Abstract:
A method (500) for obtaining structural information from a digital image by an electronic device is described. The method includes determining (510) an iris position in a region of interest based on a gradient direction transform (504, 506, 508). Determining the iris position may include determining a first dimension position and a second dimension position corresponding to a maximum value in the transform space.