Abstract:
Un procedimiento (1100) para realizar una operación de procesamiento de imágenes usando múltiples tramas de material de vídeo pregrabado o en directo, la operación de procesamiento de imágenes comprendiendo capturar una fotografía (149) del material de vídeo o realizar una operación de edición de vídeo en una trama del material de vídeo, estando el procedimiento caracterizado por: obtener (1102, 1104) una primera área de rastreo (1233) y una segunda área de rastreo (1235) dentro de una trama; rastrear (1106, 1108) la primera área de rastreo y la segunda área de rastreo a través de múltiples tramas; y realizar (1110) la operación de procesamiento de imágenes una vez que una parte (1243) de la primera área de rastreo que se superpone a la segunda área de rastreo pasa un umbral, donde la operación de procesamiento de imágenes no se realiza hasta que se pasa el umbral.
Abstract:
Generally stated a method and an accompanying apparatus provides for a voice recognition system ( 300 ) with programmable front end processing unit ( 400 ). The front end processing unit ( 400 ) requests and receives different configuration files at different times for processing voice data in the voice recognition system ( 300 ). The configuration files are communicated to the front end unit via a communication link ( 310 ) for configuring the front end processing unit ( 400 ). A microprocessor may provide the front end configuration files on the communication link at different times.
Abstract:
An electronic device is described. The electronic device includes a processor. The processor is configured to obtain a plurality of images. The processor is also configured to obtain global motion information indicating global motion between at least two of the plurality of images. The processor is further configured to obtain object tracking information indicating motion of a tracked object between the at least two of the plurality of images. The processor is additionally configured to perform automatic zoom based on the global motion information and the object tracking information. Performing automatic zoom produces a zoom region including the tracked object. The processor is configured to determine a motion response speed for the zoom region based on a location of the tracked object within the zoom region.
Abstract:
A method of processing data includes receiving, at a computing device, data representative of an image captured by an image sensor. The method also includes determining a first scene clarity score. The method further includes determining whether the first scene clarity score satisfies a threshold, and if the first scene clarity score satisfies the threshold, determining a second scene clarity score based on second data extracted from the data.
Abstract:
A method for picture processing is described. A first tracking area is obtained. A second tracking area is also obtained. The method includes beginning to track the first tracking area and the second tracking area. Picture processing is performed once a portion of the first tracking area overlapping the second tracking area passes a threshold.
Abstract:
This disclosure describes adaptive filtering techniques to improve the quality of captured image information, such as video or still images. An image sensor captures image information and determines a plurality of parameter values based on a current exposure index and a current scaling factor of the image information. The adaptive spatial image filter includes both horizontal and vertical sharpening filters and configures, i.e., adapts, the horizontal sharpening filter and the vertical sharpening filter based on the plurality of parameter values determined by the image sensor. The adaptive spatial image filter applies the horizontal and vertical sharpening filters to at least one channel of the image information to generate filtered image information.
Abstract:
A two-dimensional (2D) mesh is applied over a distortion surface to approximate a lens roll-off distortion pattern. The process to apply the 2D mesh distributes a plurality of grid points among the distortion pattern and sub-samples the distortion pattern to derive corrected digital gains at each grid location. Non-grid pixels underlying grid blocks having a grid point at each corner are adjusted based on the approximation of the lens roll-off for the grid points of the grid block. In one example, bilinear interpolation is used. The techniques universally correct lens roll-off distortion irregardless of the distortion pattern shape or type. The technique may also correct for green channel imbalance.
Abstract:
Techniques for improving the quality of images are described. A first histogram of intensity values may be obtained for an input image and diffused to obtain a second histogram with better intensity coverage. The diffusion may be achieved by filtering the first histogram for multiple iterations with a diffusion function obtained based on a filter function and a diffusion control function. The filter function may control the rate and/or characteristics of the diffusion. The diffusion control function may control shifts in positions of lobes in the first histogram. A transformation function may be determined based on a first cumulative distribution function (CDF) for the first histogram and an inverse function for a second CDF for the second histogram. An output image may be generated by mapping each pixel value in the input image to a corresponding pixel value in the output image based on the transformation function.
Abstract:
A method performed by an electronic device is described. The method includes generating a plurality of bounding regions based on an image. The method also includes determining a subset of the plurality of bounding regions based on at least one criterion and a selected area in the image. The method further includes processing the image based on the subset of the plurality of bounding regions.
Abstract:
A method of generating a temporal saliency map is disclosed. In a particular embodiment, the method includes receiving an object bounding box from an object tracker. The method includes cropping a video frame based at least in part on the object bounding box to generate a cropped image. The method further includes performing spatial dual segmentation on the cropped image to generate an initial mask and performing temporal mask refinement on the initial mask to generate a refined mask. The method also includes generating a temporal saliency map based at least in part on the refined mask.