Abstract:
PROBLEM TO BE SOLVED: To solve a problem that individual or thumbnail display of digital photographs gives a viewer a very limited sense of what the person who has took the photographs has experienced when the images have been captured. SOLUTION: A method includes steps of: receiving a first image captured by an image capture device and first meta data related thereto, including geographical information for contents of the first image; receiving a second image and second metadata related thereto, including geographical information for contents of the second image; determining a viewpoint of the first image which represents the location and orientation of the image capture device when the first image has been captured; and creating a view including the first image and the second image. The placement of the first image in the view is based on the first metadata and the viewpoint of the first image, and the placement of the second image relative to the first image is based on the second metadata and the viewpoint of the first image. COPYRIGHT: (C)2011,JPO&INPIT
Abstract:
In some embodiments, a method of processing a video sequence may include receiving an input video sequence having an input video sequence resolution, aligning images from the input video sequence, reducing noise in the aligned images, and producing an output video sequence from the reduced noise images, wherein the output video sequence has the same resolution as the input video sequence resolution. Other embodiments are disclosed and claimed.
Abstract:
Briefly, in accordance with one or more embodiments, an image processing system is capable of receiving an image containing text, applying optical character recognition to the image, and then audibly reproducing the text via text-to-speech synthesis. Prior to optical character recognition, an orientation corrector is capable of detecting an amount of angular rotation of the text in the image with respect to horizontal, and then rotating the image by an appropriate amount to sufficiently align the text with respect to horizontal for optimal optical character recognition. The detection may be performed using steerable filters to provide an energy versus orientation curve of the image data. A maximum of the energy curve may indicate the amount of angular rotation that may be corrected by the orientation corrector.
Abstract:
Methods and systems to create an image in which objects at different focal depths all appear to be in focus. In an embodiment, all objects in the scene may appear in focus. Non-stationary cameras may be accommodated, so that variations in the scene resulting from camera jitter or other camera motion may be tolerated. An image alignment process may be used, and the aligned images may be blended using a process that may be implemented using logic that has relatively limited performance capability. The blending process may take a set of aligned input images and convert each image into a simplified Laplacian pyramid (LP). The LP is a data structure that includes several processed versions of the image, each version being of a different size. The set of aligned images is therefore converted into a corresponding set of LPs. The LPs may be combined into a composite LP, which may then undergo Laplacian pyramid reconstruction (LPR). The output of the LPR process is the final blended image.
Abstract:
In some embodiments, provided are procedures for processing images that may have different font sizes. In some embodiments, it involves OCR'ing with multiple passes at different resolutions.
Abstract:
Methods and systems to for generating high accuracy estimates of the 3D orientation of a camera within a global frame of reference. Orientation estimates may be produced from an image-based alignment method. Other orientation estimates may be taken from a camera-mounted orientation sensor. The alignment-derived estimates may be input to a high pass filter. The orientation estimates from the orientation sensor may be processed and input to a low pass filter. The outputs of the high pass and low pass filters are fused, producing a stabilized video sequence.
Abstract:
In some embodiments, a method of processing a video sequence may include receiving an input video sequence having an input video sequence resolution, aligning images from the input video sequence, reducing noise in the aligned images, and producing an output video sequence from the reduced noise images, wherein the output video sequence has the same resolution as the input video sequence resolution. Other embodiments are disclosed and claimed.
Abstract:
In some embodiments, a method of processing a video sequence may include receiving an input video sequence having an input video sequence resolution, aligning images from the input video sequence, reducing noise in the aligned images, and producing an output video sequence from the reduced noise images, wherein the output video sequence has the same resolution as the input video sequence resolution. Other embodiments are disclosed and claimed.
Abstract:
Systeme und Verfahren zum Bestimmen von Punkt-zu-Punkt-Abständen aus 3D-Bilddaten. Bei manchen Ausführungsformen repräsentieren zwei Messpunkte, die zum Beispiel durch einen Benutzer spezifiziert werden, Endpunkte auf einem Objekt von Interesse innerhalb eines Einzelbildes. Unter der Annahme, dass alle Punkte, die zwischen diesen Endpunkten liegen, auch zum Objekt von Interesse gehören, können zusätzliche 3D-Daten, die mit Punkten assoziiert sind, die entlang einer durch die Messpunkte definierten Messlinie liegen, wirksam eingesetzt werden, um eine robuste Abstandsmessung bereitzustellen. Bei manchen Ausführungsformen wird eine Total-Least-Squares-Anpassung durchgeführt, zum Beispiel über RPCA (Robust Principal Component Analysis), um lineare Strukturen innerhalb des Satzes der 3D-Koordinaten auf der Messlinie zu identifizieren. Bei manchen Ausführungsbeispielen wird die MCD(Minimum Covariance Determinant)-Schätzfunktion der Kovarianzmatrix der Daten für eine hoch robuste Schätzung des Mehrvariablenorts und der Mehrvariablenstreuung berechnet.