Abstract:
A system and method is provided for automatically recognizing building numbers in street level images. In one aspect, a processor selects a street level image that is likely to be near an address of interest. The processor identifies those portions of the image that are visually similar to street numbers, and then extracts the numeric values of the characters displayed in such portions. If an extracted value corresponds with the building number of the address of interest such as being substantially equal to the address of interest, the extracted value and the image portion are displayed to a human operator. The human operator confirms, by looking at the image portion, whether the image portion appears to be a building number that matches the extracted value. If so, the processor stores a value that associates that building number with the street level image.
Abstract:
A system and method is provided for automatically recognizing building numbers in street level images. In one aspect, a processor selects a street level image that is likely to be near an address of interest. The processor identifies those portions of the image that are visually similar to street numbers, and then extracts the numeric values of the characters displayed in such portions. If an extracted value corresponds with the building number of the address of interest such as being substantially equal to the address of interest, the extracted value and the image portion are displayed to a human operator. The human operator confirms, by looking at the image portion, whether the image portion appears to be a building number that matches the extracted value. If so, the processor stores a value that associates that building number with the street level image.
Abstract:
Examples of methods and systems for using eye gesture duration to provide calibration for eye gesture detection are described. In some examples, calibration can be executed using a head-mountable device. The head-mountable device may be configured to determine a duration range indicative of an eye gesture and receive a plurality of reference signals indicative of the eye gesture. The plurality of reference signals may comprise duration information indicative of a plurality of reference durations of the eye gesture. The head-mountable device may determine, based on the plurality of reference durations, a reference duration range associated with the eye gesture that is within the duration range, and adjust the duration range for the eye gesture based on the reference duration range.
Abstract:
A system and method is provided for automatically recognizing building numbers in street level images. In one aspect, a processor selects a street level image that is likely to be near an address of interest. The processor identifies those portions of the image that are visually similar to street numbers, and then extracts the numeric values of the characters displayed in such portions. If an extracted value corresponds with the building number of the address of interest such as being substantially equal to the address of interest, the extracted value and the image portion are displayed to a human operator. The human operator confirms, by looking at the image portion, whether the image portion appears to be a building number that matches the extracted value. If so, the processor stores a value that associates that building number with the street level image.
Abstract:
A computer implemented system for identifying license plates and faces in street-level images is disclosed. The system includes an object detector configured to determine a set of candidate objects in the image, a feature vector module configured to generate a set of feature vectors using the object detector to generate a feature vector for each candidate object in the set of candidate objects, a composite feature vector module to generate a set of composite feature vectors by combining each generated feature vector with a corresponding road or street description of the object in question, and an identifier module configured to identify objects of a particular type using a classifier that takes a set of composite feature vectors as input and returns a list of candidate objects that are classified as being of the particular type as output.
Abstract:
The present disclosure provides a computing device including an image-capture device and a control system. The control system may be configured to receive sensor data from one or more sensors, and analyze the sensor data to detect at least one image-capture signal. The control system may also be configured to cause the image-capture device to capture an image in response to detection of the at least one image-capture signal. The control system may also be configured to enable one or more speech commands relating to the image-capture device in response to capturing the image. The control system may also be configured to receive one or more verbal inputs corresponding to the one or more enabled speech commands. The control system may also be configured to perform an image-capture function corresponding to the one or more verbal inputs.
Abstract:
Disclosed are methods and devices for varying functionality of a wearable computing device. An example device includes a first sensor and a second sensor. An example method includes, while a device is operating in a first state, receiving an indication of a touch input at the first sensor. The second sensor is configured in an idle mode based on the device operating in the first state. The method further includes, in response to receiving the indication of the touch input, triggering the second sensor to operate in an active mode and receiving data from the second sensor. The method further includes determining, based on the data, whether the device is being worn.
Abstract:
This disclosure relates to winking to capture image data using an image capture device that is associated with a head-mountable device (HMD). An illustrative method includes detecting a wink gesture at an HMD. The method also includes causing an image capture device to capture image data, in response to detecting the wink gesture at the HMD.
Abstract:
This disclosure involves proximity sensing of eye gestures using a machine-learned model. An illustrative method comprises receiving training data that includes proximity-sensor data. The data is generated by at least one proximity sensor of a head-mountable device (HMD). The data is indicative of light received by the proximity sensor(s). The light is received by the proximity sensor(s) after a reflection of the light from an eye area. The reflection occurs while an eye gesture is being performed at the eye area. The light is generated by at least one light source of the HMD. The method further comprises applying a machine-learning process to the training data to generate at least one classifier for the eye gesture. The method further comprises generating an eye-gesture model that includes the at least one classifier for the eye gesture. The model is applicable to subsequent proximity-sensor data for detection of the eye gesture.
Abstract:
A system and method is provided for automatically recognizing building numbers in street level images. In one aspect, a processor selects a street level image that is likely to be near an address of interest. The processor identifies those portions of the image that are visually similar to street numbers, and then extracts the numeric values of the characters displayed in such portions. If an extracted value corresponds with the building number of the address of interest such as being substantially equal to the address of interest, the extracted value and the image portion are displayed to a human operator. The human operator confirms, by looking at the image portion, whether the image portion appears to be a building number that matches the extracted value. If so, the processor stores a value that associates that building number with the street level image.