Abstract:
A device is configured to receive a translation query that requests a translation of terms from a source language to a target language; determine translation features associated with the translation query; assign a feature value to each of the translation features to form feature values; apply a feature weight to each of the feature values to generate a final value; and determine whether to provide a dialog translation user interface or a non-dialog translation user interface based on whether the final value satisfies a threshold. The dialog translation user interface may facilitate translation of a conversation, the non-dialog translation user interface may provide translation search results, and the non-dialog translation user interface may be different than the dialog translation user interface. The device also configured to provide the dialog translation user interface for display when the final value satisfies the threshold
Abstract:
Methods, systems, and apparatus, including computer programs encoded on computer storage media, are provided for using audio features to classify audio for information retrieval. In general, one aspect of the subject matter described in this specification can be embodied in methods that include the actions of generating a collection of auditory images, each auditory image being generated from respective audio files according to an auditory model; extracting sparse features from each auditory image in the collection to generate a sparse feature vector representing the corresponding audio file; and ranking the audio files in response to a query including one or more words using the sparse feature vectors and a matching function relating sparse feature vectors to words in the query.
Abstract:
Techniques for summarizing media are described. A viewer-interaction analyzer receives a media file containing media, the media file including a plurality of segments. A segment of the media file is scored based on interactions of a set of raters. Viewer metrics on the segment of the media file are measured based on interactions with the segment of the media file by a set of viewers. A set of feature vectors are formed based on the measured viewer interactions, where feature vectors in the set of feature vectors are based on interactions of the set of viewers. A model is trained based on the set of feature vectors and the score assigned to the segment of the media file.
Abstract:
A system, computer readable storage medium, and computer-implemented method presents video search results responsive to a user keyword query. The video hosting system uses a machine learning process to learn a feature-keyword model associating features of media content from a labeled training dataset with keywords descriptive of their content. The system uses the learned model to provide video search results relevant to a keyword query based on features found in the videos. Furthermore, the system determines and presents one or more thumbnail images representative of the video using the learned model.
Abstract:
Methods and systems for providing rich content with an answer to a question query. A method includes receiving a query determined to be a question query and a corresponding answer generated in response to the question query, generating a contextual query that includes an element relating to the question query and an element relating to the answer; submitting the contextual query to a rich content search process and receiving data specifying a first set of rich content items responsive to the contextual query, determining first rich content item in the first set of rich content items that meet a context condition that is indicative of a rich content item providing contextual information of both elements of the question query and the answer query; and preferentially selecting from the first content items relative to the second rich content items to be provided as one or more answer rich content items.
Abstract:
A system, computer readable storage medium, and computer-implemented method presents video search results responsive to a user keyword query. The video hosting system uses a machine learning process to learn a feature-keyword model associating features of media content from a labeled training dataset with keywords descriptive of their content. The system uses the learned model to provide video search results relevant to a keyword query based on features found in the videos. Furthermore, the system determines and presents one or more thumbnail images representative of the video using the learned model.
Abstract:
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for identifying a query for selected content. In one aspect, a method includes receiving gesture data specifying a user gesture interacting with a portion of displayed content. A subset of the content is identified based on the gesture data. A set of candidate search queries is identified based on the subset of the content. A likelihood score is determined for each candidate search query. The likelihood score for a candidate search query indicates a likelihood that the candidate search query is an intended search query specified by the user gesture. The likelihood score for each candidate search query is adjusted using a normalization factor. The normalization factor can be based on a number of characters included in the candidate search query. One or more of the candidate search queries are selected based on the adjusted likelihood scores.
Abstract:
Systems and methods are described herein for comparing products in a marketplace. An image or video of the products may be captured using a camera associated with a mobile device. User input may be received to select two or more products within the image. Machine vision techniques may be applied to specifically identify the selected products. Product features associated with each of the identified products may be retrieved and formatted into a comparison of product features. The comparison may be presented to the user.
Abstract:
Methods, systems, and apparatus, including computer programs encoded on a computer storage medium, for identifying a query for selected content. In one aspect, a method includes receiving gesture data specifying a user gesture interacting with a portion of displayed content. A subset of the content is identified based on the gesture data. A set of candidate search queries is identified based on the subset of the content. A likelihood score is determined for each candidate search query. The likelihood score for a candidate search query indicates a likelihood that the candidate search query is an intended search query specified by the user gesture. The likelihood score for each candidate search query is adjusted using a normalization factor. The normalization factor can be based on a number of characters included in the candidate search query. One or more of the candidate search queries are selected based on the adjusted likelihood scores.