Abstract:
A method, performed in an electronic device, for tracking a piece of music in an audio stream is disclosed. The method may receive a first portion of the audio stream and extract a first sound feature based on the first portion of the audio stream. Also, the method may determine whether the first portion of the audio stream is indicative of music based on the first sound feature. In response to determining that the first portion of the audio stream is indicative of music, a piece of music may be identified based on the first portion of the audio stream. Further, upon receiving a second portion of the audio stream, the method may extract a second sound feature based on the second portion of the audio stream and determine whether the second portion of the audio stream is indicative of the first piece of music.
Abstract:
A method, which is performed in an electronic device, for activating a target application is disclosed. The method may include receiving an input sound stream including an activation keyword for activating the target application and a speech command indicative of a function of the target application. The method may also detect the activation keyword from the input sound stream. If the activation keyword is detected, a portion of the input sound stream including at least a portion of the speech command may be buffered in a buffer memory. In addition, in response to detecting the activation keyword, the target application may be activated to perform the function of the target application.
Abstract:
A method of detecting a target keyword for activating a function in an electronic device is disclosed. The method includes receiving an input sound starting from one of the plurality of portions of the target keyword. The input sound may be periodically received based on a duty cycle. The method extracts a plurality of sound features from the input sound, and obtains state information on a plurality of states associated with the portions of the target keyword. Based on the extracted sound features and the state information, the input sound may be detected as the target keyword. The plurality of states includes a predetermined number of entry states indicative of a predetermined number of the plurality of portions.
Abstract:
A method, which is performed by an electronic device, for obtaining a speaker-independent keyword model of a keyword designated by a user is disclosed. The method may include receiving at least one sample sound from the user indicative of the keyword. The method may also generate a speaker-dependent keyword model for the keyword based on the at least one sample sound, send a request for the speaker-independent keyword model of the keyword to a server in response to generating the speaker-dependent keyword model, and receive the speaker-independent keyword model adapted for detecting the keyword spoken by a plurality of users from the server.
Abstract:
A method, performed in an electronic device, for assigning a target keyword to a function is disclosed. In this method, a list of a plurality of target keywords is received at the electronic device via a communication network, and a particular target keyword is selected from the list of target keywords. Further, the method may include receiving a keyword model for the particular target keyword via the communication network. In this method, the particular target keyword is assigned to a function of the electronic device such that the function is performed in response to detecting the particular target keyword based on the keyword model in an input sound received at the electronic device.
Abstract:
A method of detecting a target keyword from an input sound for activating a function in a mobile device is disclosed. In this method sound features are extracted from the input stream and statistics are generated (340) including the mean value and the variance of a particular sound feature. Based on the statistics the method adaptively skips the normalisation of a sound feature when the difference between the present value and the previous value is not significant, this has the effect of lowering the process load. In detail a first plurality of sound features is received in a buffer (330), and a second plurality of sound features is received in the buffer 330). While receiving each of the second plurality of sound features in the buffer, a first number of the sound features are processed from the buffer. The first number of the sound features includes two or more sound features. Further, the method includes determining a keyword score (360) for each of the processed sound features and detecting the input sound as the target keyword (370) if at least one of the keyword scores is greater than a threshold score.
Abstract:
A method for communicating messages by a mobile device via a sound medium is disclosed. The mobile device receives input sounds from at least one mobile device via the sound medium. From the input sounds, an input sound signal carrying a first message encoded with a first key is detected. The mobile device decodes the first message based on a matching key. An output sound signal carrying a second message encoded with a second key is generated. Further, the mobile device transmits an output sound corresponding to the output sound signal via the sound medium.
Abstract:
A method, which is performed in an electronic device, for activating a target application is disclosed. The method may include receiving an input sound stream including an activation keyword for activating the target application and a speech command indicative of a function of the target application. The method may also detect the activation keyword from the input sound stream. If the activation keyword is detected, a portion of the input sound stream including at least a portion of the speech command may be buffered in a buffer memory. In addition, in response to detecting the activation keyword, the target application may be activated to perform the function of the target application.
Abstract:
According to an aspect of the present disclosure, a method for controlling display of a region on a touch screen display of a mobile device is disclosed. The method includes receiving a command indicative of zooming by a first sensor, sensing at least one image including at least one eye by a camera, determining a direction of a gaze of the at least one eye based on the at least one image, determining a target region to be zoomed on the touch screen display based on the direction of the gaze, and zooming the target region on the touch screen display.
Abstract:
A method for activating a voice assistant function in a mobile device is disclosed. The method includes receiving an input sound stream by a sound sensor and determining a context of the mobile device. The method may determine the context based on the input sound stream. For determining the context, the method may also obtain data indicative of the context of the mobile device from at least one of an acceleration sensor, a location sensor, an illumination sensor, a proximity sensor, a clock unit, and a calendar unit in the mobile device. In this method, a threshold for activating the voice assistant function is adjusted based on the context. The method detects a target keyword from the input sound stream based on the adjusted threshold. If the target keyword is detected, the method activates the voice assistant function.