Abstract:
A phoneme series-utterance sequence conversion method and an apparatus thereof in which the morpheme for the Korean text-to-speech system is provided to a phoneme series-utterance sequence transform system by using a sampling method. A training part(110) comprises phoneme series arranged with a phoneme series-utterance sequence sorter(111) and a translation rule extracting unit(112) extracting regulation from the utterance sequence. The phoneme series-utterance sequence sorter arranges the phoneme series and utterance sequence of the language corpus. The translation rule extracting unit extracts regulation from the phoneme series and the arranged utterance sequence. A transform unit(120) comprises a rule application part(121). A conversion rule tree(130) uses the enunciation generating of the transform unit. When the phoneme series and the given utterance sequence have the phoneme series-utterance sequence sorter, it determines whether any kind of phoneme is corresponded to any kind of utterance.
Abstract:
An automatic speech interpretation system based on a statistical automatic interpretation mode, and interpretation processing and training methods thereof are provided to increase performance by performing pre treatment, post treatment, and rearrangement when a statistical automatic interpreter is applied. A statistical automatic interpreter(40) interprets a target language in an interpreted language by receiving a target language model(s6) and an interpretation model(s4). A pre-interpretation processing module(02) determines the final voice recognition and interpretation sentences by transferring a plurality of candidate voice recognition sentences to the statistical automatic interpreter without determining the voice recognition sentence after execution of a voice recognizer, and considering voice recognition and interpretation scores. A post-interpretation module(03) outputs the final interpretation result by arranging the candidate voice recognition sentences in order of a sum of the voice recognition and interpretation scores.
Abstract:
A method and an apparatus for comprehending spoken words by using an information extraction method are provided to comprehend essential elements selectively in comprehending the spoken words on the basis of a meaning structure suitable for each specific domain, thereby improving a degree of comprehension for the spoken words. A method for comprehending spoken words by using an information extraction method comprises the following steps of: standardizing the meaning structure of the spoken words previously(210); embodying the standardized meaning structure to be suitable for a specific domain(220); inputting spoken words recognized through a voice recognition unit(230); performing the natural language processing of the inputted spoken words(240); selecting quality with a specific meaning for determining the meaning structure by a result analyzed through the natural language processing(250); performing mechanical studying by using the selected quality(260); and comprehending the spoken words based on the meaning structure formed by determining corresponding elements configuring the meaning structure through the mechanical studying(270).
Abstract:
A device and a method for managing dialog of a chatting agent are provided to improve performance of a chatting robot continuously by providing an answer to the chatting robot based on dialog stacks in consideration of dialog context. A dialog analyzer(120) analyzes a speech sound(110) received from a user. A domain determiner(130) determines and verifies domains to determine the domain of the speech sound based on a speech sound analysis result. A dialog expert(140) provides a system speech sound(170) in response to the user speech sound by searching a dialog example database constructed from a dialog corpus. The domain determiner includes a language/semantic quality extractor(121,122) and a keyword extractor(123), and determines the domain by integrating and applying extracted qualities to a quality based probability model. The dialog expert includes a dialog stack(150) storing dialog example session information of the previous user speech sound and dialog/main flow analysis result, and a dialog example selector(160) indexing information such as semantic information of the previous user speech sound and scenario session information added from a scenario based dialog corpus.
Abstract:
A system and a method for dialog management are provided to improve the responding accuracy of the system by using a dialog example extracted from dialog corpus so as to predict the response of the system. An input unit(113) receives voice or text from user speech. A language comprehension unit(115) extracts a meaning frame of the user speech by using character string information and morpheme analysis result from a character string signal inputted from the input unit(113). An agent determination unit(117) analyzes the meaning frame of the user speech to determine whether the user speech is for simple chatting or purposeful dialog based on a trained possibility model using keyword feature extracted from dialog corpus, language analysis feature, and meaning analysis feature. A domain determination unit(118) determines the domain type of the user speech. A dialog example selection unit(120) selects the most appropriate dialog example by constructing dialog example database from the dialog corpus and searching for dialog examples using the meaning frame and dialog history. A response generation unit(122) generates device speech by using the selected dialog example.