Class / Patent application number | Description | Number of patent applications / Date published |
704276000 | Pattern display | 30 |
20080221904 | COARTICULATION METHOD FOR AUDIO-VISUAL TEXT-TO-SPEECH SYNTHESIS - A method for generating animated sequences of talking heads in text-to-speech applications wherein a processor samples a plurality of frames comprising image samples. The processor reads first data comprising one or more parameters associated with noise-producing orifice images of sequences of at least three concatenated phonemes which correspond to an input stimulus. The processor reads, based on the first data. second data comprising images of a noise-producing entity. The processor generates an animated sequence of the noise-producing entity. | 09-11-2008 |
20080228497 | Method For Communication and Communication Device - The invention describes a method for communication by means of a communication device (DS), in which synthesized speech (ss) is output from the communication device (DS), and in which light signals (ls) are output simultaneously with the synthesized speech (ss) in accordance with the semantic content of the synthesized speech (ss). Furthermore, an appropriate communication device (DS) is described. | 09-18-2008 |
20080228498 | Enhanced coordinated signal generation apparatus - A portable information system that, when activated, presents certain data to a user, wherein the information is prepositioned within the portable system or provided to it. The information may be resident within fixed memory, added by means of a smart card, or wirelessly transmitted to the information system. The information conveyed to the user may be visual, such as a text screen or a video display as well as audible such as a play by play broadcast. The information may include local areas of interest or locations within the venue in which the apparatus is used. | 09-18-2008 |
20090099850 | Vocal Command Directives To Compose Dynamic Display Text - Methods, apparatus, products are disclosed for displaying speech for a user of a surface computer, the surface computer comprising a surface, the surface computer capable of receiving multi-touch input through the surface and rendering display output on the surface, that include: registering, by the surface computer, a plurality of users with the surface computer; allocating, by the surface computer to each registered user, a portion of the surface for interaction between that registered user and the surface computer; detecting, by the surface computer, a speech utterance from one of the plurality of users; determining, by the surface computer using a speech engine, speech text in dependence upon the speech utterance; creating, by the surface computer, display text in dependence upon the speech text; and rendering, by the surface computer, the display text on at least one of the allocated portions of the surface. | 04-16-2009 |
20090125312 | METHOD AND SYSTEM FOR PROVIDING NEWS INFORMATION BY USING THREE DIMENSIONAL CHARACTER FOR USE IN WIRELESS COMMUNICATION NETWORK - Disclosed is a method for providing by a news information-providing server news information using a 3D character to a wireless communication terminal having accessed the news information-providing server through a wireless communication network, the method including the steps of: (a) generating voice information by converting news information received in real-time into voice data, and analyzing content of the voice information; (b) extracting mouth shape data and facial expression data corresponding to the content of the voice information analyzed at step (a); (c) applying the mouth shape data and facial expression data to the 3D character, and generating 3D character data by synthesizing the 3D character with an background image and/or background music; (d) generating 3D character news by synchronizing the voice information with the 3D character data; and (e) transmitting the 3D character news to the wireless communication terminal in a streaming mode. | 05-14-2009 |
20090132257 | System and method for inputting edited translation words or sentence - A system and a method for inputting edited translation words or sentence are provided to solve the problem that editing translation words or sentence and inputting the edited translation words or sentence cannot be performed by successive actions. In this system and method, the input words or sentence input into an input region by a user are intercepted and translated into translation words or sentence, and a function of editing the translation words or sentence is provided in a display region for displaying the translation words or sentence, thereby achieving the efficacy of inputting the edited translation words or sentence into the input region directly. | 05-21-2009 |
20090248421 | Arrangement for Creating and Using a Phonetic-Alphabet Representation of a Name of a Party to a Call - A first party creates and edits a phonetic-alphabet representation of its name. The phonetic representation is conveyed to a second party as “caller-identification” information by messages that set up a call between the parties. The phonetic representation of the name is displayed to the second party, converted to speech, and/or converted to an alphabet of a language of the second party and then displayed to the second party. | 10-01-2009 |
20090259475 | VOICE QUALITY CHANGE PORTION LOCATING APPARATUS - A text edit apparatus which presents, based on language analysis information regarding a text, a portion of the text where voice quality may change when the text is read aloud has advantages of predicting likelihood of the voice quality change and judging whether or not the voice quality change will occur. The apparatus includes: a voice quality change estimation unit ( | 10-15-2009 |
20090281810 | SYSTEM AND METHOD FOR VISUALLY PRESENTING AUDIO SIGNALS - A method of visually presenting audio signals includes receiving an audio signal to be presented; generating a predetermined number of discrete frequency components from the audio signal; assigning a graphical object to each of the frequency components, each of the graphical objects being specified by a geometrical shape, a position information and a size information; and all of the graphical objects associated with all of the frequency components are displayed simultaneously on a graphic display. The system includes a microphone for generating audio signals; an audio interface unit for sampling the audio signals and transforming them into digital signals; a processing unit for translating digital signals into a predetermined number of discrete frequency components and for assigning a graphical object to each of the discrete frequency components; a video interface unit for generating a video signal; and a graphic display for displaying a sonogram based on the video signal. | 11-12-2009 |
20100211397 | FACIAL EXPRESSION REPRESENTATION APPARATUS - An avatar facial expression representation technology is provided. The avatar facial expression representation technology estimates changes in emotion and emphasis in a user's voice from vocal information, and changes in mouth shape of the user from pronunciation information of the voice. The avatar facial expression technology tracks a user's facial movements and changes in facial expression from image information and may represent avatar facial expressions based on the result of the these operations. Accordingly, the avatar facial expressions can be obtained which are similar to actual facial expressions of the user. | 08-19-2010 |
20100250256 | CHARACTER MOUTH SHAPE CONTROL METHOD | 09-30-2010 |
20110087493 | Communication System and Method for Representing Information in a Communication - The invention relates to a communication system having a display unit ( | 04-14-2011 |
20110093274 | APPARATUS AND METHOD OF MANUFACTURING ARTICLE USING SOUND - Disclosed is an apparatus and method of manufacturing an article using sound that modifies sound waveforms for sound of living things (including human voice) in various shapes and manufactures articles corresponding to the shapes. An apparatus for manufacturing an article using sound generates a sampling waveform based on the sound waveform. Next, the sampling waveform is converted into a two-dimensional image file and the two-dimensional image is again converted into a three-dimensional image file. Thereafter, an article is manufactured based on the two-dimensional or three-dimensional image file. According to the invention, the apparatus and method of manufacturing an article using sound manufactures an article based on the sampling waveform generated by sampling the sound waveform, thereby manufacturing a simplified article. | 04-21-2011 |
20110231194 | Interactive Speech Preparation - In an embodiment, a method of interactive speech preparation is disclosed. The method may include or comprise displaying an interactive speech application on a display device, wherein the interactive speech application has a text display window. The method may also include or comprise accessing text stored in an external storage device over a communication network, and displaying the text within the text display window while capturing video and audio data with video and audio data capturing devices, respectively. | 09-22-2011 |
20120130720 | INFORMATION PROVIDING DEVICE - An information providing device takes an image of a predetermined area and obtains the taken image in the form of image data, while externally obtaining voice data representing speech. The information providing device obtains text in a preset language corresponding to the speech in the form of text data, based on the obtained voice data, generates a composite image including the taken image and the text in the form of composite image data, based on the image data and the text data, and outputs the composite image data. | 05-24-2012 |
20120323581 | Systems and Methods for Voice Personalization of Video Content - Systems and methods are disclosed for performing voice personalization of video content. The personalized media content may include a composition of a background scene having a character, head model data representing an individualized three-dimensional (3D) head model of a user, audio data simulating the user's voice, and a viseme track containing instructions for causing the individualized 3D head model to lip sync the words contained in the audio data. The audio data simulating the user's voice can be generated using a voice transformation process. In certain examples, the audio data is based on a text input or selected by the user (e.g., via a telephone or computer) or a textual dialogue of a background character. | 12-20-2012 |
20130054249 | VISUALIZING, NAVIGATING AND INTERACTING WITH AUDIO CONTENT - Methods and arrangements for visually representing audio content in a voice application. A display is connected to a voice application, and an image is displayed on the display, the image comprising a main portion and at least one subsidiary portion, the main portion representing a contextual entity of the audio content and the at least one subsidiary portion representing at least one participatory entity of the audio content. The at least one subsidiary portion is displayed without text, and the image is changed responsive to changes in audio content in the voice application. | 02-28-2013 |
20130054250 | VISUALIZING, NAVIGATING AND INTERACTING WITH AUDIO CONTENT - Methods and arrangements for visually representing audio content in a voice application. A display is connected to a voice application, and an image is displayed on the display, the image comprising a main portion and at least one subsidiary portion, the main portion representing a contextual entity of the audio content and the at least one subsidiary portion representing at least one participatory entity of the audio content. The at least one subsidiary portion is displayed without text, and the image is changed responsive to changes in audio content in the voice application. | 02-28-2013 |
20130124212 | Method and Apparatus for Time Synchronized Script Metadata - A method includes receiving script data including script words for dialogue, receiving audio data corresponding to at least a portion of the dialogue, wherein the audio data includes timecodes associated with dialogue words, generating a sequential alignment of the script words to the dialogue words, matching at least some of the script words to corresponding dialogue words to determine hard alignment points, partitioning the sequential alignment of script words into alignment sub-sets, wherein the bounds of the alignment sub-subsets are defined by adjacent hard-alignment points, and wherein the alignment subsets includes a sub-set of the script words and a corresponding sub-set of dialogue words that occur between the hard-alignment points, determining corresponding timecodes for a sub-set of script words in a sub-subset based on the timecodes associated with the sub-set of dialogue words, and generating time-aligned script data including the sub-set of script words and their corresponding timecodes. | 05-16-2013 |
20130124213 | Method and Apparatus for Interpolating Script Data - Provided in some embodiments is a computer implemented method that includes providing script data including script words indicative of dialogue words to be spoken, providing audio data corresponding to at least a portion of the dialogue words to be spoken, wherein the audio data includes timecodes associated with dialogue words, generating a sequential alignment of the script words to the dialogue words, matching at least some of the script words to corresponding dialogue words to determine alignment points, determining corresponding timecodes for unmatched script words using interpolation based on the timecodes associated with matching script words, and generating time-aligned script data including the script words and their corresponding time codes. | 05-16-2013 |
20130226593 | AUDIO PROCESSING APPARATUS - An apparatus comprising: an audio source determiner configured to determine at least one audio source; a visualizer configured to generate a visual representation associated with the at least one audio source; and a controller configured to process an audio signal associated with the at least one audio source dependent on interaction with the visual representation. | 08-29-2013 |
20140129235 | AUDIO TRACKER APPARATUS - Apparatus comprising a receiver configured to receive a first audio signal, a signal characteriser configured to determine at least one characteristic associated with the first audio signal, a comparator configured to compare the at least one characteristic against at least one characteristic associated with at least one further audio signal, and a display configured to display the at least one characteristic associated with at least one further audio signal dependent on the first audio signal characteristic. | 05-08-2014 |
20140142954 | SOUNDTRACK FOR ELECTRONIC TEXT - A soundtrack creation method and user playback system for soundtracks synchronized to electronic text. Synchronization is achieved by maintaining a reading speed variable indicative of the user's reading speed. The system provides for multiple channels of audio to enable concurrent playback of two or more partially or entirely overlapping audio regions so as to create an audio output having, for example, sound effects, ambience, music or other audio features that are triggered to playback at specific portions in the electronic text to enhance the reading experience. | 05-22-2014 |
20140172432 | DISPLAY DEVICE, HEAD-MOUNT TYPE DISPLAY DEVICE, METHOD OF CONTROLLING DISPLAY DEVICE, AND METHOD OF CONTROLLING HEAD-MOUNT TYPE DISPLAY DEVICE - A transmissive display device includes an image display section adapted to generate image light representing an image, allow a user to visually recognize the image light, and transmit an external sight, a sound acquisition section adapted to obtain a sound, a conversion section adapted to convert the sound into a character image expressing the sound as an image using characters, a specific direction setting section adapted to set a specific direction, and a display position setting section adapted to set an image display position, which is a position where character image light representing the character image is made to be visually recognized in a visual field of the user, based on the specific direction. | 06-19-2014 |
20140288939 | METHOD AND APPARATUS FOR OPTIMIZING TIMING OF AUDIO COMMANDS BASED ON RECOGNIZED AUDIO PATTERNS - An approach is provided for timing application information presentation based on audio patterns. The audio platform processes and/or facilitates a processing of one or more audio samples to determine a conversational state of one or more users. Next, the audio platform determines a timing for at least one presentation of application information on a device associated with at least one of the one or more users based, at least in part, on the conversational state. | 09-25-2014 |
20150032460 | TERMINAL AND SPEECH-RECOGNIZED TEXT EDIT METHOD THEREOF - A terminal and speech-recognized text edit method edit the text input through writing recognition or speech recognition function efficiently. The text edit method includes displaying at least one letter input through speech recognition; detecting one of touch and speech inputs; analyzing the detected input; and performing a certain operation corresponding to the at least one letter based on the analysis result. The terminal and speech-recognized text edit method are advantageous in editing misrecognized speech-input text efficiently though finger or pen gesture-based or speech recognition-based input. | 01-29-2015 |
20150364146 | Method for Providing Visual Feedback for Vowel Quality - A method for obtaining visual feedback for vowel quality is disclosed. By filtering and decomposing audio signals in formants, one or more software modules and/or applications may determine the type of phoneme in an audio signal that corresponds to a vowel. First and second formants may be measured in their frequencies; subsequently frequencies values of first formant and second formant may be assigned as X and Y coordinates, respectively. These (X,Y) coordinates may be used for graphing points and trajectories of the points in a Cartesian coordinate system while the user is pronouncing a vowel. Thus, graphic feedback for vowel pronunciation may be obtained. | 12-17-2015 |
20160035352 | VOICE RECOGNITION SYSTEM AND RECOGNITION RESULT DISPLAY APPARATUS - A voice recognition system, when deciding that a user is about to select a voice recognition result displayed on a screen while the system continuously recognizes a voice uttered, stops a display update of the recognition result on the display unit, thereby being able to prevent the recognition result from disappearing or from changing to other recognition result at the moment when the user is about to select the recognition result. This makes it possible to prevent a user from carrying out an unwanted operation, and to improve the convenience of the user. | 02-04-2016 |
20160379666 | EMPLOYING USER INPUT TO FACILITATE INFERENTIAL SOUND RECOGNITION BASED ON PATTERNS OF SOUND PRIMITIVES - The disclosed embodiments provide a system that generates sound primitives to facilitate sound recognition. First, the system performs a feature-detection operation on sound samples to detect a set of sound features, wherein each sound feature comprises a measurable characteristic of a window of consecutive sound samples. Next, the system creates feature vectors from coefficients generated by the feature-detection operation, wherein each feature vector comprises a set of coefficients for sound features detected in a window. The system then performs a clustering operation on the feature vectors to produce feature-vector clusters, wherein each feature-vector cluster comprises a set of feature vectors that are proximate to each other in a feature-vector space that contains the feature vectors. After the clustering operation, the system defines a set of sound primitives, wherein each sound primitive is associated with a feature-vector cluster. Finally, the system associates semantic labels with the set of sound primitives. | 12-29-2016 |
20160379668 | STRESS REDUCTION AND RESILIENCY TRAINING TOOL - A computer-implemented method, a computing device, and a computer program product (such as a mobile application) are described that implement a CBT and mindfulness training tool. The method can be embodied in the product and executed by the computing device. The method can include receiving first input from a user that represents a description of cognitive conditions of the user. The first input can include factual information and at least one conclusion drawn by the user from the factual information. The method can also include receiving a second input from the user that includes factual information inconsistent with the at least one conclusion. The method can also include determining impugning material based on the factual information and the at least one conclusion. The impugning material can be configured to be consistent with the factual information of the first input and the factual information of the second input. The impugning material can be output to the user. | 12-29-2016 |