Patent application number | Description | Published |
20140085215 | PROGRESS DISPLAY OF HANDWRITING INPUT - A computer-implemented method includes: receiving, at a user device, user input corresponding to handwritten text to be recognized using a recognition engine; and receiving, at the user device, a representation of the handwritten text. The representation includes the handwritten text parsed into individual handwritten characters. The method further includes: displaying, on a display of the user device, the handwritten characters using a first indicator; receiving, at the user device, an identification of a text character recognized as one of the handwritten characters; displaying, on the display, the text character; and adjusting, at the user device, the one of the handwritten characters from being displayed using the first indicator to using a second indicator in response to the received identification. The first and second indicators are different. | 03-27-2014 |
20140258263 | Predictive Information Retrieval - A computer-implemented method for generating results for a client-requested query involves receiving a query produced by a client communication device, generating a result for the query in response to reception of the query, determining one or more predictive follow-up requests before receiving an actual follow-up request from the client device, and initiating retrieval of information associated with the one or more predictive follow-up requests, and transmitting at least part of the result to the client device, and then transmitting to the client device at least part of the information associated with the one or more predictive follow-up requests. | 09-11-2014 |
20140270329 | EXTRACTION OF FINANCIAL ACCOUNT INFORMATION FROM A DIGITAL IMAGE OF A CARD - Capturing information from payment instruments comprises receiving, using one or more computer devices, an image of a back side of a payment instrument, the payment instrument comprising information imprinted thereon such that the imprinted information protrudes from a front side of the payment instrument and the imprinted information is indented into the back side of the payment instrument; extracting sets of characters from the image of the back side of the payment instrument based on the imprinted information indented into the back side of the payment instrument and depicted in the image of the back side of the payment instrument; applying a first character recognition application to process the sets of characters extracted from the image of the back side of the payment instrument; and categorizing each of the sets of characters into one of a plurality of categories relating to information required to conduct a payment transaction. | 09-18-2014 |
20150023604 | CARD ART DISPLAY - Providing improved card art for display comprises receiving, by one or more computing devices, an image of a card and performing an image recognition algorithm on the image. The computing device identifies images represented on the card image and comparing the identified images to an image database. The computing device determines a standard card art image associated with the identified image based at least in part on the comparison and associates the standard card art image with an account of a user, the account being associated with the card in the image. The computing device displays the standard card art as a representation of the account. | 01-22-2015 |
20150086069 | EXTRACTING CARD DATA USING IIN DATABASE - Extracting card data comprises receiving, by one or more computing devices, a digital image of a card; perform an image recognition process on the digital representation of the card; identifying an image in the digital representation of the card; comparing the identified image to an image database comprising a plurality of images and determining that the identified image matches a stored image in the image database; determining a card type associated with the stored image and associating the card type with the card based on the determination that the identified image matches the stored image; and performing a particular optical character recognition algorithm on the digital representation of the card, the particular optical character recognition algorithm being based on the determined card type. Another example uses an issuer identification number to improve data extraction. Another example compares extracted data with user data to improve accuracy. | 03-26-2015 |
20150161174 | CONTENT-BASED IMAGE RANKING - Methods, systems, and apparatus, including computer program products, for ranking search results for queries. The method includes calculating a visual similarity score for one or more pairs of images in a plurality of images based on visual features of images in each of the one or more pairs; building a graph of images by linking each of one or more images in the plurality of images to one or more nearest neighbor images based on the visual similarity scores; associating a respective score with each of one or more images in the graph based on data indicative of user behavior relative to the image as a search result for a query; and determining a new score for each of one or more images in the graph based on the respective score of the image, and the respective scores of one or more nearest neighbors to the image. | 06-11-2015 |
20150186718 | Segmentation of Overwritten Online Handwriting Input - Implementations of the disclosed subject matter provide methods and systems for identifying a candidate character cut for an overwritten character. A method may include providing a handwriting input area. The handwriting input area may be divided into multiple sections and a first portion of the multiple sections may be located in an end point region. A first handwritten input comprising a first stroke that ends in a section located in the end point region may be received. A second handwritten input comprising a second stroke that begins in a section that is not located in the end point region may be received. As a result, a first candidate character cut may be identified between the first stroke and the second stroke. | 07-02-2015 |
20150186740 | EXTRACTING CARD DATA WITH CARD MODELS - Embodiments herein provide computer-implemented techniques for allowing a user computing device to extract financial card information using optical character recognition (“OCR”). Extracting financial card information may be improved by applying various classifiers and other transformations to the image data. For example, applying a linear classifier to the image to determine digit locations before applying the OCR algorithm allows the user computing device to use less processing capacity to extract accurate card data. The OCR application may train a classifier to use the wear patterns of a card to improve OCR algorithm performance. The OCR application may apply a linear classifier and then a nonlinear classifier to improve the performance and the accuracy of the OCR algorithm. The OCR application uses the known digit patterns used by typical credit and debit cards to improve the accuracy of the OCR algorithm. | 07-02-2015 |
20150186797 | DATA REDUCTION IN NEAREST NEIGHBOR CLASSIFICATION - A set S is initialized. Initially, S is empty; but, as the disclosed process is performed, items are added to it. It may contain one or more samples (e.g., items) from each class. One or more labeled samples for one or more classes may be obtained. A series of operations may be performed, iteratively, until a stopping criterion is reach to obtain the reduced set. For each class of the one or more classes, a point may be generated based on at least one sample in the class having a nearest neighbor in a set S with a different class label than the sample. The point may be added to the set S. The process may be repeated unless a stopping criterion is reached. A nearest neighbor for a submitted point in the set S may be identified and a candidate nearest neighbor may be output for the submitted point. | 07-02-2015 |