Patent application number | Description | Published |
20120233159 | HIERARCHICAL RANKING OF FACIAL ATTRIBUTES - In response to a query of discernible facial attributes, the locations of distinct and different facial regions are estimated from face image data, each relevant to different attributes. Different features are extracted from the estimated facial regions from database facial images, which are ranked in base layer rankings by matching feature vectors in a bipartite graph to a base layer ranking sequence as a function of edge weights parameterized by an associated base layer parameter vector. Second-layer rankings define second-layer attribute vectors as bilinear combinations of the base-layer feature vectors and associated base layer parameter vectors for common attributes, which are matched in a bipartite graph to a second-layer ranking sequence as a function of edge weights parameterized by associated second-layer parameter vectors. The images are thus ranked for relevance to the query as a function of fusing the second-layer rankings. | 09-13-2012 |
20120237081 | ANOMALOUS PATTERN DISCOVERY - A trajectory of movement of an object is tracked in a video data image field that is partitioned into a plurality of different grids. Global image features from video data relative to the trajectory are extracted and compared to a learned trajectory model to generate a global anomaly detection confidence decision value as a function of fitting to the learned trajectory model. Local image features are also extracted for each of the image field grids that include object trajectory, which are compared to learned feature models for the grids to generate local anomaly detection confidence decisions for each grid as a function of fitting to the learned feature models for the grids. The global anomaly detection confidence decision value and the local anomaly detection confidence decision values for the grids are into a fused anomaly decision with respect to the tracked object. | 09-20-2012 |
20120242832 | VEHICLE HEADLIGHT MANAGEMENT - A method, data processing system, and computer program product for managing a headlight on a vehicle are presented. An image of an area in front of the vehicle is received. A first set of features is identified in the received image. The first set of features in the received image is compared with a number of sets of features from a plurality of previous images. Each image in the plurality of previous images is associated with a headlight setting. A second set of features from a previous image in the plurality of previous images matching the first set of features in the received image is identified. A determination is made whether to change a setting for the headlight on the vehicle based on the headlight setting associated with the previous image. | 09-27-2012 |
20120263346 | VIDEO-BASED DETECTION OF MULTIPLE OBJECT TYPES UNDER VARYING POSES - Training data object images are clustered as a function of motion direction attributes and resized from respective original into same aspect ratios. Motionlet detectors are learned for each of the sets from features extracted from the resized object blobs. A deformable sliding window is applied to detect an object blob in input by varying window size, shape or aspect ratio to conform to a shape of the detected input video object blob. A motion direction of an underlying image patch of the detected input video object blob is extracted and motionlet detectors selected and applied that have similar motion directions. An object is thus detected within the detected blob and semantic attributes of an underlying image patch extracted if a motionlet detectors fires, the extracted semantic attributes available for use for searching for the detected object. | 10-18-2012 |
20120281873 | INCORPORATING VIDEO META-DATA IN 3D MODELS - A moving object detected and tracked within a field of view environment of a 2D data feed of a calibrated video camera is represented by a 3D model through localizing a centroid of the object and determining an intersection with a ground-plane within the field of view environment. An appropriate 3D mesh-based volumetric model for the object is initialized by using a back-projection of a corresponding 2D image as a function of the centroid and the determined ground-plane intersection. Nonlinear dynamics of a tracked motion path of the object are represented as a collection of different local linear models. A texture of the object is projected onto the 3D model, and 2D tracks of the object are upgraded to 3D motion to drive the 3D model by learning a weighted combination of the different local linear models that minimizes an image re-projection error of model movement. | 11-08-2012 |
20120284211 | IDENTIFYING ABNORMALITIES IN RESOURCE USAGE - A method, data processing system, and computer program product for identifying abnormalities in data. A model representing a plurality of modes for an activity generated from training data is retrieved. The training data includes a first plurality of measurements of a first performance of the activity over a period of time. Each of the plurality of modes is identified as one of normal and abnormal. Activity data including a second plurality of measurements of a second performance of the activity is received. A portion of the activity data is compared with the plurality of modes in the model. A notification of an abnormality in the second performance of the activity is generated in response to an identification that the portion of the activity data matches a mode in the plurality of modes identified as abnormal. Confirmation of the abnormality is requested via a user interface. | 11-08-2012 |
20120294511 | EFFICIENT RETRIEVAL OF ANOMALOUS EVENTS WITH PRIORITY LEARNING - Local models learned from anomaly detection are used to rank detected anomalies. The local models include image feature values extracted from an image field of video image data with respect to different predefined spatial and temporal local units, wherein anomaly results are determined by failures to fit to applied anomaly detection module local models. Image features values extracted from the image field local units associated with anomaly results are normalized, and image feature values extracted from the image field local units are clustered. Weights for anomaly results are learned as a function of the relations of the normalized extracted image feature values to the clustered image feature values. The normalized values are multiplied by the learned weights to generate ranking values to rank the anomalies. | 11-22-2012 |
20120308121 | IMAGE RANKING BASED ON ATTRIBUTE CORRELATION - Images are retrieved and ranked according to relevance to attributes of a multi-attribute query through training image attribute detectors for different attributes annotated in a training dataset. Pair-wise correlations are learned between pairs of the annotated attributes from the training dataset of images. Image datasets may then be searched via the trained attribute detectors for images comprising attributes in a multi-attribute query, wherein images are retrieved from the searching that each comprise one or more of the query attributes and also in response to information from the trained attribute detectors corresponding to attributes that are not a part of the query but are relevant to the query attributes as a function of the learned plurality of pair-wise correlations. The retrieved images are ranked as a function of respective total numbers of attributes within the query subset attributes. | 12-06-2012 |
20120314030 | ESTIMATION OF OBJECT PROPERTIES IN 3D WORLD - Objects within two-dimensional (2D) video data are modeled by three-dimensional (3D) models as a function of object type and motion through manually calibrating a 2D image to the three spatial dimensions of a 3D modeling cube. Calibrated 3D locations of an object in motion in the 2D image field of view of a video data input are computed and used to determine a heading direction of the object as a function of the camera calibration and determined movement between the computed 3D locations. The 2D object image is replaced in the video data input with an object-type 3D polygonal model having a projected bounding box that best matches a bounding box of an image blob, the model oriented in the determined heading direction. The bounding box of the replacing model is then scaled to fit the object image blob bounding box, and rendered with extracted image features. | 12-13-2012 |
20130124514 | HIERARCHICAL RANKING OF FACIAL ATTRIBUTES - In response to a query of discernable facial attributes, the locations of distinct and different facial regions are estimated from face image data, each relevant to different attributes. Different features are extracted from the estimated facial regions from database facial images, which are ranked in base layer rankings by matching feature vectors to a base layer ranking sequence as a function of edge weights. Second-layer rankings define second-layer attribute vectors as combinations of the base-layer feature vectors and associated base layer parameter vectors for common attributes, which are matched to a second-layer ranking sequence as a function of edge weights. The images are thus ranked for relevance to the query as a function of the second-layer rankings. | 05-16-2013 |
20130241928 | INCORPORATING VIDEO META-DATA IN 3D MODELS - A moving object detected and tracked within a field of view environment of a two-dimensional data feed of a calibrated video camera is represented by a three-dimensional model through localizing a centroid of the object and determining an intersection with a ground-plane within the field of view environment. An appropriate three-dimensional mesh-based volumetric model for the object is initialized by using a back-projection of a corresponding two-dimensional image as a function of the centroid and the determined ground-plane intersection. A texture of the object is projected onto the three-dimensional model, and two-dimensional tracks of the object are upgraded to three-dimensional motion to drive a three-dimensional model. | 09-19-2013 |
20130336581 | MULTI-CUE OBJECT DETECTION AND ANALYSIS - Foreground objects of interest are distinguished from a background model by dividing a region of interest of a video data image into a grid array of individual cells that are each smaller than that a foreground object of interest. More particularly, image data of the foreground object of interest spans a contiguous plurality of the cells. Each of the cells are labeled as foreground if accumulated edge energy within the cell meets an edge energy threshold, if color intensities for different colors within each cell differ by a color intensity differential threshold, or as a function of combinations of said determinations in view of one or more combination rules. | 12-19-2013 |
20140003708 | OBJECT RETRIEVAL IN VIDEO DATA USING COMPLEMENTARY DETECTORS | 01-02-2014 |
20140056476 | INCORPORATING VIDEO META-DATA IN 3D MODELS - A moving object tracked within a field of view environment of a two-dimensional data feed of a calibrated video camera is represented by a three-dimensional model. An appropriate three-dimensional mesh-based volumetric model for the object is initialized by using a back-projection of a corresponding two-dimensional image. A texture of the object is projected onto the three-dimensional model, and two-dimensional tracks of the object are upgraded to three-dimensional motion to drive a three-dimensional model. | 02-27-2014 |
20140098221 | APPEARANCE MODELING FOR OBJECT RE-IDENTIFICATION USING WEIGHTED BRIGHTNESS TRANSFER FUNCTIONS - An approach for re-identifying an object in a first test image is presented. Brightness transfer functions (BTFs) between respective pairs of training images are determined. Respective similarity measures are determined between the first test image and each of the training images captured by the first camera (first training images). A weighted brightness transfer function (WBTF) is determined by combining the BTFs weighted by weights of the first training images. The weights are based on the similarity measures. The first test image is transformed by the WBTF to better match one of the training images captured by the second camera. Another test image, captured by the second camera, is identified because it is closer in appearance to the transformed test image than other test images captured by the second camera. An object in the identified test image is a re-identification of the object in the first test image. | 04-10-2014 |
20140098989 | MULTI-CUE OBJECT ASSOCIATION - Multiple discrete objects within a scene image captured by a single camera track are distinguished as un-labeled from a background model within a first frame of a video data input. Object position and object appearance and/or object size attributes are determined for each of the blobs, and costs determined to assign to existing blobs of existing object tracks as a function of the determined attributes and combined to generate respective combination costs. The un-labeled object blob that has a lowest combined cost of association with any of the existing object tracks is labeled with the label of that track having the lowest combined cost, said track is removed from consideration for labeling remaining un-labeled object blobs, and the process iteratively repeated until each of the track labels have been used to label one of the un-labeled blobs. | 04-10-2014 |
20140122470 | HIERARCHICAL RANKING OF FACIAL ATTRIBUTES - In response to a query of discernable facial attributes, the locations of distinct and different facial regions are estimated from face image data, each relevant to different attributes. Different features are extracted from the estimated facial regions from database facial images, which are ranked in base layer rankings as a function of relevance of extracted features to attributes relevant to the estimated regions, and in second-layer rankings as a function of combinations of the base layer rankings and relevance of the extracted features to common ones of the attributes relevant to the estimated regions. The images are ranked in relevance to the query as a function of the second-layer rankings. | 05-01-2014 |
20140147041 | IMAGE COLOR CORRECTION - Color-correcting a digital image comprising P pixels (P≧4) is presented. Each of the P pixels has a respective color. Color strengths of the P pixels are determined based at least on respective intensities, respective saturations, or both respective intensities and respective saturations of the P pixels. A subset of the P pixels less than all of the P pixels is determined. The pixels in the subset have respective color strengths in a range of respective color strength. All other pixels of the P pixels have respective color strengths outside of the range of respective color strengths. Color correction is determined for the P pixels based in part on the colors of the respective pixels in the subset which are the only pixels of the P pixels used for determining the color correction. The colors of the P pixels are corrected based on the color correction. | 05-29-2014 |
20140164306 | PATHWAY MANAGEMENT USING MODEL ANALYSIS AND FORCASTING - A computer generates a three dimensional map of a pathway area using a plurality of overhead images. The computer determines a forecasted weather pattern to occur in the pathway area. The computer analyzes the three dimensional map and the forecasted weather pattern to predict one or more violations of the pathway. The computer generates a priority for the one or more predicted violations of the pathway. The computer generates a plan for pathway management of the pathway area. | 06-12-2014 |
20140185925 | BOOSTING OBJECT DETECTION PERFORMANCE IN VIDEOS - A method and system for training a special object detector to distinguish a foreground object appearing in a sequence of frames for a target domain. The sequence of frames depicts motion of the foreground object in a non-uniform background. The foreground object is detected in a high-confidence subwindow of an initial frame of the sequence, which includes computing a measure of confidence that the high-confidence subwindow includes the foreground object and determining that the measure of confidence exceeds a specified confidence threshold. The foreground object is tracked in respective positive subwindows of subsequent frames appearing after the initial frame. The subsequent frames are within a specified short period of time. The positive subwindows are used to train the special object detector to detect the foreground object in the target domain. The positive subwindows include the subwindow of the initial frame and the respective subwindows of the subsequent frames. | 07-03-2014 |
20140253732 | TOPOLOGY DETERMINATION FOR NON-OVERLAPPING CAMERA NETWORK - Image-matching tracks the movements of the objects from initial camera scenes to ending camera scenes in non-overlapping cameras. Paths are defined through scenes for pairings of initial and ending cameras by different respective scene entry and exit points. For each of said camera pairings a combination path having a highest total number of tracked movements relative to all other combinations of one path through the initial and ending camera scene is chosen, and the scene exit point of the selected path through the initial camera and the scene entry point of the selected path into the ending camera define a path connection of the initial camera scene to the ending camera scene. | 09-11-2014 |
20140293043 | DETERMINING CAMERA HEIGHT USING DISTRIBUTIONS OF OBJECT HEIGHTS AND OBJECT IMAGE HEIGHTS - A camera at a fixed vertical height positioned above a reference plane, with an axis of a camera lens at an acute angle with respect to a perpendicular of the reference plane. One or more processors receive images of different people. The vertical measurement values of the images of different people are determined. The one or more processors determine a first statistical measure associated with a statistical distribution of the vertical measurement values. The known heights of people from a known statistical distribution of heights of people are transformed to normalized measurements, based in part on a focal length of the camera lens, the angle of the camera, and a division operator in an objective function of differences between the normalized measurements and the vertical measurement values. The fixed vertical height of the camera is determined, based at least on minimizing the objective function. | 10-02-2014 |
20140294231 | AUTOMATICALLY DETERMINING FIELD OF VIEW OVERLAP AMONG MULTIPLE CAMERAS - Field of view overlap among multiple cameras is automatically determined as a function of the temporal overlap of object tracks determined within their fields-of-view. Object tracks with the highest similarity value are assigned into pairs, and portions of the assigned object track pairs having a temporally overlapping period of time are determined. Scene entry points are determined from object locations on the tracks at a beginning of the temporally overlapping period of time, and scene exit points from object locations at an ending of the temporally overlapping period of time. Boundary lines for the overlapping fields-of-view portions within the corresponding camera fields-of-view are defined as a function of the determined entry and exit points in their respective fields-of-view. | 10-02-2014 |
20140314277 | INCORPORATING VIDEO META-DATA IN 3D MODELS - A moving object tracked within a field of view environment of a two-dimensional data feed of a calibrated video camera is represented by a three-dimensional model. An appropriate three-dimensional mesh-based volumetric model for the object is initialized by using a back-projection of a corresponding two-dimensional image. A texture of the object is projected onto the three-dimensional model, and two-dimensional tracks of the object are upgraded to three-dimensional motion to drive a three-dimensional model. | 10-23-2014 |
20140376775 | ESTIMATION OF OBJECT PROPERTIES IN 3D WORLD - Objects within two-dimensional video data are modeled by three-dimensional models as a function of object type and motion through manually calibrating a two-dimensional image to the three spatial dimensions of a three-dimensional modeling cube. Calibrated three-dimensional locations of an object in motion in the two-dimensional image field of view of a video data input are determined and used to determine a heading direction of the object as a function of the camera calibration and determined movement between the determined three-dimensional locations. The two-dimensional object image is replaced in the video data input with an object-type three-dimensional polygonal model having a projected bounding box that best matches a bounding box of an image blob, the model oriented in the determined heading direction. The bounding box of the replacing model is then scaled to fit the object image blob bounding box, and rendered with extracted image features. | 12-25-2014 |
20150023560 | MULTI-CUE OBJECT ASSOCIATION - Multiple discrete objects within a scene image captured by a single camera track are distinguished as un-labeled from a background model within a first frame of a video data input. Object position, object appearance and/or object size attributes are determined for each of the blobs, and costs determined to assign to existing blobs of existing object tracks as a function of the determined attributes. The un-labeled object blob that has a lowest cost of association with any of the existing object tracks is labeled with the label of that track having the lowest cost, said track is removed from consideration for labeling remaining un-labeled object blobs, and the process iteratively repeated until each of the track labels have been used to label one of the un-labeled blobs. | 01-22-2015 |
20150039542 | IMAGE RANKING BASED ON ATTRIBUTE CORRELATION - Images are retrieved and ranked according to relevance to attributes of a multi-attribute query through training image attribute detectors for different attributes annotated in a training dataset. Pair-wise correlations are learned between pairs of the annotated attributes from the training dataset of images. Image datasets may are searched via the trained attribute detectors for images comprising attributes in a multi-attribute query. The retrieved images are ranked as a function of comprising attributes that are not within the query subset plurality of attributes but are paired to one of the query subset plurality of attributes by the pair-wise correlations, wherein the ranking is an order of likelihood that the different ones of the attributes will appear in an image with the paired one of the query subset plurality of attributes. | 02-05-2015 |
20150055830 | AUTOMATICALLY DETERMINING FIELD OF VIEW OVERLAP AMONG MULTIPLE CAMERAS - Field of view overlap among multiple cameras are automatically determined as a function of the temporal overlap of object tracks determined within their fields-of-view. Object tracks with the highest similarity value are assigned into pairs, and portions of the assigned object track pairs having a temporally overlapping period of time are determined. Scene entry points are determined from object locations on the tracks at a beginning of the temporally overlapping period of time, and scene exit points from object locations at an ending of the temporally overlapping period of time. Boundary lines for the overlapping fields-of-view portions within the corresponding camera fields-of-view are defined as a function of the determined entry and exit points in their respective fields-of-view. | 02-26-2015 |
20150062340 | High Occupancy Toll Lane Compliance - A method and system for demonstrating compliance with a requirement of a high occupancy lane in a vehicle for a reduced toll charge for the vehicle is provided. The system includes a housing, an infrared camera within the housing, a GPS unit, a transceiver and a control within the housing. The infrared camera images one or more people in the vehicle. The transceiver detects an RF signal indicating that the vehicle is located at or near a toll booth for the high occupancy lane. The control triggers the infrared camera to image the one or more people and transmit a current time, a current location of the vehicle from the GPS unit, and the triggered image of the one or more people, to a server to demonstrate compliance with the requirement of the high occupancy lane for the reduced toll charge for the vehicle. | 03-05-2015 |
20150063689 | MULTI-CUE OBJECT DETECTION AND ANALYSIS - Foreground objects of interest are distinguished from a background model by dividing a region of interest of a video data image into a grid array of individual cells. Each of the cells are labeled as foreground if accumulated edge energy within the cell meets an edge energy threshold, or if color intensities for different colors within each cell differ by a color intensity differential threshold, or as a function of combinations of said determinations. | 03-05-2015 |