Patent application number | Description | Published |
20110142370 | GENERATING A COMPOSITE IMAGE FROM VIDEO FRAMES - A method described herein includes acts of receiving a sequence of images of a scene and receiving an indication of a reference image in the sequence of images. The method further includes an act of automatically assigning one or more weights independently to each pixel in each image in the sequence of images of the scene. Additionally, the method includes an act of automatically generating a composite image based at least in part upon the one or more weights assigned to each pixel in each image in the sequence of images of the scene. | 06-16-2011 |
20110211758 | MULTI-IMAGE SHARPENING AND DENOISING USING LUCKY IMAGING - The multi-image sharpening and denoising technique described herein creates a clean (low-noise, high contrast), detailed image of a scene from a temporal series of images of the scene. The technique employs a process of image alignment to remove global and local camera motion plus a novel weighted image averaging procedure that avoids sacrificing sharpness to create a resultant high-detail, low-noise image from the temporal series. In addition, the multi-image sharpening and denoising technique can employ a dehazing procedure that uses a spatially varying airlight model to dehaze an input image. | 09-01-2011 |
20110304687 | GENERATING SHARP IMAGES, PANORAMAS, AND VIDEOS FROM MOTION-BLURRED VIDEOS - A “Blur Remover” provides various techniques for constructing deblurred images from a sequence of motion-blurred images such as a video sequence of a scene. Significantly, this deblurring is accomplished without requiring specialized side information or camera setups. In fact, the Blur Remover receives sequential images, such as a typical video stream captured using conventional digital video capture devices, and directly processes those images to generate or construct deblurred images for use in a variety of applications. No other input beyond the video stream is required for a variety of the embodiments enabled by the Blur Remover. More specifically, the Blur Remover uses joint global motion estimation and multi-frame deblurring with optional automatic video “duty cycle” estimation to construct deblurred images from video sequences for use in a variety of applications. Further, the automatically estimated video duty cycle is also separately usable in a variety of applications. | 12-15-2011 |
20120314899 | NATURAL USER INTERFACES FOR MOBILE IMAGE VIEWING - The mobile image viewing technique described herein provides a hands-free interface for viewing large imagery (e.g., 360 degree panoramas, parallax image sequences, and long multi-perspective panoramas) on mobile devices. The technique controls the imagery displayed on a display of a mobile device by movement of the mobile device. The technique uses sensors to track the mobile device's orientation and position, and front facing camera to track the user's viewing distance and viewing angle. The technique adjusts the view of a rendered imagery on the mobile device's display according to the tracked data. In one embodiment the technique can employ a sensor fusion methodology that combines viewer tracking using a front facing camera with gyroscope data from the mobile device to produce a robust signal that defines the viewer's 3D position relative to the display. | 12-13-2012 |
20130003196 | NON-PLANAR FOCAL SURFACE LENS ASSEMBLY - A lens assembly includes a plurality of component lens elements, and a fiber optic face plate having a back surface and a non-planar front surface. The plurality of component lens elements are configured to direct a focused image onto the non-planar front surface of the fiber optic face plate, and the fiber optic face plate is configured to transmit the focused image through the back surface. | 01-03-2013 |
20130229581 | JUXTAPOSING STILL AND DYNAMIC IMAGERY FOR CLIPLET CREATION - Various technologies described herein pertain to juxtaposing still and dynamic imagery to create a cliplet. A first subset of a spatiotemporal volume of pixels in an input video can be set as a static input segment, and the static input segment can be mapped to a background of the cliplet. Further, a second subset of the spatiotemporal volume of pixels in the input video can be set as a dynamic input segment based on a selection of a spatial region, a start time, and an end time within the input video. Moreover, the dynamic input segment can be refined spatially and/or temporally and mapped to an output segment of the cliplet within at least a portion of output frames of the cliplet based on a predefined temporal mapping function, and the output segment can be composited over the background for the output frames of the cliplet. | 09-05-2013 |
20140293074 | GENERATING A COMPOSITE IMAGE FROM VIDEO FRAMES - A method described herein includes acts of receiving a sequence of images of a scene and receiving an indication of a reference image in the sequence of images. The method further includes an act of automatically assigning one or more weights independently to each pixel in each image in the sequence of images of the scene. Additionally, the method includes an act of automatically generating a composite image based at least in part upon the one or more weights assigned to each pixel in each image in the sequence of images of the scene. | 10-02-2014 |
20140327680 | AUTOMATED VIDEO LOOPING WITH PROGRESSIVE DYNAMISM - Various technologies described herein pertain to generating a video loop. An input video can be received, where the input video includes values at pixels over a time range. An optimization can be performed to determine a respective input time interval within the time range of the input video for each pixel from the pixels in the input video. The respective input time interval for a particular pixel can include a per-pixel loop period and a per-pixel start time of a loop at the particular pixel within the time range from the input video. Moreover, an output video can be created based upon the values at the pixels over the respective input time intervals for the pixels in the input video. | 11-06-2014 |