Patent application number | Description | Published |
20100182396 | DATA CAPTURE SYSTEM - A system that facilitates collecting data is described herein. The system includes a digital camera that is configured to capture images in a visible light spectrum and a near-infrared camera that is configured to capture near infrared images, wherein a field of view of the digital camera and the field of view of the near-infrared camera are substantially similar. The system further includes a trigger component that is configured to cause the digital camera and the near-infrared camera to capture images at a substantially similar point in time, and also includes a mounting mechanism that facilitates mounting the digital camera and the near-infrared camera to an automobile. | 07-22-2010 |
20140184585 | REDUNDANT PIXEL MITIGATION - Among other things, one or more techniques and/or systems are provided for mitigating redundant pixel texture contribution for texturing a geometry. That is, the geometry may represent a multidimensional surface of a scene, such as a city. The geometry may be textured using one or more texture images (e.g., an image comprising color values and/or depth values) depicting the scene from various view directions (e.g., a top-down view, an oblique view, etc.). Because more than one texture image may contribute to texturing a pixel of the geometry (e.g., due to overlapping views of the scene), redundant pixel texture contribution may arise. Accordingly, a redundant textured pixel within a texture image may be knocked out (e.g., in-painted) from the texture image to generate a modified texture image that may be relatively efficient to store and/or stream to a client due to enhanced compression of the modified texture image. | 07-03-2014 |
20140184596 | IMAGE BASED RENDERING - Among other things, one or more techniques and/or systems are provided for generating geometry using one or more depth images and/or for texturing geometry using one or more texture imagery. That is, geometry (e.g., a three-dimensional representation of a city) may be generated based upon depth information within a depth image. The geometry may be textured by assigning color values to pixels within the geometry based upon texture imagery (e.g., a video and/or an image comprising depth values and/or color values). For example, a 3D point associated with a pixel of the geometry may be projected to a location within texture imagery. If the depth of the pixel corresponds to a depth of the location, then texture information (e.g., a color value) from the texture imagery may be assigned to the pixel. In this way, the textured geometry may be used to generate a rendered image. | 07-03-2014 |
20140300736 | MULTI-SENSOR CAMERA RECALIBRATION - One or more techniques and/or systems are providing for facilitating recalibration of a multi-sensor camera. That is, a multi-sensor camera may comprise a nadir sensor and one or more oblique sensors. Temperature, mechanical stress, and other factors can lead to misalignment of one or more sensors within the multi-sensor camera. Accordingly, a set of tie points and/or observations may be generated based upon a search matching technique, a densification technique, and/or a virtual matching technique. A bundle technique may be utilized to generate updated eccentricity information based upon the set of tie points and/or observations. The updated eccentricity information (e.g., orientation and/or position information of a sensor, such as an oblique sensor, with respect to a nadir view) may be used to recalibrate the multi-sensor camera, such as in real-time (e.g., during a flight mission that utilizes the multi-sensor camera to capture aerial images of a city or other scene). | 10-09-2014 |
20140368651 | MOBILE IMAGING PLATFORM CALIBRATION - Mobile platforms are used to capture an area using a variety of sensors (e.g., cameras and laser scanners) while traveling through the area, in order to create a representation (e.g., a navigable set of panoramic images, or a three-dimensional reconstruction). However, such sensors are often precisely calibrated in a controlled setting, and miscalibration during travel (e.g., due to a physical jolt) may result in a corruption of data and/or a recalibration that leaves the platform out of service for an extended duration. Presented herein are techniques for verifying sensor calibration during travel. Such techniques involve the identification of a sensor path for each sensor over time (e.g., a laser scanner path, a camera path, and a location sensor path) and a comparison of the paths, optionally after registration with a static coordinate system, to verify that the continued calibration of the sensors during the mobile operation of the platform. | 12-18-2014 |
20150326783 | THREE DIMENSIONAL IMAGE CAPTURE SYSTEM FOR IMAGING BUILDING FACADES USING A DIGITAL CAMERA, A NEAR-INFRARED CAMERA, AND LASER RANGE FINDER - A system that facilitates collecting data is described herein. The system includes a digital camera that is configured to capture images in a visible light spectrum and a near-infrared camera that is configured to capture near infrared images, wherein a field of view of the digital camera and the field of view of the near-infrared camera are substantially similar. The system further includes a trigger component that is configured to cause the digital camera and the near-infrared camera to capture images at a substantially similar point in time, and also includes a mounting mechanism that facilitates mounting the digital camera and the near-infrared camera to an automobile. | 11-12-2015 |
20150339843 | VIEW DIRECTION DETERMINATION - Among other things, one or more techniques and/or systems are provided for defining a view direction for a texture image used to texture a geometry. That is, a geometry may represent a multi-dimensional surface of a scene, such as a city. The geometry may be textured using one or more texture images depicting the scene from various view directions. Because more than one texture image may contribute to texturing portions of the geometry, a view direction for a texture image may be selectively defined based upon a coverage metric associated with an amount of non-textured geometry pixels that are textured by the texture image along the view direction. In an example, a texture image may be defined according to a customized configuration, such as a spherical configuration, a cylindrical configuration, etc. In this way, redundant texturing of the geometry may be mitigated based upon the selectively identified view direction(s). | 11-26-2015 |
20160071306 | REDUNDANT PIXEL MITIGATION - Among other things, one or more techniques and/or systems are provided for mitigating redundant pixel texture contribution for texturing a geometry. That is, the geometry may represent a multidimensional surface of a scene, such as a city. The geometry may be textured using one or more texture images (e.g., an image comprising color values and/or depth values) depicting the scene from various view directions (e.g., a top-down view, an oblique view, etc.). Because more than one texture image may contribute to texturing a pixel of the geometry (e.g., due to overlapping views of the scene), redundant pixel texture contribution may arise. Accordingly, a redundant textured pixel within a texture image may be knocked out (e.g., in-painted) from the texture image to generate a modified texture image that may be relatively efficient to store and/or stream to a client due to enhanced compression of the modified texture image. | 03-10-2016 |