Patent application number | Description | Published |
20110208591 | Forecasting Online Advertising Inventory of Day Parting Queries - Disclosed is a system to forecast a supply of online advertising demand contracts having day parting targets. The system may receive an online advertising demand contract having a flight duration and a day parting target. The received online advertising demand contract may be processed by converting the day parting target into an hour-of-week day parting target vector. If a first day of the flight duration would generate a conflict in view of the hour-of-week day parting target vector, then the system may generate a new first day contract for the first day of the flight duration. | 08-25-2011 |
20120004957 | FORECASTING SUPPLY FOR ADVERTISEMENTS ACCORDING TO A NON-PARAMETRIC SUPPLY MODEL - Methods, systems, and apparatuses for generating supply forecasts for advertisement impressions are provided. Supply forecasts for future advertisement inventory are enabled to be determined that match attribute values received from a requester in a query or “contract.” Such supply forecasts may be determined based on cumulative or non-cumulative supply forecast values, and may be determined based on time series of supply forecast values for individual impressions or for blocks of impressions. | 01-05-2012 |
20120221591 | System for Processing Complex Queries - A method of processing a supply query is described. The method includes receiving a first supply query comprising a first set of attributes. A size of the first set of attributes is compared to a threshold value. A second supply query comprising a second set of attributes is generated when the comparison indicates that the size of the first set of attributes is greater than the threshold value. | 08-30-2012 |
20120239468 | HIGH-PERFORMANCE SUPPLY FORECASTING USING OVERRIDE RULES IN DISPLAY ADVERTISING SYSTEMS - A method, system and computer readable medium to adjust impression supply sampling weights in a display advertising environment while providing high-performance supply forecasting using trends adjustments and override rules. A forecasting system initiates the method upon submitting a campaign query (with a campaign query predicate to be matched with impression supply predicates) and a campaign query time period (specifying time-wise aspects of an advertising campaign). Upon receiving initial campaign query results (e.g. impression supply datastructure with base weights), the method proceeds to identify applicable weight adjustment rules for applying to the impression supply node, then, after accounting for trends or other time-wise aspects of the weight adjustment rule, applying the adjustment rule to the impression supply base weight of an impression supply node, resulting in at least one adjusted trend weight. Given adjusted trend weights, the forecasting system can more accurately predict future supply and pricing characteristics of the campaign. | 09-20-2012 |
Patent application number | Description | Published |
20110022817 | Mapping Processing Logic Having Data-Parallel Threads Across Processors - A method for executing a plurality of data-parallel threads of a processing logic on a processor core includes grouping the plurality of data-parallel threads into one or more workgroups, associating a first workgroup from the one or more workgroups with an operating system thread on the processor core, and configuring threads from the first workgroup as user-level threads within the operating system thread. In an example, a method enables the execution of GPU-kernels that has been previously configured for a GPU, to execute on a CPU such as a multi-core CPU. The mapping of the numerous data-parallel threads to the CPU is done in such a manner as to reduce the number of costly operating system threads instantiated on the CPU, and to enable efficient debugging. | 01-27-2011 |
20130160016 | Allocating Compute Kernels to Processors in a Heterogeneous System - A system and method embodiments for optimally allocating compute kernels to different types of processors, such as CPUs and GPUs, in a heterogeneous computer system are disclosed. These include comparing a kernel profile of a compute kernel to respective processor profiles of a plurality of processors in a heterogeneous computer system, selecting at least one processor from the plurality of processors based upon the comparing, and scheduling the compute kernel for execution in the selected at least one processor. | 06-20-2013 |
20140137104 | Cooperative Application Workload Scheduling for a Consolidated Virtual Environment - Application resource scheduler module is provided to achieve cooperative application workload scheduling for a consolidated virtual environment. The application resource scheduler aids an application workload scheduler that is part of a distributed computing application, such as Hadoop, to achieve a specified relative priority of the application workload virtual machines to other virtual machines in the virtual environment. The application resource scheduler assists in achieving cooperative workload scheduling by revising the amount of resources that the application workload scheduler sees as available and by setting resource controls for the virtual machines of the distributed computing application to influence the resources the virtual machines receive from the underlying consolidated virtual environment. | 05-15-2014 |
20150058843 | VIRTUAL HADOOP MANAGER - A distributed computing application is described that provides a highly elastic and multi-tenant platform for Hadoop applications and other workloads running in a virtualized environment. Multiple instances of a distributed computing framework, such as Hadoop, may be executed concurrently. A centralized manager detects when contention for computing resources, such as memory and CPU, causes tasks to run slower on VMs executing on a given host, and scales up or scales down a cluster based on the detected resource contention. | 02-26-2015 |
20150120791 | MULTI-TENANT PRODUCTION AND TEST DEPLOYMENTS OF HADOOP - A distributed computing application is described that provides a highly elastic and multi-tenant platform for Hadoop applications and other workloads running in a virtualized environment. Production, test, and development deployments of a Hadoop application may be executed using multiple compute clusters and a shared instance of a distributed filesystem, or in other cases, multiple instances of the distributed filesystem. Data nodes executing as virtual machines (VMs) for test and development deployments can be linked clones of data nodes executing as VMs for a production deployment to reduce duplicated data and provide a shared storage space. | 04-30-2015 |
20150120928 | CONTAINER VIRTUAL MACHINES FOR HADOOP - A distributed computing application is described that provides a highly elastic and multi-tenant platform for Hadoop applications and other workloads running in a virtualized environment. Data and compute nodes are separated into different virtual machines (VM). Compute VMs are used to launch containers from different tenants. Compute VMs are organized in pools of hot spare VMs that are immediately available for launching a container and executing a task, and pools of cold spare VMs. Each compute VM may include a mounted network filesystem provided by a node manager to share intermediate outputs across VMs executing on the same host. | 04-30-2015 |
20150121371 | MULTI-TENANT DISTRIBUTED COMPUTING AND DATABASE - A distributed computing application is described that provides a highly elastic and multi-tenant platform for Hadoop applications and other workloads running in a virtualized environment. Deployments of a distributed computing application, such as Hadoop, may be executed concurrently with a distributed database application, such as HBase, using a shared instance of a distributed filesystem, or in other cases, multiple instances of the distributed filesystem. Computing resources allocated to region server nodes executing as VMs may be isolated from compute VMs of the distributed computing application, as well as from data nodes executing as VMs of the distributed filesystem. | 04-30-2015 |
20150178107 | Elastic Compute Fabric Using Virtual Machine Templates - Embodiments include an infrastructure shared among cloud services that supports fast provisioning of virtual machines (VMs). A set of powered-on parent VM templates and a set of powered-off child VMs are maintained by the infrastructure in a hierarchy. The child VMs are instantiated from the parent VM templates, and pre-registered to a cloud operating system in some embodiments. In response to requests from the cloud services for the child VMs, where the requests specify child VM configurations, child VMs from the set of powered-off child VMs are selected and customized based on the child VM configurations, and then deployed for use by the cloud services. In some embodiments, the fast provisioning of VMs is supported by forking operations in the infrastructure. | 06-25-2015 |
20150178108 | Fast Instantiation of Virtual Machines - Embodiments support instant forking of virtual machines (VMs) and state customization. Virtual device state and persistent storage of a child VM are defined based on virtual device state and persistent storage of parent VMs. After forking, a state of the child VM is customized based on configuration data. Customizing the state includes configuring one or more identities of the child VM, before bootup completes on the child VM. | 06-25-2015 |
20150178110 | State Customization of Forked Virtual Machines - Embodiments support instant forking of virtual machines (VMs) and state customization. A computing device initiates execution of a first group of services (e.g., identity-independent) in a first VM. A second VM is instantiated from the first VM. The second VM shares memory and storage with the first VM. The computing device customizes the second VM based on configuration data associated with the second VM. A second group of services (e.g., identity-dependent) starts executing on the second VM after configuring the identity of the second VM. Customizing the second VM includes configuring one or more identities of the second VM. In some embodiments, a domain identity is selected from a pool of previously-created identities and applied to the second VM, before bootup completes on the second VM. | 06-25-2015 |
Patent application number | Description | Published |
20130132735 | APPARATUS AND METHOD FOR HARDWARE-BASED SECURE DATA PROCESSING USING BUFFER MEMORY ADDRESS RANGE RULES - Disclosed is a processor for processing data from a buffer memory. The processor, implemented in hardware, may allow writing of output data, processed based on input data from at least one secure location associated with a secure address range of the buffer memory, to one or more secure locations associated with the secure address range. Further, the processor may block writing of output data, processed based on input data from at least one secure location associated with the secure address range, to one or more insecure locations associated with an insecure address range of the buffer memory. | 05-23-2013 |
20150106630 | APPARATUS AND METHOD FOR HARDWARE-BASED SECURE DATA PROCESSING USING BUFFER MEMORY ADDRESS RANGE RULES - Disclosed is a processor for processing data from a buffer memory. The processor, implemented in hardware, may allow writing of output data, processed based on input data from at least one secure location associated with a secure address range of the buffer memory, to one or more secure locations associated with the secure address range. Further, the processor may block writing of output data, processed based on input data from at least one secure location associated with the secure address range, to one or more insecure locations associated with an insecure address range of the buffer memory. | 04-16-2015 |
Patent application number | Description | Published |
20110208636 | MATCHING PARTIES TO A TRANSACTION FOR AN AGRICULTURAL COMMODITY - In general, in one aspect, the invention relates to a method for matching parties to a transaction of an agricultural commodity. The method involves receiving buyer criteria (e.g., purchase quantity, buyer price, transaction location at which to buy the agricultural commodity) sent from a mobile device used by a primary and secondary buyer. The method further involves receiving seller criteria sent from a seller mobile device. The method further involves matching, within a predetermined period of time after receiving the buyer criteria and the seller criteria, the primary and secondary buyer with the seller based on determining that a yield is sufficient to meet the purchase quantity and a transaction location at which to buy the agricultural commodity falls within a number of locations common the primary/secondary buyers and the seller. The method further involves sending the buyer criteria to the seller mobile device using short message service (SMS) format. | 08-25-2011 |
20160125551 | METHOD AND SYSTEM FOR PUBLIC AND PRIVATE TEMPLATE SHARING - A method for sharing templates for use with a financial management application (FMA) includes receiving, from each of multiple financial professionals, a company profile including a chart of accounts for a business operating within a commercial industry and at a geographic location, generating profile templates including, for each profile, a template including the commercial industry, the geographic location, and a redacted chart of accounts of the profile, receiving, from each of the financial professionals, a request to publicly share the template corresponding to the profile, receiving, from a small business representative, a profile creation request specifying a first commercial industry and a first geographic location, identifying a first template, and providing, in response to the profile creation request, the first template to the representative, where the representative uses the first template to create a first profile for the business. | 05-05-2016 |