Patent application number | Description | Published |
20140156910 | Automated Space Management for Server Flash Cache - Techniques for automatically allocating space in a flash storage-based cache are provided. In one embodiment, a computer system collects I/O trace logs for a plurality of virtual machines or a plurality of virtual disks and determines cache utility models for the plurality of virtual machines or the plurality of virtual disks based on the I/O trace logs. The cache utility model for each virtual machine or each virtual disk defines an expected utility of allocating space in the flash storage-based cache to the virtual machine or the virtual disk over a range of different cache allocation sizes. The computer system then calculates target cache allocation sizes for the plurality of virtual machines or the plurality of virtual disks based on the cache utility models and allocates space in the flash storage-based cache based on the target cache allocation sizes. | 06-05-2014 |
20140208001 | Techniques for Achieving Crash Consistency when Performing Write-Behind Caching Using a Flash Storage-Based Cache - Techniques for achieving crash consistency when performing write-behind caching using a flash storage-based cache are provided. In one embodiment, a computer system receives from a virtual machine a write request that includes data to be written to a virtual disk and caches the data in a flash storage-based cache. The computer system further logs a transaction entry for the write request in the flash storage-based cache, where the transaction entry includes information usable for flushing the data from the flash storage-based cache to a storage device storing the virtual disk. The computer system then communicates an acknowledgment to the VM indicating that the write request has been successfully processed. | 07-24-2014 |
20140244898 | I/O Hint Framework for Server Flash Cache - An I/O hint framework is provided. In one embodiment, a computer system can receive an I/O command originating from a virtual machine (VM), where the I/O command identifies a data block of a virtual disk. The computer system can further extract hint metadata from the I/O command, where the hint metadata includes one or more characteristics of the data block that are relevant for determining how to cache the data block in a flash storage-based cache. The computer system can then make the hint metadata available to a caching module configured to manage the flash storage-based cache. | 08-28-2014 |
20140245298 | Adaptive Task Scheduling of Hadoop in a Virtualized Environment - A control module is introduced to communicate with an application workload scheduler of a distributed computing application, such as a Job Tracker node of a Hadoop cluster, and with the virtualized computing environment underlying the application. The control module periodically queries for resource consumption data, such as CPU utilization, and uses the data to calculate how MapReduce task slots should be allocated on each task node of the Hadoop cluster. The control module passes the task slot allocation to the application workload scheduler, which honors the allocation by adjusting task assignments to task nodes accordingly. The task nodes may also activate and deactivate task slots according to the changed slot allocation. As a result, the distributed computing application is able to scale up and down when other workloads sharing the virtualized computing environment change. | 08-28-2014 |
20150067262 | THREAD CACHE ALLOCATION - Systems and techniques are described for thread cache allocation. A described technique includes monitoring input and output accesses for a plurality of threads executing on a computing device that includes a cache comprising a quantity of memory blocks, determining a respective reuse intensity for each of the threads, determining a respective read ratio for each of the threads, determining a respective quantity of memory blocks for each of the partitions by optimizing a combination of cache utilities, each cache utility being based on the respective reuse intensity, the respective read ratio, and a respective hit ratio for a particular partition, and resizing one or more of the partitions to be equal to the respective quantity of the memory blocks for the partition. | 03-05-2015 |
Patent application number | Description | Published |
20080244073 | METHOD AND SYSTEM FOR STORAGE-AWARE FLOW RESOURCE MANAGEMENT - A method and structure for managing resources in a storage system that is based on server resource usage is disclosed. The method allocates a command window to each of the servers accessing the resources, achieves an equilibrium state that balances the size of each server's command window and the amount of free resources, and maintains the equilibrium state by continually adjusting the size of each command window. The command window defines a limit on the number of outstanding storage requests that each server may maintain. The invention initially allocates a minimum sized command window. | 10-02-2008 |
20090307166 | METHOD AND SYSTEM FOR AUTOMATED INTEGRATED SERVER-NETWORK-STORAGE DISASTER RECOVERY PLANNING - An automated disaster recovery (DR) planning system for a computing environment is provided. A discovery module discovers servers, networks, and storage devices in a computing environment. An expert knowledge base module captures best practices in planning, and capabilities, interoperability, limitation and boundary values for different DR technologies. A match-making module determines multiple DR plans as combinations of one or more replication technologies that can be used to satisfy DR requirements. And, an optimizer configured for assessing a feasible DR plan from said multiple DR plans, to deploy for DR planning of a primary computing environment. | 12-10-2009 |
20100205133 | SYSTEM AND METHOD FOR PROVIDING AUTONOMIC MANAGEMENT OF A NETWORKED SYSTEM FOR USING AN ACTION-CENTRIC APPROACH - According to the present invention, there is provided a SAN management system to provide autonomic management of a storage system using an action-centric approach. The SAN management system includes a policy specification logic block to maintain a policy specification associated with the managed system. In addition, the SAN management system includes a reasoning logic block to provide for the determining of action rules using a combination of logic and information obtained from the policy specification. Also, the SAN management system includes a learning logic block to couple the policy specification logic block with the reasoning logic block to improve an understanding of a managed system. The learning is continuous and provides for autonomic evolvement of the system in which reliance on manual input from a user is lessened. | 08-12-2010 |
20110010343 | OPTIMIZATION AND STAGING METHOD AND SYSTEM - An optimization method and system. The method includes receiving by a computing system a data footprint associated with data and a human resource model. The data footprint comprises a primary data section, a secondary data section, and an archive data section. A plurality of data storage strategies are associated with the primary data section, said secondary data section, and said archive data section. The plurality of data storage strategies are compared to each other. A data staging orchestrator software module is executed. The computing system determines based on results of executing the data staging orchestrator software module, an optimal migration time, an optimal migration speed, and an optimal migration cost for managing storage for portions of the data. The computing system executes a risk modulation software module and determines a risk associated with the managing. | 01-13-2011 |
20110167424 | INTEGRATED DIFFERENTIATED OPERATION THROTTLING - A method and system for throttling a plurality of operations of a plurality of applications that share a plurality of resources. A difference between observed and predicted workloads is computed. If the difference does not exceed a threshold, a multi-strategy finder operates in normal mode and applies a recursive greedy pruning process with a look-back and look-forward optimization to select actions for a final schedule of actions that improve the utility of a data storage system. If the difference exceeds the threshold, the multi-strategy finder operates in unexpected mode and applies a defensive action selection process to select actions for the final schedule. The selected actions are performed according to the final schedule and include throttling of a CPU, network, and/or storage. | 07-07-2011 |
20110208622 | DATA CENTER POWER COST ACCOUNTING SYSTEM - A data center power cost accounting system uses server and storage and cooling power consumption models and device maps, together with runtime application maps, to estimate the equipment power consumption and cooling power consumption of individual applications. An approximation of the cooling cost over a period of time, for any given application, can be pieced together by adding up the equipment utilized by the application and applying the cooling estimates obtained from computational fluid dynamics (CFD) simulations. The cooling estimates can further account for changes or variability in resource usage over time since the cooling estimates are based directly on utilization. The per application power consumption costs are obtained without having to install or depend on power measurement instruments or other hardware in the datacenters. | 08-25-2011 |
20110296237 | SELECTING A DATA RESTORE POINT WITH AN OPTIMAL RECOVERY TIME AND RECOVERY POINT - Embodiments of the invention relate to selecting a data restore point with an optimal recovery time and recovery point. An exemplary embodiment includes generating a problem search criterion for an entity with corrupted data. Dependencies relied on by the entity to function are determined. At least one event signature match is found that comprises information for an event being logged in a event log, and is associated with the dependencies. At least one data restore point created prior to an occurrence of a particular event in the at least one event signature match is selected. The particular event having caused the data to be corrupted. The at least one data restore point is selected to restore data to a storage system with the corrupted data. | 12-01-2011 |
20120191661 | OPTIMIZATION AND STAGING - An optimization method and system. The method includes receiving by a computing system a data footprint associated with data and a human resource model. The data footprint comprises a primary data section, a secondary data section, and an archive data section. A plurality of data storage strategies are associated with the primary data section, said secondary data section, and said archive data section. The plurality of data storage strategies are compared to each other. A data staging orchestrator software module is executed. The computing system determines based on results of executing the data staging orchestrator software module, an optimal migration time, an optimal migration speed, and an optimal migration cost for managing storage for portions of the data. The computing system executes a risk modulation software module and determines a risk associated with the managing. | 07-26-2012 |