Patent application number | Description | Published |
20100069126 | GENERATING, AT LEAST IN PART, AND/OR RECEIVING, AT LEAST IN PART, AT LEAST ONE REQUEST - In an embodiment, an apparatus is provided that may include circuitry to generate, at least in part, and/or receive, at least in part, at least one request that at least one network node generate, at least in part, information. The information may be to permit selection, at least in part, of (1) at least one power consumption state of the at least one network node, and (2) at least one time period. The at least one time period may be to elapse, after receipt by at least one other network node of at least one packet, prior to requesting at least one change in the at least one power consumption state. The at least one packet may be to be transmitted to the at least one network node. Of course, many alternatives, modifications, and variations are possible without departing from this embodiment. | 03-18-2010 |
20100077243 | CONSERVING POWER IN A COMPUTER SYSTEM - A power management unit (PMU) may determine an optimal power saving state using a break-even period of a power saving state and an expected idle duration based on a first policy. The PMU may determine the optimal power saving state using a first break even period and actual idle duration based on a second policy. The break-even period may equal a minimum time a computer system should remain in a power saving state to compensate for the power consumed by the system to enter and exit that power saving state. The expected idle time duration is determined as an average of idle duration and a recent sample of idle duration. The actual idle duration is the difference of a first and second time point that represents entry and exit points to and from the power saving state. The PMU may transition the system to the optimal power saving state. | 03-25-2010 |
20100169683 | Platform and processor power management - The present invention relates to platform power management. | 07-01-2010 |
20100250982 | CONSERVING POWER IN A PLATFORM SUPPORTING NETWORK INTERFACES - A computer system may comprise a platform in which a processing block may be provisioned. The processing block may determine an optimal compression ratio such that the optimal compression ratio may cause a minimum of a total power to be consumed by the computer platform. The total power may comprise total compression power consumption and total transmission power consumption. The processing block may generate compressed frames from a plurality of frames generated by an application. The compressed frames may be generated by encoding the plurality of frames using the optimal compression ratio. The processing block may select a network interface from multiple network interfaces supported by the computer system to transmit the compressed frames. | 09-30-2010 |
20100329276 | Energy efficient network forwarding - Includes a method that determines utilization of a link and determines a weight for the link based on an energy metric for the determined utilization. The method also includes transmitting the determined weight to at least one node adjacent in the network to the network forwarding device, and determining forwarding based on the transmitted weight. | 12-30-2010 |
20110154073 | SYSTEMS AND METHODS FOR ENERGY EFFICIENT LOAD BALANCING AT SERVER CLUSTERS - Methods and systems to balance the load among a set of processing units, such as servers, in a manner that allows the servers periods of low power consumption. This allows energy efficient operation of the set of processing units. Moreover, the process is adaptable to variations in systemic response times, so that systemic response times may be improved when operational conditions so dictate. | 06-23-2011 |
20110154080 | METHOD AND APPARATUS FOR I/O DEVICES ASSISTED PLATFORM POWER MANAGEMENT - Embodiments of an apparatus, system and method are described for input/output (I/O) device assisted platform power management. An apparatus may comprise, for example, power management logic operative to receive idle duration information from one or more input/output (I/O) devices and to modify a power state for one or more components based on the idle information. Other embodiments are described and claimed. | 06-23-2011 |
20110199948 | MINIMIZING POWER CONSUMPTION IN A NETWORK DEVICE - A network interface device (NID) may determine whether the received data units of the computer system are to be compressed before transmitting the data units. The NID may determine the compression energy value consumed to compress the first K | 08-18-2011 |
20110222397 | PACKET BUFFERING BASED AT LEAST IN PART UPON PACKET RECEIPT TIME INTERVAL WEIGHTED MOVING AVERAGE - An embodiment may include circuitry to be comprised in a node. The node may be communicatively coupled to a network and may include a host processor to execute at least one communication protocol-related process. The circuitry may determine, at least in part, whether to buffer, at least in part, at least one packet received from the network based at least in part upon whether a weighted moving average exceeds a first threshold and is less than a second threshold. The average may be determined based at least in part upon one or more time intervals between one or more packet receptions from the network. Many alternatives, variations, and modifications are possible. | 09-15-2011 |
20110222537 | NETWORK CONTROLLER CIRCUITRY TO ISSUE AT LEAST ONE PORTION OF PACKET PAYLOAD TO DEVICE IN MANNER THAT BY-PASSES COMMUNICATION PROTOCOL STACK INVOLVEMENT - An embodiment may include network controller to be comprised in a first node. The node may be communicatively coupled to a network and may include a host processor to execute an operating system environment. The operating system environment may include, at least in part, a communication protocol stack and an application. The circuitry may receive, at least in part, a packet from the network. The packet may include, at least in part, a header and payload. At least one portion of the payload may be associated with the application. The circuitry may issue at least one portion of the header to the stack. The circuitry may issue the at least one portion of the payload to a destination device in a manner that by-passes involvement of the stack. The destination device may be specified, at least in part, by the application. Many alternatives, variations, and modifications are possible. | 09-15-2011 |
20120079482 | COORDINATING DEVICE AND APPLICATION BREAK EVENTS FOR PLATFORM POWER SAVING - Systems and methods of managing break events may provide for detecting a first break event from a first event source and detecting a second break event from a second event source. In one example, the event sources can include devices coupled to a platform as well as active applications on the platform. Issuance of the first and second break events to the platform can be coordinated based on at least in part runtime information associated with the platform. | 03-29-2012 |
20120131370 | CONSERVING POWER IN A COMPUTER SYSTEM - A power management unit (PMU) may determine an optimal power saving state using a break-even period of a power saving state and an expected idle duration based on a first policy. The PMU may determine the optimal power saving state using a first break even period and actual idle duration based on a second policy. The break-even period may equal a minimum time a computer system should remain in a power saving state to compensate for the power consumed by the system to enter and exit that power saving state. The expected idle time duration is determined as an average of idle duration and a recent sample of idle duration. The actual idle duration is the difference of a first and second time point that represents entry and exit points to and from the power saving state. The PMU may transition the system to the optimal power saving state. | 05-24-2012 |
20120166731 | COMPUTING PLATFORM POWER MANAGEMENT WITH ADAPTIVE CACHE FLUSH - In some embodiments, an adaptive break-even time, based on the load level of the cache, may be employed. | 06-28-2012 |
20120198030 | Storing Data Using A Direct Data Path Architecture To Reduce Energy Consumption And Improve Performance - In one embodiment, the present invention includes a method for obtaining file information regarding a file to be downloaded from a remote location to a computing device, creating at least one empty file in a destination storage based on the file information and communicating block information regarding the empty file to a network interface, and receiving a data packet of the file in the network interface and directly sending a payload of the data packet from the network interface to the destination storage according to the block information, while a host processor of the computing device is in a low power state. Other embodiments are described and claimed. | 08-02-2012 |
20120210032 | TECHNIQUES FOR MANAGING POWER CONSUMPTION STATE OF A PROCESSOR - Techniques are described for determining a temporary latency tolerance report (tLTR) value. A processing unit has to respond to a device interrupt within a duration specified by tLTR to ensure no incoming data is lost due to device buffer overflow. The tLTR value can be used to prevent the processing unit from entering too deep a sleep state when a device driver anticipates multiple sequential interrupts for a transaction. | 08-16-2012 |
20130077483 | MINIMIZING POWER CONSUMPTION IN A NETWORK DEVICE - A network interface device (NID) may determine whether the received data units of the computer system are to be compressed before transmitting the data units. The NID may determine the compression energy value consumed to compress the first K | 03-28-2013 |
20130138985 | POWER MANAGEMENT USING RELATIVE ENERGY BREAK-EVEN TIME - Systems and methods may provide for determining an absolute energy break-even time for a first low power state with respect to a current state of a system. A relative energy break-even time may also be determined for the first low power state with respect to a second low power state based on at least in part the absolute energy break-even time. In addition, an operating state may be selected for the system based on at least in part the relative energy break-even time. | 05-30-2013 |
20130170349 | GENERATING, AT LEAST IN PART, AND/OR RECEIVING, AT LEAST IN PART, AT LEAST ONE REQUEST - In an embodiment, an apparatus is provided that may include circuitry to generate, at least in part, and/or receive, at least in part, at least one request that at least one network node generate, at least in part, information. The information may be to permit selection, at least in part, of (1) at least one power consumption state of the at least one network node, and (2) at least one time period. The at least one time period may be to elapse, after receipt by at least one other network node of at least one packet, prior to requesting at least one change in the at least one power consumption state. The at least one packet may be to be transmitted to the at least one network node. Of course, many alternatives, modifications, and variations are possible without departing from this embodiment. | 07-04-2013 |
20130275789 | CREDIT BASED POWER MANAGEMENT - An embodiment may include circuitry to determine whether to issue at least one credit to at least one sender of at least one packet. The credit(s) may be to grant permission to the at least one sender to issue the at least one packet to at least one receiver of the at least one packet. The determination of whether to issue the credit(s) may be based, at least in part, upon whether a time in which the at least one receiver is in a relatively lower power state prior to issuance of the credit(s) is at least sufficient to provide at least a predetermined amount of reduction in power consumption. The relatively lower power state may be relative to a relatively higher power state of the at least one receiver that prevails at the issuance of the credit(s). Additionally or alternatively, the circuitry may be to receive such credit(s). | 10-17-2013 |
20130290546 | MECHANISM FOR EMPLOYING AND FACILITATING DYNAMIC AND REMOTE MEMORY COLLABORATION AT COMPUTING DEVICES - A mechanism is described for facilitating dynamic and remote memory collaboration at computing devices according to one embodiment of the invention. A method of embodiments of the invention includes dynamically classifying a computing device of a plurality of computing devices as a memory server, where the plurality of computing devices are coupled to each other over a network. The method may further include offering, by the memory server, of memory to be used by one or more of the plurality of computing devices classified as one or more memory clients, and remotely granting, by the memory server, of the memory to the one or more memory clients. | 10-31-2013 |
20130311808 | PLATFORM AND PROCESSOR POWER MANAGEMENT - The present invention relates to platform power management. | 11-21-2013 |
20140006713 | Cache Collaboration in Tiled Processor Systems | 01-02-2014 |
20140006819 | JOINT OPTIMIZATION OF PROCESSOR FREQUENCIES AND SYSTEM SLEEP STATES | 01-02-2014 |
20140006824 | USING DEVICE IDLE DURATION INFORMATION TO OPTIMIZE ENERGY EFFICIENCY | 01-02-2014 |
20140007114 | MONITORING ACCESSES OF A THREAD TO MULTIPLE MEMORY CONTROLLERS AND SELECTING A THREAD PROCESSOR FOR THE THREAD BASED ON THE MONITORING | 01-02-2014 |
20140029617 | PACKET PROCESSING APPROACH TO IMPROVE PERFORMANCE AND ENERGY EFFICIENCY FOR SOFTWARE ROUTERS - Methods, apparatus and systems for improved performance and energy efficiency of software-based routers. A software router running on a host computer system employing multiple Network Interface Controllers (NICs) maintains a routing table wherein packet flows are classified as managed flows (MFs) under which packets are received at and forwarded from the same NIC and unmanaged flows UFs under which packets are received at and forwarded from different NICs. Forwarding table data is employed by a NIC to facilitate packet identification and flow classification operations under which the NIC determines whether a received packet is an MF, UF, or an unclassified flow. Under various schemes, packet forwarding for MFs is handled by the software router architecture such that either only the packet header is copied into memory in the host or the entire packet forwarding is handled by the NIC. | 01-30-2014 |
20140071866 | OPTIMIZING SEMI-ACTIVE WORKLOADS - Devices and methods for optimizing semi-active workloads are described herein. A network interface device may be configured to offload data packet acknowledgment responsibilities of a host platform by transmitting, to the sender of the packets, acknowledgements of packets received throughout a time duration. Upon completion of the time duration, the network interface device may trigger the host platform to perform batch processing of the data packets received during the time duration. | 03-13-2014 |
20140092740 | ADAPTIVE PACKET DEFLECTION TO ACHIEVE FAIR, LOW-COST, AND/OR ENERGY-EFFICIENT QUALITY OF SERVICE IN NETWORK ON CHIP DEVICES - Methods and apparatus for provision of adaptive packet deflection to achieve fair, low-cost, and/or energy-efficient Quality of Service (QoS) in Network-on-Chip (NoC) devices are described. In some embodiments, it is determined whether a target port of a packet has reached a threshold utilization value and the packet is routed to an alternate port in response to a deflection probability value that is to be determined based on a utilization value of the target port and a priority level value of the packet. Other embodiments are also claimed and/or disclosed. | 04-03-2014 |
20140095695 | CLOUD AWARE COMPUTING DISTRIBUTION TO IMPROVE PERFORMANCE AND ENERGY FOR MOBILE DEVICES - An intelligent cloud aware computing distribution architecture for a device. A network conditions monitor is to observe and identify decision impact factors of tasks in a runtime environment. A dynamic profiler, coupled to the network conditions monitor, is to receive runtime information regarding the decision impact factors identified by the network conditions monitor and produce a profile based on the decision impact factors. Runtime offload decision making logic is to process the profile produced by the dynamic profiler based on the received decision impact factors according a predetermined policy and to determine final offloading decisions based on the predetermined policy and the processed decision impact factors. The runtime offload decision making logic is to provide the final offloading decisions to the applications on the device for executing the tasks locally or remotely based on the determined final offloading decision. | 04-03-2014 |
20140101674 | COORDINATING DEVICE AND APPLICATION BREAK EVENTS FOR PLATFORM POWER SAVING - Systems and methods of managing break events may provide for detecting a first break event from a first event source and detecting a second break event from a second event source. In one example, the event sources can include devices coupled to a platform as well as active applications on the platform. Issuance of the first and second break events to the platform can be coordinated based on at least in part runtime information associated with the platform. | 04-10-2014 |
20140149766 | ROUTER PARKING IN POWER-EFFICIENT INTERCONNECT ARCHITECTURES - A method and apparatus for selectively parking routers used for routing traffic in mesh interconnects. Various router parking (RP) algorithms are disclosed, including an aggressive RP algorithm where a minimum number of routers are kept active to ensure adequate network connectivity between active nodes and/or intercommunicating nodes, leading to a maximum reduction in static power consumption, and a conservative RP algorithm that favors network latency considerations over static power consumption while also reducing power. An adaptive RP algorithm is also disclosed that implements aspects of the aggressive and conservative RP algorithms to balance power consumption and latency considerations in response to ongoing node utilization and associated traffic. The techniques may be implemented in internal network structures, such as for single chip computers, as well as external network structures, such as computing clusters and massively parallel computer architectures. Performance modeling has demonstrated substantial power reduction may be obtained using the router parking techniques while maintaining Quality of Service performance objectives. | 05-29-2014 |
20140169281 | CHANNEL AWARE JOB SCHEDULING - Methods and systems may provide for determining quality of service (QoS) information for a job associated with an application, and determining a condition prediction for a wireless channel of a mobile platform. Additionally, the job may be scheduled for communication over the wireless channel based at least in part on the QoS information and the condition prediction. In one example, scheduling the job includes imposing a delay in the communication if the condition prediction indicates that a throughput of the wireless channel is below a threshold and the delay complies with a latency constraint of the QoS information. | 06-19-2014 |
20140173206 | Power Gating A Portion Of A Cache Memory - In an embodiment, a processor includes multiple tiles, each including a core and a tile cache hierarchy. This tile cache hierarchy includes a first level cache, a mid-level cache (MLC) and a last level cache (LLC), and each of these caches is private to the tile. A controller coupled to the tiles includes a cache power control logic to receive utilization information regarding the core and the tile cache hierarchy of a tile and to cause the LLC of the tile to be independently power gated, based at least in part on this information. Other embodiments are described and claimed. | 06-19-2014 |
20140173207 | Power Gating A Portion Of A Cache Memory - In an embodiment, a processor includes multiple tiles, each including a core and a tile cache hierarchy. This tile cache hierarchy includes a first level cache, a mid-level cache (MLC) and a last level cache (LLC), and each of these caches is private to the tile. A controller coupled to the tiles includes a cache power control logic to receive utilization information regarding the core and the tile cache hierarchy of a tile and to cause the LLC of the tile to be independently power gated, based at least in part on this information. Other embodiments are described and claimed. | 06-19-2014 |
20140173221 | CACHE MANAGEMENT - The present disclosure provides techniques for cache management. A data block may be received from an IO interface. After receiving the data block, the occupancy level of a cache memory may be determined. The data block may be directed to a main memory if the occupancy level exceeds a threshold. The data block may be directed to a cache memory if the occupancy level is below a threshold. | 06-19-2014 |
20140173314 | ADAPTIVELY DISABLING AND ENABLING SLEEP STATES FOR POWER AND PERFORMANCE - Methods and systems may provide for determining whether a runtime disablement condition is met with respect to a sleep state and disabling the sleep state if the runtime disablement condition is met. Additionally, the sleep state may be enabled if a runtime reinstatement condition is met. In one example, determining whether the runtime disablement condition is met includes determining a false entry rate for the sleep state, and comparing the false entry rate to an energy-based threshold, wherein the sleep state is disabled if the false entry rate exceeds the energy-based threshold. | 06-19-2014 |
20140173636 | USING PLATFORM IDLE DURATION INFORMATION TO NOTIFY PLATFORM DEVICES OF UPCOMING ACTIVE PERIODS - Methods and systems may provide for determining a next active window for a platform and notifying one or more of a plurality of devices of the platform of the next active window being determined. Additionally, one or more of the plurality of devices may be notified of an onset of the next active window. In one example, a pre-warm message is issued to notify one or more of the plurality of devices of the next active window being determined. | 06-19-2014 |
20140177494 | CLOUD-AWARE COLLABORATIVE MOBILE PLATFORM POWER MANAGEMENT USING MOBILE SENSORS - Methods and systems may provide for accumulating sensor-based information from a plurality of mobile devices and identifying context information in the sensor-based information, wherein the context information is relevant to a particular mobile device in the plurality of mobile devices. Additionally, an operational recommendation may be sent to the particular mobile device based on the context information. In one example, the operational recommendation identifies one or more of a power management action, a performance management action, a wireless interface selection and a wireless access point selection. | 06-26-2014 |
20140187264 | SITUATION AWARE SENSOR POWER MANAGEMENT - Methods and systems may provide for determining a status of a mobile platform, wherein the status indicates whether the mobile platform is stationary, and adapting a detection schedule of one or more location sensors on the mobile platform based at least in part on whether the mobile platform is stationary. Additionally, one or more location updates may be generated based at least in part on information from the one or more location sensors. In one example, a location request is received, wherein the detection schedule is adapted further based on quality of service (QoS) information associated with the location request, and wherein the one or more location updates are generated in response to the location request. | 07-03-2014 |
20140195833 | ADAPTIVE LOW-POWER LINK-STATE ENTRY POLICY FOR ACTIVE INTERCONNECT LINK POWER MANAGEMENT - Methods and apparatus for implementing active interconnect link power management using an adaptive low-power link-state entry policy. The power state of an interconnect link or fabric is changed in response to applicable conditions determined by low-power link-state entry policy logic in view of runtime traffic on the interconnect link or fabric. The low-power link-state policy logic may be configured to include consideration of operating system input and Quality of Service (QoS) requirements for applications and devices employing the link or fabric, and device latency tolerance requirements. | 07-10-2014 |
20140289546 | POWER MANAGEMENT USING REWARD-BASED SLEEP STATE SELECTION - Systems and methods may provide for conducting a reward determination for a plurality of sleep states to obtain a plurality of reward determinations with respect to a device. In addition, a sleep state may be selected for the device from the plurality of sleep states based at least in part on the plurality of reward determinations. In one example, false entry and missed opportunity probabilities may be determined for stochastic interrupts, wherein the reward determination is conducted based at least in part on the false entry and missed opportunity probabilities. | 09-25-2014 |
20140301199 | METHODS, SYSTEMS, AND COMPUTER PROGRAM PRODUCTS FOR PROCESSING A PACKET - A system for processing a packet may include, for each of a network interface controller and a central processing unit, a measurement of the processing time, a determination of the amount of energy consumed to process a unit of information in the packet, and a measurement of the load. A user may provide the system with signals to perform networking processes for the packet in a manner to reduce the processing time of the system or in a manner to reduce the amount of energy consumed by the system for processing the packet. A portion of the system may receive at least one of the measurements, determinations, and signals and may cause one of the network interface controller and the central processing unit to perform networking processes for the packet. The networking processes may include establishing a connection to a network. | 10-09-2014 |
20140365797 | MINIMIZING POWER CONSUMPTION IN A NETWORK DEVICE - A network interface device (NID) may determine whether the received data units of the computer system are to be compressed before transmitting the data units. The NID may determine the compression energy value consumed to compress the first K1 data units and a second transmission energy value to transmit the compressed first K1 data units. The NID may then estimate a first transmission energy value that may be consumed by the NID to transmit uncompressed first K1 data units using the second transmission energy value. The NID may then use the first and second transmission energy value and the compression energy value to determine if the remaining (N−K1) data units of the first data stream. | 12-11-2014 |
20150067259 | MANAGING SHARED CACHE BY MULTI-CORE PROCESSOR - Systems and methods for managing shared cache by multi-core processor. An example processing system comprises: a plurality of processing cores, each processing core communicatively coupled to a last level cache (LLC) slice; and a cache control logic coupled to the plurality of processing cores, the cache control logic configured to perform one of: making an LLC slice of an inactive processing core available to an active processing core or power gating the LLC slice, based on estimating cache requirements by active processing cores. | 03-05-2015 |