Patent application number | Description | Published |
20090132582 | PROCESSOR-SERVER HYBRID SYSTEM FOR PROCESSING DATA - The present invention relates to a processor-server hybrid system that comprises (among other things) a set (one or more of servers (e.g., mainframes) and a set of front-end application optimized processors. Moreover, implementations of the invention provide a server and processor hybrid system and method for distributing and managing the execution of applications at a fine-grained level via an I/O-connected hybrid system. This method allows one system to be used to manage and control the system functions, and one or more other systems to co-processor. | 05-21-2009 |
20090132638 | SERVER-PROCESSOR HYBRID SYSTEM FOR PROCESSING DATA - The present invention relates to a server-processor hybrid system that comprises (among other things) a set (one or more) of front-end servers (e.g., mainframes) and a set of back-end application optimized processors. Moreover, implementations of the invention provide a server and processor hybrid system and method for distributing and managing the execution of applications at a fine-grained level via an I/O-connected hybrid system. This method allows one system to be used to manage and control the system functions, and one or more other systems to co-processor. | 05-21-2009 |
20090150555 | MEMORY TO MEMORY COMMUNICATION AND STORAGE FOR HYBRID SYSTEMS - The present invention relates to memory to memory communication and storage for hybrid systems. Under the present invention, a data stream is received on a first computing device of a hybrid system. An attempt is made to store the data stream on the first computing device up to a per stream limit and a total storage limit of the first computing device. It is then determined whether to store at least a portion of the data stream on a second computing device of the hybrid system that is in communication with the first computing device. This decision is based on the per stream limit and the total storage limit of the first computing device as well as a per stream limit and a total storage limit of the second computing device. Thereafter, the at least a portion of the data stream and a control signal are communicated to the second computing device for storage. | 06-11-2009 |
20090150556 | MEMORY TO STORAGE COMMUNICATION FOR HYBRID SYSTEMS - Under the present invention, a hybrid system having multiple computing devices and storage devices is provided. The “multiple computing devices” typically include at least one server and at least one processor, both of which include local memory. Thus, the hybrid system will typically have at least two different types of computing devices. The “multiple storage devices” are typically implemented within a storage area network, and include at least one staging storage device and at least one processed data storage device. These devices will be utilized to store incoming data streams in the event that either computing device lacks sufficient space and/or sufficient credits for transmission to another computing device. | 06-11-2009 |
20090157961 | TWO-SIDED, DYNAMIC CACHE INJECTION CONTROL - A method, system, and computer program product for two-sided, dynamic cache injection control are provided. An I/O adapter generates an I/O transaction in response to receiving a request for the transaction. The transaction includes an ID field and a requested address. The adapter looks up the address in a cache translation table stored thereon, which includes mappings between addresses and corresponding address space identifiers (ASIDs). The adapter enters an ASID in the ID field when the requested address is present in the cache translation table. IDs corresponding to device identifiers, address ranges and pattern strings may also be entered. The adapter sends the transaction to one of an I/O hub and system chipset, which in turn, looks up the ASID in a table stored thereon and injects the requested address and corresponding data in a processor complex when the ASID is present in the table, indicating that the address space corresponding to the ASID is actively running on a processor in the complex. The ASIDs are dynamically determined and set in the adapter during execution of an application in the processor complex. | 06-18-2009 |
20090157962 | CACHE INJECTION USING CLUSTERING - A method and system for cache injection using clustering are provided. The method includes receiving an input/output (I/O) transaction at an input/output device that includes a system chipset or input/output (I/O) hub. The I/O transaction includes an address. The method also includes looking up the address in a cache block indirection table. The cache block indirection table includes fields and entries for addresses and cluster identifiers (IDs). In response to a match resulting from the lookup, the method includes multicasting an injection operation to processor units identified by the cluster ID. | 06-18-2009 |
20090157966 | CACHE INJECTION USING SPECULATION - A method, system, and computer program product for cache injection using speculation are provided. The method includes creating a cache line indirection table at an input/output (I/O) hub, which includes fields and entries for addresses, processor ID, and cache type and includes cache level line limit fields. The method also includes setting cache line limits to the CLL fields and receiving a stream of contiguous addresses at the table. For each address in the stream, the method includes: looking up the address in the table; if the address is present in the table, inject the cache line corresponding to the address in the processor complex; if the address is not present in the table, search limit values from the lowest level cache to the highest level cache; and inject addresses not present in the table to the cache hierarchy of the processor last injected from the contiguous address stream. | 06-18-2009 |
20090157977 | DATA TRANSFER TO MEMORY OVER AN INPUT/OUTPUT (I/O) INTERCONNECT - A method, system, and computer program product for data transfer to memory over an input/output (I/O) interconnect are provided. The method includes reading a mailbox stored on an I/O adapter in response to a request to initiate an I/O transaction. The mailbox stores a directive that defines a condition under which cache injection for data values in the I/O transaction will not be performed. The method also includes embedding a hint into the I/O transaction when the directive in the mailbox matches data received in the request, and executing the I/O transaction. The execution of the I/O transaction causes a system chipset or I/O hub for a processor receiving the I/O transaction, to directly store the data values from the I/O transaction into system memory and to suppress the cache injection of the data values into a cache memory upon presence of the hint in a header of the I/O transaction. | 06-18-2009 |
20090157978 | TARGET COMPUTER PROCESSOR UNIT (CPU) DETERMINATION DURING CACHE INJECTION USING INPUT/OUTPUT (I/O) ADAPTER RESOURCES - A method, system, and computer program product for target computer processor unit (CPU) determination during cache injection using input/output (I/O) adapter resources are provided. The method includes storing locations of cache lines for pinned or affinity scheduled processes in a table on an input/output (I/O) adapter. The method also includes setting a cache injection hint in an input/output (I/O) transaction when an address in the I/O transaction is found in the table. The cache injection hint is set for performing direct cache injection. The method further includes entering a central processing unit (CPU) identifier and cache type in the I/O transaction, and updating a cache by injecting data values of the I/O transaction into the cache as determined by the CPU identifier and the cache type associated with the address in the table. | 06-18-2009 |
20090157979 | TARGET COMPUTER PROCESSOR UNIT (CPU) DETERMINATION DURING CACHE INJECTION USING INPUT/OUTPUT (I/O) HUB/CHIPSET RESOURCES - A method, system, and computer program product for target computer processor unit (CPU) determination during cache injection using I/O hub/chipset resources are provided. The method includes creating a cache injection indirection table on the input/output (I/O) hub or chipset. The cache injection indirection table includes fields for address or address range, CPU identifier, and cache type. In response to receiving an input/output (I/O) transaction, the hub/chipset reads the address in an address field of the I/O transaction, looks up the address in the cache injection indirection table, and injects the address and data of the I/O transaction to a target cache associated with a CPU as identified in the CPU identifier field when, in response to the look up, the address is present in the address field of the cache injection indirection table. | 06-18-2009 |
20090213127 | GUIDED ATTACHMENT OF ACCELERATORS TO COMPUTER SYSTEMS - A method of guided attachment of hardware accelerators to slots of a computing system includes dividing a first group of hardware accelerators into a plurality of priority classes, dividing a first group of slots of the computing system into a plurality of hierarchical tiers, and attaching each hardware accelerator of the first group of hardware accelerators to a slot matched to the hardware accelerators based on comparison of a priority class of the hardware accelerator and a hierarchical tier of the slot. | 08-27-2009 |
20090217266 | STREAMING ATTACHMENT OF HARDWARE ACCELERATORS TO COMPUTER SYSTEMS - A method of streaming attachment of hardware accelerators to a computing system includes receiving a stream for processing, identifying a stream handler based on the received stream, activating the identified stream handler, and steering the stream to an associated hardware accelerator. | 08-27-2009 |
20090217275 | PIPELINING HARDWARE ACCELERATORS TO COMPUTER SYSTEMS - A method of pipelining hardware accelerators of a computing system includes associating hardware addresses to at least one processing unit (PU) or at least one logical partition (LPAR) of the computing system, receiving a work request for an associated hardware accelerator address, and queuing the work request for a hardware accelerator using the associated hardware accelerator address. | 08-27-2009 |
20090307711 | INTEGRATING COMPUTATION AND COMMUNICATION ON SERVER ATTACHED ACCELERATORS - In a call-return-communicate scheme an OS/hypervisor/inter-partition shared memory usage is replaced by a software abstraction or mailbox router implemented on an accelerator which handles LPAR communication needs, thereby obviating the need to invoke the OS/hypervisor/inter-partition shared memory. By eliminating the need for the OS/hypervisor/shared memory, system latency is reduced by removing communication and hypervisor invocation time. | 12-10-2009 |
20100262882 | PROTOCOLS FOR HIGH PERFORMANCE COMPUTING VISUALIZATION, COMPUTATIONAL STEERING AND FORWARD PROGRESS - Methods, systems and computer program product for reducing latency and increasing throughput of data transmissions along a switch network path. Exemplary embodiments include a method in a network accelerator device having a memory buffer, a method including identifying a data transmission, copying data packets from the data transmission into the memory buffer, and in response to at least one of a missing data packet and a corrupt data packet identified during the data transmission, sending a copied data packet corresponding to the at least one of the missing data packet and the corrupt data packet. | 10-14-2010 |
20110107035 | CROSS-LOGICAL ENTITY ACCELERATORS - A cross-logical entity group is created that includes one or more accelerators to be shared by a plurality of logical entities. Instantiated on the accelerators are functions that are common across multiple logical entities. The functions to be instantiated are determined, for instance, dynamically during run-time. | 05-05-2011 |
20110107066 | CASCADED ACCELERATOR FUNCTIONS - Accelerator functions are cascaded, such that a result of one accelerator function is directly forwarded to another accelerator function, bypassing the processor requesting the functions to be performed. The cascading may be provided during compilation of a program specifying the functions to be performed, but can be dynamically reversed during runtime of the program. | 05-05-2011 |
20110119444 | ADAPTIVE CACHING OF DATA - Data access is facilitated by employing local caches and an adaptive caching strategy. Specific data is stored in each local cache and consistency is maintained between the caches. To maintain consistency, adaptive caching structures are used. The members of an adaptive caching structure are selected based on a sharing context, such as those members having a chosen association identifier or those members not having the chosen association identifier. | 05-19-2011 |
20110131430 | MANAGING ACCELERATORS OF A COMPUTING ENVIRONMENT - Accelerators of a computing environment are managed in order to optimize energy consumption of the accelerators. To facilitate the management, virtual queues are assigned to the accelerators, and a management technique is used to enqueue specific tasks on the queues for execution by the corresponding accelerators. The management technique considers various factors in determining which tasks to be placed on which virtual queues in order to manage energy consumption of the accelerators. | 06-02-2011 |
20110131580 | MANAGING TASK EXECUTION ON ACCELERATORS - Execution of tasks on accelerator units is managed. The managing includes multi-level grouping of tasks into groups based on defined criteria, including start time of tasks and/or deadline of tasks. The task groups and possibly individual tasks are mapped to accelerator units to be executed. During execution, redistribution of a task group and/or an individual task may occur to optimize a defined energy profile. | 06-02-2011 |
20110145318 | INTERACTIVE ANALYTICS PROCESSING - An end-to-end record, request, response token-based protocol is used to facilitate processing of client jobs. This allows the client to forward analytical tasks of a job directly to an analytics cluster and to record an indication of such at a server. The accelerators of the cluster to perform the tasks are specified in a token provided by the server to the client. | 06-16-2011 |
20110145366 | CONCURRENT EXECUTION OF REQUEST PROCESSING AND ANALYTICS OF REQUESTS - Request processing within a computing environment is facilitated. Request processing and analytics processing for the request are performed substantially concurrently in order to improve efficiency of request execution. The analytics processing is at least commenced, and may complete, prior to receiving an indication of success or failure of the request processing. If request processing fails, analytics processing ceases, if not already complete, and results of the analytic processing are not used. | 06-16-2011 |
20110145429 | MULTI-GRANULAR STREAM PROCESSING - Stream processing is facilitated by distributing responsibility for processing the stream to multiple components of a computing environment. A programmable unit receives one or more streams and determines the operations to be performed for the one or more streams and which components of the computing environment are to perform those operations. It forwards data relating to the one or more streams to one or more components of the computing environment for processing and/or information purposes. | 06-16-2011 |
20110145603 | ENERGY MANAGEMENT FOR DATACENTERS - A datacenter schedules and executes requests to conserve energy. The datacenter uses an event-based opportunistic approach to schedule and run the requests, which provides energy efficiency. The requests are hierarchically batched and sent to the datacenter for scheduling and execution. They are selectively sent over low power links and selectively serviced by low power processors. | 06-16-2011 |
20110317743 | SPREAD SPECTRUM WIRELESS COMMUNICATION CODE FOR DATA CENTER ENVIRONMENTS - A method for wirelessly transmitting management data without cables in a data center is provided. Management data is obtained at a device in the data center. A wireless management data signal is generated from the management data. The wireless management data signal is configured to be transmitted at frequencies matching an ambient electromagnetic noise emitted by the data center, such that the wireless management data signal is masked in the ambient electromagnetic noise emitted by the data center. The wireless management data signal is wirelessly transmitted to another device. | 12-29-2011 |
20110320523 | SPECULATIVE AND COORDINATED DATA ACCESS IN A HYBRID MEMORY SERVER - A method, accelerator system, and computer program product, for prefetching data from a server system in an out-of-order processing environment. A plurality of prefetch requests associated with one or more given data sets residing on the server system are received from an application on the server system. Each prefetch request is stored in a prefetch request queue. A score is assigned to each prefetch request. A set of the prefetch requests are selected from the prefetch queue that comprise a score above a given threshold. A set of data, for each prefetch request in the set of prefetch requests, is prefetched from the server system that satisfies each prefetch request, respectively. | 12-29-2011 |
20110320796 | REDUNDANT POWER SUPPLY CONFIGURATION FOR A DATA CENTER - A redundant power supply configuration for a data center is provided. A method includes receiving instructions to operate power supplies at a high current mode. An individual current for each of the power supplies is calculated to total a high current at the high current mode. The power supplies are operated at the high current mode to provide the high current at the high current mode. In response to operation at the high current mode being complete, the power supplies are operated at a normal mode to provide a normal current at the normal current mode. | 12-29-2011 |
20110320804 | DATA ACCESS MANAGEMENT IN A HYBRID MEMORY SERVER - A method, accelerator system, and computer program access data in an out-of-core processing environment. A data access configuration is received from a server system managing a plurality of data sets. A determination is made that data sets retrieved from the server system are to be stored locally based on the data access configuration. A request to interact with a given data set is received from a user client. At least a portion of the given data set is retrieved from the server system. The at least a portion of the given data set is stored locally a memory based on the data access configuration that has been received. | 12-29-2011 |
20120096109 | Hierarchical Pre-fetch Pipelining in a Hybrid Memory Server - A method, hybrid server system, and computer program product, prefetch data. A set of prefetch requests associated with one or more given datasets residing on the server system are received from a set of accelerator systems. A set of data is prefetched from a memory system residing at the server system for at least one prefetch request in the set of prefetch requests. The set of data satisfies the at least one prefetch request. The set of data that has been prefetched is sent to at least one accelerator system, in the set of accelerator systems, associated with the at least one prefetch request. | 04-19-2012 |
20120102138 | Multiplexing Users and Enabling Virtualization on a Hybrid System - A method, hybrid server system, and computer program product, support multiple users in an out-of-core processing environment. At least one accelerator system in a plurality of accelerator systems is partitioned into a plurality of virtualized accelerator systems. A private client cache is configured on each virtualized accelerator system in the plurality of virtualized accelerator systems. The private client cache of each virtualized accelerator system stores data that is one of accessible by only the private client cache and accessible by other private client caches associated with a common data set. Each user in a plurality of users is assigned to a virtualized accelerator system from the plurality of virtualized accelerator systems. | 04-26-2012 |
20120117312 | Hybrid Server with Heterogeneous Memory - A method, hybrid server system, and computer program product, for managing access to data stored on the hybrid server system. A memory system residing at a server is partitioned into a first set of memory managed by the server and a second set of memory managed by a set of accelerator systems. The set of accelerator systems are communicatively coupled to the server. The memory system comprises heterogeneous memory types. A data set stored within at least one of the first set of memory and the second set of memory that is associated with at least one accelerator system in the set of accelerator systems is identified. The data set is transformed from a first format to a second format, wherein the second format is a format required by the at least one accelerator system. | 05-10-2012 |
20120213507 | ADAPTOR SYSTEM FOR AN ETHERNET NETWORK - An Ethernet adapter system may include a transmitter to insert a payload type identifier sequence in a generic frame procedure header to indicate that a network is a converged enhanced Ethernet network. The transmitter may insert idle sequences in a stream of data frames transmitted along a link. The system may include a receiver to recognize a condition and to force a loss of synchronization condition on the link that will be converted by the receiver into a loss of light condition. The receiver may scan the transmitted stream of data frames for invalid data frames and introduce a code into the stream of data frames whenever an invalid data frame is detected. | 08-23-2012 |
20120215902 | SPREAD SPECTRUM WIRELESS COMMUNICATION CODE FOR DATA CENTER ENVIRONMENTS - A technique for wirelessly transmitting management data without cables in a data center is provided. Management data is obtained at a device in the data center. A wireless management data signal is generated from the management data. The wireless management data signal is configured to be transmitted at frequencies matching an ambient electromagnetic noise emitted by the data center, such that the wireless management data signal is masked in the ambient electromagnetic noise emitted by the data center. The wireless management data signal is wirelessly transmitted to another device. | 08-23-2012 |
20120278503 | ENERGY MANAGEMENT SYSTEM FOR A DATA CENTER NETWORK - An energy management system for a data center network may include a central computer to establish an energy use policy for the computer data center network. The system may also include computer nodes in the computer data center network to receive a packet which is a query that obtains energy information from the nodes and/or a transmission reservation that provides instructions for the energy use policy to be implemented at the computer nodes. The computer nodes may reallocate data traffic on the computer data center network based on the energy use policy to improve energy consumption of the computer data center network. | 11-01-2012 |
20120284730 | SYSTEM TO PROVIDE COMPUTING SERVICES - A system is provided. The system includes a computing device by which first and second commands are inputted, first and second resources disposed in communication with the computing device to be receptive of the first command and responsive to the first command with first and second energy demands in first and second response times, respectively and a managing unit. The managing unit is disposed in communication with the computing device to be receptive of the first and second commands and with the first and second resources to allocate tasks associated with the first command to one of the first and second resources. The tasks are allocated in accordance with the second command and the second command is based on the first and second energy demands and the first and second response times. | 11-08-2012 |
20120300611 | Soft Error Recovery for Converged Networks - Detecting and recovering from soft errors in a network comprising a first device. A first device receives a first data packet. Responsive to receiving a second data packet, the first device determines whether the two data packets are identical. Responsive to the determination that the two data packets are not identical, the first device discards the two data packets, and requests retransmission of the two data packets. | 11-29-2012 |
20130007412 | UNIFIED, WORKLOAD-OPTIMIZED, ADAPTIVE RAS FOR HYBRID SYSTEMS - A method, system, and computer program product for maintaining reliability in a computer system. In an example embodiment, the method includes managing workloads on a first processor with a first processor architecture by an agent process executing on a second processor with a second processor architecture. The method proceeds by activating redundant computation on the second processor by the agent process. The method continues by performing a same computation from a workload of the workloads at least twice. Finally, the method includes comparing results of the same computation. In this embodiment the first processor is coupled the second processor by a network, and the first processor architecture and second processor architecture are different architectures. | 01-03-2013 |
20130007759 | UNIFIED, WORKLOAD-OPTIMIZED, ADAPTIVE RAS FOR HYBRID SYSTEMS - A method, system, and computer program product for maintaining reliability in a computer system. In an example embodiment, the method includes performing a first data computation by a first set of processors, the first set of processors having a first computer processor architecture. The method continues by performing a second data computation by a second processor coupled to the first set of processors, the second processor having a second computer processor architecture, the first computer processor architecture being different than the second computer processor architecture. Finally, the method includes dynamically allocating computational resources of the first set of processors and the second processor based on at least one metric while the first set of processors and the second processor are in operation such that the accuracy and processing speed of the first data computation and the second data computation are optimized. | 01-03-2013 |
20130007762 | PROCESSING WORKLOADS USING A PROCESSOR HIERARCHY SYSTEM - Workload processing is facilitated by use of a processor hierarchy system. The processor hierarchy system includes a plurality of processor hierarchies, each including one or more processors (e.g., accelerators). Each processor hierarchy has associated therewith a set of characteristics that define the processor hierarchy, and the processors of the hierarchy also have a set of characteristics associated therewith. Workloads are assigned to processors of processor hierarchies depending on characteristics of the workload, characteristics of the processor hierarchies and/or characteristics of the processors. | 01-03-2013 |
20130073668 | SPECULATIVE AND COORDINATED DATA ACCESS IN A HYBRID MEMORY SERVER - A method, accelerator system, and computer program product, for prefetching data from a server system in an out-of-order processing environment. A plurality of prefetch requests associated with one or more given data sets residing on the server system are received from an application on the server system. Each prefetch request is stored in a prefetch request queue. A score is assigned to each prefetch request. A set of the prefetch requests are selected from the prefetch queue that comprise a score above a given threshold. A set of data, for each prefetch request in the set of prefetch requests, is prefetched from the server system that satisfies each prefetch request, respectively. | 03-21-2013 |
20130097407 | UNIFIED, WORKLOAD-OPTIMIZED, ADAPTIVE RAS FOR HYBRID SYSTEMS - A method, system, and computer program product for maintaining reliability in a computer system. In an example embodiment, the method includes managing workloads on a first processor with a first processor architecture by an agent process executing on a second processor with a second processor architecture. The method proceeds by activating redundant computation on the second processor by the agent process. The method continues by performing a same computation from a workload of the workloads at least twice. Finally, the method includes comparing results of the same computation. In this embodiment the first processor is coupled the second processor by a network, and the first processor architecture and second processor architecture are different architectures. | 04-18-2013 |
20130097611 | UNIFIED, WORKLOAD-OPTIMIZED, ADAPTIVE RAS FOR HYBRID SYSTEMS - A method, system, and computer program product for maintaining reliability in a computer system. In an example embodiment, the method includes performing a first data computation by a first set of processors, the first set of processors having a first computer processor architecture. The method continues by performing a second data computation by a second processor coupled to the first set of processors, the second processor having a second computer processor architecture, the first computer processor architecture being different than the second computer processor architecture. Finally, the method includes dynamically allocating computational resources of the first set of processors and the second processor based on at least one metric while the first set of processors and the second processor are in operation such that the accuracy and processing speed of the first data computation and the second data computation are optimized. | 04-18-2013 |
20130101284 | ADAPTOR SYSTEM FOR AN ETHERNET NETWORK - An Ethernet adapter system may include a transmitter to insert a payload type identifier sequence in a generic frame procedure header to indicate that a network is a converged enhanced Ethernet network. The transmitter may insert idle sequences in a stream of data frames transmitted along a link. The system may include a receiver to recognize a condition and to force a loss of synchronization condition on the link that will be converted by the receiver into a loss of light condition. The receiver may scan the transmitted stream of data frames for invalid data frames and introduce a code into the stream of data frames whenever an invalid data frame is detected. | 04-25-2013 |
20130103999 | High Performance Virtual Converged Enhanced Ethernet With Persistent State Flow Control - Maintaining bandwidth in links betweens servers and storage arrays comprising a device. The device establishes the links. The device identifies a first link from the links. The first link has a high response time. The device transmits a plurality of data packets on the first link. Each data packet is associated with a corresponding acknowledgment (ACK). The transmission is performed without waiting for the corresponding ACK to be received. The device tracks the ACK received in response to each of the transmitted data packets. The device detects a failure of the first link. In response to the detection, the device identifies invalid data packets. The invalid data packets comprise data packets transmitted on the first link after the detected failure. | 04-25-2013 |
20130114396 | High Performance Virtual Converged Enhanced Ethernet With Persistent State Flow Control - Maintaining bandwidth in links betweens servers and storage arrays comprising a device. The device establishes the links. The device identifies a first link from the links. The first link has a high response time. The device transmits a plurality of data packets on the first link. Each data packet is associated with a corresponding acknowledgment (ACK). The transmission is performed without waiting for the corresponding ACK to be received. The device tracks the ACK received in response to each of the transmitted data packets. The device detects a failure of the first link. In response to the detection, the device identifies invalid data packets. The invalid data packets comprise data packets transmitted on the first link after the detected failure. | 05-09-2013 |
20130114397 | Soft Error Recovery for Converged Networks - Detecting and recovering from soft errors in a network comprising a first device. A first device receives a first data packet. Responsive to receiving a second data packet, the first device determines whether the two data packets are identical. Responsive to the determination that the two data packets are not identical, the first device discards the two data packets, and requests retransmission of the two data packets. | 05-09-2013 |
20130128721 | SYSTEM TO IMPROVE AN ETHERNET NETWORK - A system to improve a Fibre Channel over Convergence Enhanced Ethernet (FCoCEE) network may include a sender in an FCoCEE network in which data packets having different data link layer structures are transmitted by the sender on a single data link. The system may also include a receiver to receive the data packets at the data link layer and to transmit an ACK and/or NAK in response to a sequence number in the data packets. The system may further include a replay buffer to retransmit the data packets where the replay buffer is sized by the length of the data link, data rate of the data link, the ACK and/or NAK processing time at either the sender and/or the receiver, and/or a threshold time for transmission and/or reception of the data packets. | 05-23-2013 |
20130128884 | SYSTEM TO IMPROVE AN ETHERNET NETWORK - A system to improve a Fibre Channel over Convergence Enhanced Ethernet (FCoCEE) network may include a sender in an FCoCEE network in which data packets having different data link layer structures are transmitted by the sender on a single data link. The system may also include a receiver to receive the data packets at the data link layer and to transmit an ACK and/or NAK in response to a sequence number in the data packets. The system may further include a replay buffer to retransmit the data packets where the replay buffer is sized by the length of the data link, data rate of the data link, the ACK and/or NAK processing time at either the sender and/or the receiver, and/or a threshold time for transmission and/or reception of the data packets. | 05-23-2013 |
20130151628 | Time Based System for Urgent Email Messages - An email client receives an urgent tagged email sent by said sender. An urgent timestamp is obtained based on a stored timestamp of a previous urgent tagged email by said sender. Based on the urgent timestamp, it is determined whether said urgent tagged email is allowed. The allowed determining comprises determining whether the urgent tagged email was received within a defined time interval and determining whether the previous timestamp is an allowed value. If the urgent tagged email is allowed, the urgent tagged email is displayed in the email client with an urgent response; a user evaluates whether the urgent tagged email was worth being tagged as urgent; if the urgent tagged email is not worth being tagged urgent, the stored timestamp is updated to the urgent tagged email date; and if the urgent tagged email is worth being tagged urgent, the stored timestamp is updated to the allowed value. | 06-13-2013 |
20130191441 | COLLABORATION AND INTERACTION WITH SYSTEM TERMINALS - A technique is provided for session aggregation in a distributed architecture having a server connected to blades. A master session is generated for collaboration by user clients, and the master session corresponds to one or more terminal sessions on the distributed architecture of the server connected to the blades. The one or more terminal sessions on the distributed architecture are aggregated in the master session. A determination is made that the user clients agree for one user client of the user clients to be authorized as a current command line user name, in response to requests that are sent to the user clients. The one user client is granted authorization to the current command line user name in the master session based on an agreement by the user clients. | 07-25-2013 |
20130191848 | Distributed Function Execution for Hybrid Systems - A system for distributed function execution, the system includes a host in operable communication with an accelerator. The system is configured to perform a method including processing an application by the host and distributing at least a portion of the application to the accelerator for execution. The method also includes instructing the accelerator to create a buffer on the accelerator, instructing the accelerator to execute the portion of the application, wherein the accelerator writes data to the buffer and instructing the accelerator to transmit the data in the buffer to the host before the application requests the data in the buffer. The accelerator aggregates the data in the buffer before transmitting the data to the host based upon one or more runtime conditions in the host. | 07-25-2013 |
20130191849 | DISTRIBUTED FUNCTION EXECUTION FOR HYBRID SYSTEMS - A method includes processing an application by a host including one or more processors and distributing at least a portion of the application to an accelerator for execution. The method includes instructing the accelerator to create a buffer on the accelerator and instructing the accelerator to execute the portion of the application, wherein the accelerator writes data to the buffer. The method also includes instructing the accelerator to transmit the data in the buffer to the host before the application requests the data in the buffer. The accelerator aggregates the data in the buffer before transmitting the data to the host based upon one or more runtime conditions in the host. | 07-25-2013 |
20130194925 | RESERVATION SYSTEM FOR AN ETHERNET NETWORK - An improved computer system may include a server carrying a computer processor in a Fibre Channel over Convergence Enhanced Ethernet (FCoCEE) network. The system may also include a reservation system in communication with the server that utilizes enhanced transmission selection (ETS) to reserve bandwidth for a priority group by scheduling bandwidth for the priority group based upon a projected workload. | 08-01-2013 |
20130194926 | RESERVATION SYSTEM FOR AN ETHERNET NETWORK - An improved computer system may include a server carrying a computer processor in a Fibre Channel over Convergence Enhanced Ethernet (FCoCEE) network. The system may also include a reservation system in communication with the server that utilizes enhanced transmission selection (ETS) to reserve bandwidth for a priority group by scheduling bandwidth for the priority group based upon a projected workload. | 08-01-2013 |
20130198258 | DATABASE IMPROVEMENT SYSTEM - An improved database system may include a root-server including a computer processor. The system may also include a segment-server including a computer processor, the segment-server to store data based upon the data's frequency of use by a client who is closer to the segment-server than the root-server and any other segment-server in the system, and the data stored is at least write data. The system may further include a consistency unit to update the root-server based upon data stored by the segment-server and client. | 08-01-2013 |
20130212376 | DATA ACCESS MANAGEMENT IN A HYBRID MEMORY SERVER - Once or more embodiments manage access to data by accelerator systems in an out-of-core processing environment. In one embodiment, a request from an accelerator system is received for access to a given data set. An access context associated with the given data set is determined. The accelerator system is dynamically configured, based on the access context that has been determined, based on the access context that has been determined, to one of access the given data set directly from the server system; locally store a portion of the given data set in a memory; and locally store all of the given data set in the memory. | 08-15-2013 |
20140101241 | COLLABORATION AND INTERACTION WITH SYSTEM TERMINALS - A technique is provided for session aggregation in a distributed architecture having a server connected to blades. A master session is generated for collaboration by user clients, and the master session corresponds to one or more terminal sessions on the distributed architecture of the server connected to the blades. The one or more terminal sessions on the distributed architecture are aggregated in the master session. A determination is made that the user clients agree for one user client of the user clients to be authorized as a current command line user name, in response to requests that are sent to the user clients. The one user client is granted authorization to the current command line user name in the master session based on an agreement by the user clients. | 04-10-2014 |
20140164644 | ENERGY MANAGEMENT SYSTEM FOR A DATA CENTER NETWORK - An energy management system for a data center network may include a central computer to establish an energy use policy for the computer data center network. The system may also include computer nodes in the computer data center network to receive a packet which is a query that obtains energy information from the nodes and/or a transmission reservation that provides instructions for the energy use policy to be implemented at the computer nodes. The computer nodes may reallocate data traffic on the computer data center network based on the energy use policy to improve energy consumption of the computer data center network. | 06-12-2014 |
20140185621 | ENERGY MANAGEMENT FOR COMMUNICATION NETWORK ELEMENTS - A method of managing data flow may include assigning, along a communication route, a predetermined total energy consumption limit for a network adaptor having a first and a second network adaptor port card, whereby the route transmits data between a first and a second node within the network. Along the route, a total energy consumption value for the network adaptor is determined based on the transmitted data. The assigned total energy consumption limit is compared with the determined total energy consumption value of the network adaptor. The transmitted data is determined as being data of a known duration and data of an unknown duration based on the comparing resulting in the determined total energy consumption value exceeding the assigned total energy consumption limit. The data of a known duration is transferred from the first network adaptor to the second network adaptor and the first network adaptor is powered off. | 07-03-2014 |
20140189157 | ENERGY MANAGEMENT FOR COMMUNICATION NETWORK ELEMENTS - A method of managing data flow may include determining data flows as one of data flows of known duration (KD) and data flows of arbitrary duration (AD). Profile energy consumption values for each of the KD flows is determined based on transmitting the KD flows between a first and a second node via a first network adaptor port, a second network adaptor port card, and switch port cards. Energy efficient routes for transmitting the KD flows between the nodes are also determined. Based on the profile energy consumption values, network element reference energy consumption values for each of network adaptor port cards and each of switch port cards located along the routes are determined, whereby transmission of the KD flows is based on energy consumption values at each of network adaptor port cards and each of switch port cards not exceeding respective network element reference energy consumption values. | 07-03-2014 |
20140223151 | KERNEL EXECUTION FOR HYBRID SYSTEMS - A method for executing kernels in a hybrid system includes running a program on a host computer and identifying in an instruction stream of the program a first instruction including a function of a target classification. The method includes generating a first kernel including the function and transmitting the first kernel to a client system to execute the first kernel based on identifying the first instruction as being of the target classification. The method also includes determining whether to store results of executing the first kernel in a read-only buffer of the client system based on determining whether a subsequent instruction of the target classification relies upon results of the first instruction. | 08-07-2014 |
20140244777 | DISK MIRRORING FOR PERSONAL STORAGE - Embodiments of the present invention provide a system for backing up personal data between two mated (i.e., paired) network attached storage (NAS) devices. The system includes a local storage device and a secondary storage device that communicate over a network (e.g., the Internet) via a network connection. Any data added or modified on the local storage device will be automatically mirrored (i.e., copied) to the secondary storage device, which may be located at a secure remote site, pursuant to a data mirroring technique. | 08-28-2014 |
20140269277 | DYNAMIC MAXIMUM TRANSMISSION UNIT SIZE ADAPTION - An aspect includes dynamically adapting a maximum transmission unit (MTU) size in a computer network. A packet is received at a first link of a switch of the computer network. An MTU size of the received packet is determined based on an application programming interface (API) crossing value associated with the first link of the switch. An associated destination MTU size of a second link of the switch is determined as a target destination of the received packet. The associated destination MTU size of the second link is adjusted based on a difference between the MTU size of the received packet and the associated destination MTU size of the second link. | 09-18-2014 |
20140269319 | NETWORK PER-FLOW RATE LIMITING - A method of monitoring data flow in a network is provided. The method includes: configuring a data flow including a plurality of data packets by a switch controller, the switch controller configured to control routing through the switch and switch configuration, wherein configuring includes storing an indication of a flow control policy in one or more of the data packets; monitoring a network switch receiving the data flow, wherein monitoring includes determining flow statistics in the switch; determining whether a congestion condition exists for the data flow based on the flow statistics and the flow control policy; and based on determining that the congestion condition exists for the data flow, performing a remedial action specific to the data flow to address the congestion condition. | 09-18-2014 |
20140269320 | Scalable Flow and Cogestion Control with OpenFlow - A method of performing congestion management in a network is provided. The method includes monitoring a congestion status at a switch in the network. It is determined at the switch that the congestion status indicates that there is congestion at the switch. Based on the congestion being related to data received at the switch from a source at a first rate, a first message is transmitted from the switch to the source requesting the source to send the data at a second rate that is slower than the first rate. Based on the congestion being related to data requests received at the switch from a destination at third rate, a second message is transmitted from the switch to the destination requesting the destination to request the data at a fourth rate slower than the third rate. | 09-18-2014 |
20140269731 | RELIABLE LINK LAYER FOR CONTROL LINKS BETWEEN NETWORK CONTROLLERS AND SWITCHES - A method for transmission of control data between a network switch and a switch controller is provided. The method includes: configuring a plurality of control data packets by the switch controller, wherein configuring includes disposing a sequence number in each of the plurality of control data packets indicating an order of data packet transmission; storing the plurality of control data packets in a replay buffer in communication with the switch controller; transmitting the plurality of control data packets to the network switch over a secure link between the switch controller and the network switch; and responsive to determining that one or more control data packets were not received by the network switch, retrieving the one or more control data packets from the replay buffer and re-transmitting the one or more control data packets to the network switch. | 09-18-2014 |
20140270764 | CONTROL PLANE FOR INTEGRATED SWITCH WAVELENGTH DIVISION MULTIPLEXING - A method of controlling optical data flow in a network is provided. The method includes: receiving, by a switch controller, a data flow including a plurality of data packets used for network management and control, the network switch including at least one wavelength tunable optical data transmission interface to a wavelength division multiplexed optical network; transmitting a control signal indicating which wavelengths are in use and which wavelengths are not in use to the network switch over a control link; receiving a return signal generated by an optical filter in the switch, the return signal indicative of an available wavelength for transmission of the data flow to the switch; allocating a data traffic flow to the available wavelength, and communicating the allocation decision to the network switch; and transmitting the data flow to the switch via an optical carrier signal having the available wavelength indicated by the return signal. | 09-18-2014 |
20140280547 | Virtual Machine Mobility Using OpenFlow - A method for migrating virtual machines (VMs) that includes creating a temporary storage area. The temporary storage area is created by a controller in a network that is in a first state. The configuration data for a VM is stored in the temporary storage area. The network is provisioned from the first state to a second state. The provisioning is by the controller and based on network locations of a source server and a target server. A message is transmitted to the target server. The message includes a request to migrate the VM from the source server to the target server, and an address of the temporary storage area. It is determined that the migration has completed and the temporary storage area is removed based on the determining that the migrant has completed. | 09-18-2014 |
20140280665 | CELL FABRIC HARDWARE ACCELERATION - An aspect includes a method for providing direct communication between a server and a network switch in a cell-based fabric. A host channel adapter of a cell fabric hardware accelerator is configured to provide the server with direct access to memory within the network switch. A plurality of data packets having a fixed size is received at the host channel adapter from the server. The host channel adapter is coupled to a bus of the server. A direct transmission is performed from the cell fabric hardware accelerator to the memory within the network switch on an interconnect bus to write the data packets directly into the memory. | 09-18-2014 |
20140280791 | DYNAMIC PORT TYPE DETECTION - An aspect includes dynamic port type detection. A link initialization frame that includes a protocol identifier is received at a port of a switch. Based on successfully decoding the protocol identifier, the port is configured to a mode of operation compatible with the protocol identifier. A table in the switch is updated to configure one or more other ports of the switch to the mode of operation compatible with the protocol identifier. A handshake response is sent to acknowledge the successful decoding of the protocol identifier. If no handshake response is received at a server prior to expiration of a wait pulse repetition time, the server periodically transmits the link initialization frame with a different protocol identifier for an alternate mode of operation until the handshake response is received or all supported modes of operation are checked. | 09-18-2014 |
20140281669 | OpenFlow Controller Master-slave Initialization Protocol - A method for network controller initialization that includes identifying a controller connected to a network as a primary controller that manages switches in the network. One or more other controllers connected to the network are identified as secondary controllers. A failover priority table is created. The failover table indicates an order that the one or more other controllers will replace the controller as the primary controller in the event that the controller enters a failure mode. The failover priority table is broadcast to the switches in the network. | 09-18-2014 |
20150016276 | RELIABLE LINK LAYER FOR CONTROL LINKS BETWEEN NETWORK CONTROLLERS AND SWITCHES - A method for transmission of control data between a network switch and a switch controller is provided. The method includes: configuring a plurality of control data packets by the switch controller, wherein configuring includes disposing a sequence number in each of the plurality of control data packets indicating an order of data packet transmission; storing the plurality of control data packets in a replay buffer in communication with the switch controller; transmitting the plurality of control data packets to the network switch over a secure link between the switch controller and the network switch; and responsive to determining that one or more control data packets were not received by the network switch, retrieving the one or more control data packets from the replay buffer and re-transmitting the one or more control data packets to the network switch. | 01-15-2015 |
20150019696 | DYNAMIC PORT TYPE DETECTION - An aspect includes dynamic port type detection. A link initialization frame that includes a protocol identifier is received at a port of a switch. Based on successfully decoding the protocol identifier, the port is configured to a mode of operation compatible with the protocol identifier. A table in the switch is updated to configure one or more other ports of the switch to the mode of operation compatible with the protocol identifier. A handshake response is sent to acknowledge the successful decoding of the protocol identifier. If no handshake response is received at a server prior to expiration of a wait pulse repetition time, the server periodically transmits the link initialization frame with a different protocol identifier for an alternate mode of operation until the handshake response is received or all supported modes of operation are checked. | 01-15-2015 |
20150019741 | Virtual Machine Mobility Using OpenFlow - A method for migrating virtual machines (VMs) that includes creating a temporary storage area. The temporary storage area is created by a controller in a network that is in a first state. The configuration data for a VM is stored in the temporary storage area. The network is provisioned from the first state to a second state. The provisioning is by the controller and based on network locations of a source server and a target server. A message is transmitted to the target server. The message includes a request to migrate the VM from the source server to the target server, and an address of the temporary storage area. It is determined that the migration has completed and the temporary storage area is removed based on the determining that the migrant has completed. | 01-15-2015 |
20150019902 | OpenFlow Controller Master-slave Initialization Protocol - A method for network controller initialization that includes identifying a controller connected to a network as a primary controller that manages switches in the network. One or more other controllers connected to the network are identified as secondary controllers. A failover priority table is created. The failover table indicates an order that the one or more other controllers will replace the controller as the primary controller in the event that the controller enters a failure mode. The failover priority table is broadcast to the switches in the network. | 01-15-2015 |
20150055476 | Scalable Flow and Cogestion Control with OpenFlow - A method of performing congestion management in a network is provided. The method includes monitoring a congestion status at a switch in the network. It is determined at the switch that the congestion status indicates that there is congestion at the switch. Based on the congestion being related to data received at the switch from a source at a first rate, a first message is transmitted from the switch to the source requesting the source to send the data at a second rate that is slower than the first rate. Based on the congestion being related to data requests received at the switch from a destination at third rate, a second message is transmitted from the switch to the destination requesting the destination to request the data at a fourth rate slower than the third rate. | 02-26-2015 |
20150055480 | DYNAMIC MAXIMUM TRANSMISSION UNIT SIZE ADAPTION - An aspect includes dynamically adapting a maximum transmission unit (MTU) size in a computer network. A packet is received at a first link of a switch of the computer network. An MTU size of the received packet is determined based on an application programming interface (API) crossing value associated with the first link of the switch. An associated destination MTU size of a second link of the switch is determined as a target destination of the received packet. The associated destination MTU size of the second link is adjusted based on a difference between the MTU size of the received packet and the associated destination MTU size of the second link. | 02-26-2015 |
20150055649 | CELL FABRIC HARDWARE ACCELERATION - An aspect includes a method for providing direct communication between a server and a network switch in a cell-based fabric. A host channel adapter of a cell fabric hardware accelerator is configured to provide the server with direct access to memory within the network switch. A plurality of data packets having a fixed size is received at the host channel adapter from the server. The host channel adapter is coupled to a bus of the server. A direct transmission is performed from the cell fabric hardware accelerator to the memory within the network switch on an interconnect bus to write the data packets directly into the memory. | 02-26-2015 |