Patent application number | Description | Published |
20090201926 | FIBRE CHANNEL PROCESSING BY A HOST CHANNEL ADAPTER - A method for data storage includes mapping a queue pair (QP) of a channel adapter to a specified Fibre Channel (FC) exchange for communication with a storage device. Upon receiving at the channel adapter from a host computer a storage command directed to the storage device, the storage command is executed by transmitting data packets over a switched network from the channel adapter to the storage device using the specified exchange and performing a remote direct memory access (RDMA) operation on the channel adapter using the mapped QP. | 08-13-2009 |
20100138840 | SYSTEM AND METHOD FOR ACCELERATING INPUT/OUTPUT ACCESS OPERATION ON A VIRTUAL MACHINE - A system and method for accelerating input/output (IO) access operation on a virtual machine, The method comprises providing a smart IO device that includes an unrestricted command queue (CQ) and a plurality of restricted CQs and allowing a guest domain to directly configure and control IO resources through a respective restricted CQ, the IO resources allocated to the guest domain. In preferred embodiments, the allocation of IO resources to each guest domain is performed by a privileged virtual switching element. In some embodiments, the smart IO device is a HCA and the privileged virtual switching element is a Hypervisor. | 06-03-2010 |
20100274876 | NETWORK INTERFACE DEVICE WITH MEMORY MANAGEMENT CAPABILITIES - An input/output (I/O) device includes a host interface for connection to a host device having a memory and a network interface, which is configured to receive, over a network, data packets associated with I/O operations directed to specified virtual addresses in the memory. Packet processing hardware is configured to translate the virtual addresses into physical addresses and to perform the I/O operations using the physical addresses, and upon an occurrence of a page fault in translating one of the virtual addresses, to transmit a response packet over the network to a source of the data packets so as to cause the source to refrain from transmitting further data packets while the page fault is serviced. | 10-28-2010 |
20110029847 | PROCESSING OF DATA INTEGRITY FIELD - A network communication device includes a host interface, which is coupled to communicate with a host processor, having a memory, so as to receive a work request to convey one or more data blocks over a network. The work request specifies a memory region of a given data size, and at least one data integrity field (DIF), having a given field size, is associated with the data blocks. Network interface circuitry is configured to execute an input/output (I/O) data transfer operation responsively to the work request so as to transfer to or from the memory a quantity of data that differs from the data size of the memory region by a multiple of the field size, while adding the at least one DIF to the transferred data or removing the at least one DIF from the transferred data. | 02-03-2011 |
20120174102 | SYSTEM AND METHOD FOR ACCELERATING INPUT/OUTPUT ACCESS OPERATION ON A VIRTUAL MACHINE - A system and method for accelerating input/output (IO) access operation on a virtual machine, The method comprises providing a smart IO device that includes an unrestricted command queue (CQ) and a plurality of restricted CQs and allowing a guest domain to directly configure and control IO resources through a respective restricted CQ, the IO resources allocated to the guest domain. In preferred embodiments, the allocation of IO resources to each guest domain is performed by a privileged virtual switching element. In some embodiments, the smart IO device is a HCA and the privileged virtual switching element is a Hypervisor. | 07-05-2012 |
20120207018 | REDUCING POWER CONSUMPTION IN A FAT-TREE NETWORK - A method for communication includes estimating a characteristic of communication traffic to be carried by a fat-tree network. Responsively to the estimated characteristic, a subset of the spine switches in the highest level of the network is selected, according to a predetermined selection order, to be active in carrying the communication traffic. In each of the levels of the spine switches below the highest level, the spine switches to be active are selected based on the selected spine switches in a next-higher level. The network is operated so as to convey the traffic between the leaf switches via the active spine switches, while the spine switches that are not selected remain inactive. | 08-16-2012 |
20120292267 | MOUNTING RAIL WITH INTERNAL POWER CABLE - An adapter kit for mounting an electrical apparatus in a rack includes a pair of rails. The rails are configured to be fitted on respective, opposing outer sides of a case of the apparatus and to slide along corresponding tracks on respective inner sides of the rack. At least one of the rails includes a cable channel configured to contain a cable passing through the cable channel between front and rear faces of the case. | 11-22-2012 |
20120311220 | COMPUTER BUS WITH ENHANCED FUNCTIONALITY - A method for computing includes connecting a host device to a peripheral device via a bus that is physically configured in accordance with a predefined standard and includes multiple connection pins that are specified by the standard, including a plurality of ground pins. At least one pin, selected from among the pins on the bus that are specified as the ground pins, is used in order to indicate to the peripheral device that the host device has an extended operational capability. | 12-06-2012 |
20140023084 | REDUCING POWER CONSUMPTION IN A FAT-TREE NETWORK - A method for communication includes configuring a multi-level fat-tree network to include at least three levels of switches, including multiple modules arranged externally in a tree topology. Each module contains a respective group of the switches arranged in an internal tree extending over at least two of the levels of the network. A subset of the modules is selected to be active in carrying the communication traffic. The network is operated so as to convey communication traffic among the switches via the active modules, while the modules that are not in the selected subset remain inactive. | 01-23-2014 |
20140075436 | SYSTEM AND METHOD FOR ACCELERATING INPUT/OUTPUT ACCESS OPERATION ON A VIRTUAL MACHINE - A system and method for accelerating input/output (IO) access operation on a virtual machine, The method comprises providing a smart IO device that includes an unrestricted command queue (CQ) and a plurality of restricted CQs and allowing a guest domain to directly configure and control IO resources through a respective restricted CQ, the IO resources allocated to the guest domain. In preferred embodiments, the allocation of IO resources to each guest domain is performed by a privileged virtual switching element. In some embodiments, the smart IO device is a HCA and the privileged virtual switching element is a Hypervisor. | 03-13-2014 |
20140129784 | METHODS AND SYSTEMS FOR POLLING MEMORY OUTSIDE A PROCESSOR THREAD - A system and method of monitoring a memory address is disclosed which may replace a polling operation on a memory by determining a memory address to monitor, notifying a cache controller of the memory address, and cause execution on a polling thread to wait. The cache controller may then monitor the memory address and notify the processor to resume execution of the thread. While the processor is waiting to be notified, it may enter a power save state or allow more time to be allocated to other threads being executed. | 05-08-2014 |
20140169169 | ROUTING SUPPORT FOR LOSSLESS DATA TRAFFIC - A method for communication in a packet data network including at least first and second subnets interconnected by routers. The method includes defining at least first and second classes of link-layer traffic within the subnets, such that the link-layer traffic in the first class is transmitted among nodes in the network without loss of packets, while at least some of the packets in the second class are dropped in case of network congestion. The routers are configured by transmitting control traffic over the network in the packets of the second class. Data traffic is transmitted between the nodes in the first and second subnets via the configured routers in the packets of the first class. | 06-19-2014 |
20140177639 | ROUTING CONTROLLED BY SUBNET MANAGERS - A method for communication in a packet data network that includes at least first and second subnets interconnected by multiple routers and having respective first and second subnet managers. The method includes assigning respective local identifiers to ports for addressing of data link traffic within each subnet, such that the first subnet manager assigns the local identifiers in the first subnet, and the second subnet manager assigns the local identifiers in the second subnet. The routers are configured by transmitting and receiving control traffic between the subnet managers and the routers. Data packets are transmitted between network nodes in the first and second subnets via one or more of the configured routers under control of the subnet managers. | 06-26-2014 |
20140185615 | SWITCH FABRIC SUPPORT FOR OVERLAY NETWORK FEATURES - A method for communication in a packet data network including a subnet containing multiple nodes having respective ports. The method includes assigning respective local identifiers to the ports in the subnet, such that each port receives a respective local identifier that is unique within the subnet to serve as an address for traffic within the subnet that is directed to the port. In addition to the local identifiers, respective port identifiers are assigned to the ports, such that at least one of the port identifiers is shared by a plurality of the ports, but not by all the ports, in the subnet. The plurality of the ports are addressed collectively using the at least one of the port identifiers. | 07-03-2014 |