Patent application number | Description | Published |
20080314979 | CABLE MANAGEMENT SYSTEMS - In order to monitoring connectivity of cabling in a system comprising chassis components and cables interconnecting the chassis components, machine readable labels are applied to cable connectors and chassis component connectors, the machine readable labels on a cable connector and a chassis component to which the cable connector is connected are scanned, and the scanned connectivity information is recorded. | 12-25-2008 |
20080315889 | FAULT ISOLATION IN INTERCONNECT SYSTEMS - A loopback connector for a system can include a connector arrangement connectable to connector of a system component and/or a cable. The loopback connector can include loopback logic for simulating cable and/or system component functionality. In an example implementation the loopback connector can also operate to protect a system component and/or cable connector during shipping. | 12-25-2008 |
20080315985 | MULTI-SWITCH CHASSIS - In a switch system L groups of the line switch elements are connectable to cables that include L links such that each of the L links within a cable connect to a switch element of a respective one of the L groups. Fabric switch elements are connected such that a fabric switch element is connected to the line switch elements of one of the group of line switch elements. | 12-25-2008 |
20080317025 | SWITCH MATRIX - A switch system comprises a switch matrix providing a configurable, protocol agnostic, packet switching core, the switch matrix providing a non-blocking switch connecting a configurable number of ports together. | 12-25-2008 |
20080318465 | CABLE INTERCONNECT SYSTEMS - In order to monitor connectivity status associated with an interconnect cable from the end-points to which either end of the cable is attached, a storage device storing a field replaceable unit identifier is provided to uniquely identify a cable end-point, which identifier can then be accessed to determine the connectivity status. | 12-25-2008 |
20100329275 | Multiple Processes Sharing a Single Infiniband Connection - A compute node with multiple transfer processes that share an Infiniband connection to send and receive messages across a network. Transfer processes are first associated with an Infiniband queue pair (QP) connection. Then send message commands associated with a transfer process are issued. This causes an Infiniband message to be generated and sent, via the QP connection, to a remote compute node corresponding to the QP. Send message commands associated with another process are also issued. This causes another Infiniband message to be generated and sent, via the same QP connection, to the same remote compute node. As mentioned, multiple processes may receive network messages received via a shared QP connection. A transfer process on a receiving compute node receives a network message through a QP connection using a receive queue. A second transfer process receives another message through the same QP connection using another receive queue. | 12-30-2010 |
20100332908 | Performing Remote Loads and Stores over Networks - A network adaptor which performs CPU loads and stores to remote memory over network fabrics. The network adaptor receives a transfer request from a compute node and converts the request to a remote transfer request, which is transmitted to the network. The network adaptor then monitors the network connection for a remote completion response. When the network adaptor receives the remote completion response within a specific time period, the network adaptor transmits a first completion response to the compute node. If the network adaptor does not receive the remote completion response within the specific time period, the network adaptor transmits an “early completion response” to the compute node. The network adaptor continues to monitor for the actual response. This allows the compute node to continue processing without having to wait for the actual response to be received. The method handles small payloads efficiently and also accounts for long completion delays. | 12-30-2010 |
20120239730 | SYSTEM INCLUDING A MIDDLEWARE MACHINE ENVIRONMENT - A system and method for providing a middleware machine or similar platform. In accordance with an embodiment the system (referred to herein in some implementations as “Exalogic”) comprises a combination of high performance hardware, together with an application server or middleware environment, to provide a complete Java EE application server complex which includes a massively parallel in-memory grid, can be provisioned quickly, and can scale on demand. In accordance with an embodiment, the system can be deployed as a full, half, or quarter rack, or other configuration, that provides an application server grid, storage area network, and InfiniBand network, which support the execution of an application server, middleware or other functionality such as, for example, WebLogic Server, JRockit or Hotspot JVM, Oracle Linux or Solaris, and Oracle VM. Additional features of the system can include, e.g. Zero Buffer Copies, Scatter/Gather I/O, T3 Connections, and Lazy Deserialization. | 09-20-2012 |
20120311597 | METHOD AND SYSTEM FOR INFINIBAND HOST CHANNEL ADAPTOR QUALITY OF SERVICE - A method for allocating resources of a host channel adapter includes the host channel adapter identifying an underlying function referenced in the first resource allocation request received from a virtual machine manager, determining that the first resource allocation request specifies a number of physical collect buffers (PCBs) allocated to the underlying function, allocating the number of PCBs to the underlying function, determining that the first resource allocation request specifies a number of virtual collect buffers (VCBs) allocated to the underlying function, and allocating the number of VCBs to the underlying function. The host channel adapter further receives command data for a command from the single virtual machine, determines that the underlying function has in use at least the number of PCBs when the command data is received, and drops the command data in the first command based on the underlying function having in use at least the number of PCBs. | 12-06-2012 |