Patent application number | Description | Published |
20080209273 | Detect User-Perceived Faults Using Packet Traces in Enterprise Networks - Exemplary methods, computer-readable media, and systems for detecting a fault by a packet trace, includes monitoring at least one packet transmitted to or received from, an computing device of an end user, between one or more computing devices implementing at least one of a service or an application on an enterprise network. The process also includes identifying whether an abnormal condition occurred on the computing device of the end user based on monitoring at least one packet transmitted to or received from, the computing device of the end user; and detecting a fault by using an algorithm based on monitoring at least one packet transmitted or received from, the computing device of the end user; wherein the fault indicates a desired course of action did not occur while the computing device of the end user uses at least one of the service or the application in the enterprise network. | 08-28-2008 |
20080222068 | Inferring Candidates that are Potentially Responsible for User-Perceptible Network Problems - Candidates that are potentially responsible for user-perceptible network problems may be inferred. In an example embodiment, a system includes an inference engine to produce a list of candidates that are potentially responsible for user-perceptible network problems, with the candidates being network components that may include both services and network elements. A response to a service request may be a non response, an incorrect response, an untimely correct response, or a timely correct response. The user-perceptible network problems may include the untimely correct response as well as the non response and the incorrect response. In another example embodiment, a method includes monitoring a network and producing a list of candidates that are potentially responsible for user-perceptible network problems. The candidates of the list may include both services and network elements of the monitored network. | 09-11-2008 |
20080222287 | Constructing an Inference Graph for a Network - Constructing an inference graph relates to the creation of a graph that reflects dependencies within a network. In an example embodiment, a method includes determining dependencies among components of a network and constructing an inference graph for the network responsive to the dependencies. The components of the network include services and hardware components, and the inference graph reflects cross-layer components including the services and the hardware components. In another example embodiment, a system includes a service dependency analyzer and an inference graph constructor. The service dependency analyzer is to determine dependencies among components of a network, the components including services and hardware components. The inference graph constructor is to construct an inference graph for the network responsive to the dependencies, the inference graph reflecting cross-layer components including the services and the hardware components. | 09-11-2008 |
20090086640 | UNCOVERING THE DIFFERENCES IN BACKBONE NETWORKS - The claimed subject matter provides systems and/or methods that discover comparative differences in performance of Internet Service Provider networks. The system can include devices, components, and/or processes that partitions measurement processes into cycles and measures at least one Internet Service Provider network every cycle, develops and distributes to interrogation nodes a list of destination Internet Protocol (IP) addresses to interrogate, and utilizes results delivered or retrieved from interrogation nodes to generate a report on the comparative performance of Internet Service Provider networks queried. | 04-02-2009 |
20090086741 | UNCOVERING THE DIFFERENCES IN BACKBONE NETWORKS - The claimed subject matter provides systems and/or methods that uncovers detailed differences in the performance of different Internet Service Provider ISP) networks. The system can include components, devices, and/or processes that utilize a list of target Internet Service Provider (ISP) networks to create a list of Internet Protocol (IP) addresses to interrogate, distribute the list of addresses to measurement nodes and obtains in return probing results from the measurement nodes, the probing results employed thereafter to provide comparative performance metrics associated with the target Internet Service Provider (ISP) networks. | 04-02-2009 |
20090204696 | SERVICE DEPENDENCY DISCOVERY IN ENTERPRISE NETWORKS - The claimed subject matter provides systems and/or techniques that identify service dependencies in enterprise networks. The system can include devices that, based on received network data packets, identify packets with common five tuples, aggregate the identified packets to form transactions associated with individual hosts, and determine delay distributions between one or more services solicited from or in response to the individual hosts. Based on the delay distributions, the system differentiates between dependent service pairs and independent service pairs and thereafter assembles and outputs dependency graphs that include dependency relationships between services from the perspective of a client-side and a server-side. | 08-13-2009 |
20100076933 | TECHNIQUES FOR RESOURCE LOCATION AND MIGRATION ACROSS DATA CENTERS - An exemplary system includes a front-end component to receive requests for resources in a data center and configured to associate each request with identifying information, to locate one or more resources for each request and to store, in a log file, the identifying information and information about the location of the one or more resources; one or more distributed computation and storage components to acquire log file information and configured to analyze log information to decide if one or more resources associated with one or more requests should be migrated to a data center in a different geographical location; and a location service component to receive decisions made by the one or more distributed computation and storage components and configured to inform the front-end component when a decision causes one or more resources to be migrated to a data center in a different geographical location to thereby allow the front-end component to re-direct future requests for the one or more migrated resources to the data center in the different geographical location. Various other devices, systems and methods are also described. | 03-25-2010 |
20110032833 | OPTIMIZATION OF TRAFFIC ROUTING FOR DATA CENTER SERVICES - Techniques and systems for providing optimization of traffic routing for data centers are disclosed herein. In some embodiments, a method may include identifying a plurality of available source sites and paths or routes between an online service provider (OSP) and a destination prefix. A traffic manager may measure a cost for each of the plurality of available paths. In some instances, the traffic manager may also measure a performance value for each of the available paths. The traffic manager may then select one of the available source sites and paths as a preferred source site and path that has a minimized cost for an instance of the performance value when compared to other paths of the plurality of available source sites and paths. In further embodiments, the traffic manager may rewrite a routing table and DNS mapping to implement the preferred source site and path as the default source site and path between the OSP and the destination prefix. | 02-10-2011 |
20110040876 | CAPACITY PLANNING FOR DATA CENTER SERVICES - Techniques and systems for providing capacity planning for data center services are disclosed herein. In some aspects, a multi-tier service is analyzed to determine a query response time (QRT) that can be used for resource planning or other service management reasons. The QRT may include a front end (FE) response time of FE servers and a back end (BE) response time of BE servers, where the BE servers are partitioned servers, and where the response times are based on various request rates from users. In various aspects, the QRT may include a network transmission time of data transmitted between an end user and each tier. In further aspects, the FE response time may be modeled using a single server queuing model while the BE response time may be modeled using a simulation or regression analysis. | 02-17-2011 |
20110054878 | AUTOMATED PERFORMANCE PREDICTION FOR CLOUD SERVICES - Embodiments of automated cloud service performance prediction are disclosed. The automated cloud service performance prediction includes extracting a parental dependency graph (PDG) for a webpage. The PDG encapsulates one or more dependency relationships for each web object in the webpage. The prediction further includes determining an original page load time (PLT) and original timing information of a webpage. The prediction also includes simulating a page loading of the webpage based on adjusted timing information of each web object and the PDG to estimate a new PLT of the webpage. The prediction additionally includes comparing the original PLT of the webpage to the new PLT of the webpage to determine whether the adjusted timing information increased or decreased the new PLT of the webpage. | 03-03-2011 |
20120030338 | WEB PAGE LOAD TIME PREDICTION AND SIMULATION - Embodiments of automated cloud service performance prediction are disclosed. The automated cloud service performance prediction includes extracting a parental dependency graph (PDG) for a webpage. The PDG encapsulates one or more dependency relationships for each web object in the webpage. The prediction further includes determining an original page load time (PLT) and original timing information of a webpage. The prediction also includes simulating a page loading of the webpage based on adjusted timing information of each web object and the PDG to estimate a new PLT of the webpage. The prediction additionally includes comparing the original PLT of the webpage to the new PLT of the webpage to determine whether the adjusted timing information increased or decreased the new PLT of the webpage. | 02-02-2012 |
20120127995 | BACKBONE NETWORK WITH POLICY DRIVEN ROUTING - A source device obtains a data packet that includes both a destination address and a payload. The source device selects an exit point address of multiple exit point addresses corresponding to the destination address based on one or more policies. The source device encapsulates the data packet with a header that includes the selected exit point address, and the encapsulated data packet is provided to the backbone network. The encapsulated data packet is routed through the backbone network based on the exit point address, and an edge router of the backbone network identifies an interface of the edge router that corresponds to the exit point address. The header is removed from the encapsulated data packet, and the data packet is added to a buffer of the interface for routing to one or more other devices outside of the backbone network. | 05-24-2012 |
20130054782 | DETERMINATION OF UNAUTHORIZED CONTENT SOURCES - A plurality of network addresses from a distributed client is obtained, at least a first portion of the obtained network addresses including resolved network address responses to distributed client requests for resolved network addresses corresponding to one or more network location indicators associated with a first web service. Test content is obtained, based on one or more of the network addresses included in the first portion. It is determined whether the obtained test content includes unauthorized content. | 02-28-2013 |
20130097313 | WEB PAGE LOAD TIME PREDICTION AND SIMULATION - Embodiments of automated cloud service performance prediction are disclosed. The automated cloud service performance prediction includes extracting one or more dependency relationships for each web object in the webpage. The prediction further includes determining an original performance metric value and original timing information associated with a page loading of a webpage. The prediction also includes simulating a page loading of the webpage based on the adjusted timing information and the dependency relationships to estimate a new performance metric value associated with the simulated page loading of the webpage. The prediction additionally includes comparing the original performance metric value to the new performance metric value to determine whether the adjusted timing information increased or decreased the new performance metric value relative to the original performance metric value. | 04-18-2013 |
20130122854 | PROVIDING PER-APPLICATION RESOURCE USAGE INFORMATION - An environment is described in which a processing system provides application-level usage information to users. In one scenario, for example, the processing system may provide personal usage information to a user who is operating a user device. The personal usage information itemizes the amount of data (and/or other resources) that has been consumed by each application run by the user device. In another scenario, the processing system may provide expected usage information associated with at least one candidate application provided by a marketplace system. The expected usage information describes an expected consumption of data (and/or other resources) by the candidate application upon running the candidate application by the user device. The processing system can tailor the expected usage information that it sends to a particular user based on user profile data. The user profile data describes a manner in which users operate applications. | 05-16-2013 |
20140078882 | Automated Datacenter Network Failure Mitigation - The subject disclosure is directed towards a technology that automatically mitigates datacenter failures, instead of relying on human intervention to diagnose and repair the network. Via a mitigation pipeline, when a network failure is detected, a candidate set of components that are likely to be the cause of the failure is identified, with mitigation actions iteratively targeting each component to attempt to alleviate the problem. The impact to the network is estimated to ensure that the redundancy present in the network will be able to handle the mitigation action without adverse disruption to the network. | 03-20-2014 |
20140195366 | INCREMENTAL VALUATION BASED NETWORK CAPACITY ALLOCATION - A bid-based network sells network capacity on a transaction-by-transaction basis in accordance with bids placed on transactions. A transaction is the transmission of a quantum of data across at least some portion of the network, where the quantum of data can be as small as a single packet. Bids for network capacity are ranked in order of monetary value, or other criteria relevant to the network service provider. The amount charged to the highest bidder is based on the maximum bid of the next highest bidder. Bids are evaluated on a real-time basis at the time when the link is ready to transmit data. An automated system makes individual bids at each link through which data is transmitted and can take into account additional criteria that can be specified as part of the bid information, including latency and routing requirements. Bid information is passed with data through the network. | 07-10-2014 |
20140195689 | SWAN: ACHIEVING HIGH UTILIZATION IN NETWORKS - Greater network utilization is implemented through dynamic network reconfiguration and allocation of network services and resources based on the data to be transferred and the consumer transferring it. A hierarchical system is utilized whereby requests from lower layers are aggregated before being provided to upper layers, and allocations received from upper layers are distributed to lower layers. To maximize network utilization, paths through the network are reconfigured by identifying specific types of packets that are to be flagged in a specific manner, and then by further identifying specific routing rules to be applied in the transmission of such packets. Network reconfiguration is performed on an incremental basis to avoid overloading a path, and capacity can be reserved along one or more paths to prevent such overloading. Background data is agnostic as to specific transmission times and is utilized to prevent overloading due to reconfiguration. | 07-10-2014 |
20150039756 | PROVIDING PER-APPLICATION RESOURCE USAGE INFORMATION - An environment is described in which a processing system provides application-level usage information to users. In one scenario, for example, the processing system may provide personal usage information to a user who is operating a user device. The personal usage information itemizes the amount of data (and/or other resources) that has been consumed by each application run by the user device. In another scenario, the processing system may provide expected usage information associated with at least one candidate application provided by a marketplace system. The expected usage information describes an expected consumption of data (and/or other resources) by the candidate application upon running the candidate application by the user device. The processing system can tailor the expected usage information that it sends to a particular user based on user profile data. The user profile data describes a manner in which users operate applications. | 02-05-2015 |