EdgeCast Networks, Inc. Patent applications |
Patent application number | Title | Published |
20160099849 | Dynamic Reconfigurable Client Side Beaconing - Some embodiments provide a monitoring system for a distributed platform. The monitoring system dynamically controls the monitoring and reporting function of client applications running on user devices using a plurality of configurations. The system distributes different configurations to different client applications to control the monitoring and reporting of each client application independently. The system distributes different configurations to the same client application at different times in order to dynamically reconfigure the reporting function of that client application. The system compiles metrics from the client applications in order to generate custom performance reports for different content providers and identify performance issues. In response to a detected issue, the system can automatically reconfigure affected client applications to better diagnose the issue with increased reporting. The system can also issue alerts for others to take corrective action. The system can also self-heal by commanding operation of the affected client applications. | 04-07-2016 |
20160080445 | Multi-Tenant Over-The-Top Multicast - Some embodiments provide a multi-tenant over-the-top multicast solution that integrates the per user stream customizability of unicast with the large scale streaming efficiencies of multicast. The solution involves an application, different multicast groups streaming an event with different customizations, and a manifest file or metadata identifying the different groups and customizations. The solution leverages the different multicast groups in order to provide different time shifts in the event stream, different quality level encodings of the event stream, and different secondary content to be included with a primary content stream. The application configured with the manifest file or metadata dynamically switches between the groups in order to customize the experience for a user or user device on which the application executes. Switching from multicast to unicast is also supported to supplement available customizations and for failover. | 03-17-2016 |
20160028598 | DISTRIBUTED PLATFORM TEST NETWORK - A test network is provided to test updates to configurations and resources of a distributed platform and to warm servers prior to their deployment in the distributed platform. The test network tests and warms using real-time production traffic of the distributed platform in a manner that does not impact users or performance of the distributed platform. At least one distributed platform caching server passes content requests that it receives to the test network using a connectionless protocol. The test network includes a test server that is loaded with any of a configuration or resource under test or whose cache is to be loaded prior to the server's deployment into the distributed platform. The test network also includes a replay server that receives the requests from the caching server, distributes the requests to the test server, and monitors the test server responses. | 01-28-2016 |
20150350365 | PROBABILITY BASED CACHING AND EVICTION - Some embodiments set forth probability based caching, whereby a probability value determines in part whether content identified by an incoming request should be cached or not. Some embodiments further set forth probability based eviction, whereby a probability value determines in part whether cached content should be evicted from the cache. Selection of the content for possible eviction can be based on recency and/or frequency of the content being requested. The probability values can be configured manually or automatically. Automatic configuration involves using a function to compute the probability values. In such scenarios, the probability values can be computed as a function of any of fairness, cost, content size, and content type as some examples. | 12-03-2015 |
20150341431 | CONTROL MESSAGE ROUTING WITHIN ANYCAST RELIANT PLATFORMS - Some embodiments provide a transport session discovery protocol that enables load balancing devices of an Anycast reliant distributed platform to route legacy control messages to destinations within the distributed platform that manage the sessions or connections implicated by the legacy control messages, even when the implicated sessions or connections cannot be directly identified from the control message headers. The modified load balancing operation as a result of the transport session discovery protocol involves identifying a message header mapping to an unrecognized session or connection, extracting session or connection identifying information and an error or condition from the message body, generating a new messaging construct to encapsulate the extracted information, and multicasting the messaging construct to other load balancing devices operating within a common point-of-presence. The load balancing device having knowledge of the server managing the implicated session then routes the error or condition to that server. | 11-26-2015 |
20150324380 | File System Enabling Fast Purges and File Access - Some embodiments provide a file system for purging content based on a single traversal of the file system that identifies the directory containing the content without performing a secondary traversal within the directory to target the operation to only the file that are associated with content such that other files contained in the directory are unaffected. The file system supplements traditional directory structures with file-level directories. Each file-level directory is created to contain a root file associated with particular content, different variants of the particular content, and supporting files. Consequently, the file system can complete an operation targeting particular content by performing that operation on the file-level directory containing the particular content, thereby eliminating the need to conduct a file-by-file traversal of the containing directory as a prerequisite to identifying the files associated with the particular content and performing the operation on the files individually. | 11-12-2015 |
20150249651 | PROVIDING LOCALIZED CONTENT DELIVERY WITH REMOTE TOKEN AUTHENTICATION - Some embodiments set forth systems and methods enabling a first network to use the resources of various second networks in order to localize delivery of the first network content from the various second networks in a secure manner. Some embodiments provide a token-based authentication scheme to ensure that any configured content access restrictions are effectuated at the first network and any of the second networks providing localized content delivery for the first network. The scheme involves a two phase user authentication, wherein the user is separately authenticated at the first network and the redirected to second network using either the same or different set of access restrictions. The first network exchanges a first encryption key with content providers for encrypting/decrypting the first access restriction and a second encryption key with a second network for encrypting/decrypting the second access restriction. | 09-03-2015 |
20150222528 | ADAPTING NETWORK CONTROL MESSAGING FOR ANYCAST RELIANT PLATFORMS - Some embodiments set forth a control message header rewriting methodology. Incoming packets are inspected to identify control messages. Each control message is then inspected to determine whether it originates from a client engaged in a session with a server or from an intermediary node along the path connecting the client and the server. The determination is predicated on a comparison of the addressing provided in the control message header and the addressing provided in the offending packet header, wherein the offending packet is the packet that triggers the intermediary node to issue the control message. If the addressing differs, the header addressing of control message is rewritten using the header addressing of the offending packet. Otherwise, a session table lookup is performed to identify which session the control message is directed to based in part on a hash of the control message header addressing. | 08-06-2015 |
20150100691 | Systems and Methods for Performing Localized Server-Side Monitoring in a Content Delivery Network - Some embodiments provide systems and methods for performing localized and real-time server-side network performance monitoring. These systems and methods leverage the distributed architecture of a content delivery network (CDN) so as to perform distributed monitoring with each Point-of-Presence of the CDN responsible for monitoring performance to a localized set of end users. These systems and methods also leverage existing traffic flows from a server to a particular end user in order to perform real-time server-side network performance monitoring without the injection of specialized monitoring packets and without active involvement of the end user in deriving the performance measurements. The performance measurements are then used to optimize delivery of existing and future traffic flows to the end user. | 04-09-2015 |
20150088964 | INSTANTANEOUS NON-BLOCKING CONTENT PURGING IN A DISTRIBUTED PLATFORM - Some embodiments provide instantaneous and non-blocking content purging across storage servers of a distributed platform. When a server receives a purge operation, it extracts an identifier from the purge operation. The server then generates a content purge pattern from the identifier and injects the pattern to its configuration. Instantaneous purging is then realized as the server averts access to any cached content identified by the pattern. The purging also occurs in a non-blocking fashion as the physical purge of the content occurs in-line with the server's cache miss operation. The content purge pattern causes the server to respond to a subsequently received content request with a cache miss, whereby the server retrieves the requested content from an origin source, serves the retrieved content to the requesting user, and replaces a previously cached copy of the content that is to be purged with the newly retrieved copy. | 03-26-2015 |
20140380454 | WHITE-LIST FIREWALL BASED ON THE DOCUMENT OBJECT MODEL - Some embodiments provide firewalls and methods for guarding against attacks by leveraging the Document Object Model (DOM). The firewall renders the DOM tree to produce a white-list rendering of the data which presents the non-executable elements of the data and, potentially, outputs of the executable elements of the data without the executable elements that could be used to carry a security threat. Some embodiments provide control over which nodes of the DOM tree are included in producing the white-list rendering. Specifically, a configuration file is specified to white-list various nodes from the DOM tree and the white-list rendering is produced by including the DOM tree nodes that are specified in the white-list of the configuration file while excluding those nodes that are not in the white-list. Some embodiments provide a hybrid firewall that executes a set of black-list rules over white-listed nodes of the DOM tree. | 12-25-2014 |
20140359092 | Capacity Exchange for the Open Content Delivery Network - Some embodiments provide a capacity exchange whereby capacity from different content delivery networks (CDNs) can be bought, sold, and traded. The capacity exchange is part of an “Open CDN” platform. The Open CDN platform federates the independent operation of CDNs and other operators of and service providers to distributed platforms participating in the Open CDN platform so that each participant can (1) dynamically scale its capacity without incurring additional infrastructure costs, (2) expand its service into previously untapped geographic regions without physically establishing points of presence (POPs) at those geographic regions, and (3) reduce sunk costs associated with unused capacity of already deployed infrastructure by selling that unused capacity to other participants that are in need of additional capacity. The Open CDN platform includes one or more APIs for facilitating intercommunication between the federation participants by performing configuration mapping, command interoperability, traffic management, and reporting aggregation. | 12-04-2014 |
20140355431 | Request Modification for Transparent Capacity Management in a Carrier Network - Some embodiments provide a capacity management agent that modifies content requests to adjust bandwidth consumption when streaming requested content from a content provider to a requesting user. The modifications include modifying a URL or header information of the request. The agent performs a process that receives a request for content of a content provider. | 12-04-2014 |
20140330924 | Efficient Cache Validation and Content Retrieval in a Content Delivery Network - Some embodiments provide systems and methods for validating cached content based on changes in the content instead of an expiration interval. One method involves caching content and a first checksum in response to a first request for that content. The caching produces a cached instance of the content representative of a form of the content at the time of caching. The first checksum identifies the cached instance. In response to receiving a second request for the content, the method submits a request for a second checksum representing a current instance of the content and a request for the current instance. Upon receiving the second checksum, the method serves the cached instance of the content when the first checksum matches the second checksum and serves the current instance of the content upon completion of the transfer of the current instance when the first checksum does not match the second checksum. | 11-06-2014 |
20140280803 | Optimized Content Distribution Based on Metrics Derived from the End User - Some embodiments provide systems and methods for determining a server of a distributed hosting system to optimally distribute content to an end user. The method includes identifying an IP address of the end user. Based on the IP address, a set of servers send packets to the end user to derive performance metrics. The performance metrics are used to determine a server from the set of servers that optimally distributes content to the end user. The method modifies a configuration for resolving end user requests such that the optimal server is identified to the end user when the end user requests content from the hosting system. Some embodiments determine the optimal server by providing downloadable content that is embedded with a monitoring tool. The monitoring tool causes the end user to derive performance metrics for the hosting system when downloading a particular object from a set of servers. | 09-18-2014 |
20140280480 | Dynamic Tag Management for Optimizing Content Delivery - Some embodiments provide a content delivery network (CDN) solution that affords the CDN control over those elements of customer content that are delivered by third parties. The CDN integrates a distributed set of monitoring agents. Each monitoring agent monitors the delivery performance of third parties to the region in which the agent operates. The CDN uses the performance monitoring information to dynamically manage the content tags to the third-party delivered elements of CDN-customer content. Specifically, a CDN server retrieves the parent page for requested CDN-customer content. The CDN server identifies the region from where the request originates and retrieves the logs from the monitoring agents monitoring from that region. The CDN server then modifies the base page by dynamically removing the tags to the third-party delivered elements that are reported in the monitoring agent logs as being unavailable, inaccessible, or underperforming in the identified region. | 09-18-2014 |
20140280479 | Dynamic Tag Management for Optimizing Content Delivery - Some embodiments provide a content delivery network (CDN) solution that affords the CDN control over those elements of customer content that are delivered by third parties. The CDN integrates a distributed set of monitoring agents. Each monitoring agent monitors the delivery performance of third parties to the region in which the agent operates. The CDN uses the performance monitoring information to dynamically manage the content tags to the third-party delivered elements of CDN-customer content. Specifically, a CDN server retrieves the parent page for requested CDN-customer content. The CDN server identifies the region from where the request originates and retrieves the logs from the monitoring agents monitoring from that region. The CDN server then modifies the base page by dynamically removing the tags to the third-party delivered elements that are reported in the monitoring agent logs as being unavailable, inaccessible, or underperforming in the identified region. | 09-18-2014 |
20140258523 | End-to-End Monitoring and Optimization of a Content Delivery Network using Anycast Routing - Some embodiments implement systems and methods to monitor CDN performance from the end user's perspective. To achieve full end-to-end monitoring, the systems and methods monitor the performance of Anycast routing for routing an end user to the optimal PoP of the CDN and monitor the performance of different routes that deliver content from the optimally selected PoP back to the end user. Systems and methods are also provided to leverage the monitored performance in order to automatically optimize the performance of the CDN by optimizing or modifying Anycast routing and by performing route grooming as needed to ensure that the optimal route is selected to deliver content to different sets of end users that request content from a particular PoP. | 09-11-2014 |
20140258240 | Scalable Content Streaming System with Server-Side Archiving - Some embodiments provide a scalable content streaming system that performs server-side archiving. The content streaming system includes a set of streaming server clusters, a remote storage server, and a set of distributed servers. Each streaming server cluster includes at least one streaming server and a local storage server. Each streaming server of a particular cluster distributes a content stream that is uploaded to the streaming server while also recording the content stream to the local storage server of that particular cluster. The remote storage server stores files from the local storage servers that are transferred when recording to those files is complete. The set of distributed servers distribute a live stream of content from a particular streaming server when the content is being uploaded to the particular streaming server and distribute a recorded stream of the content from the remote storage server when the upload is complete. | 09-11-2014 |
20140195686 | Optimized Consistent Request Distribution for Balanced Load Distribution in a Content Delivery Network - Some embodiments provide a proprietary 64-bit consistent distribution scheme that preserves the efficiencies of CARP while providing a significantly more balanced distribution of requests that is on par with schemes reliant on computationally expensive cryptographic hashes. The scheme performs hashing of requested URLs and identifiers of available servers over a 64-bit space while optimizing the hashing to remove computationally expensive operations. Some embodiments provide a variant of the scheme to provide a differentiated distribution on the basis of one or more differentiating factors. A first variant utilizes load factor values to adjust the resulting hashes and to produce a first distribution of differentiated content that varies from a second distribution of undifferentiated content. A second variant identifies requests for differentiated content from requests for undifferentiated content and utilizes a first scheme to distribute the differentiated content and a second altered scheme to distribute the undifferentiated content. | 07-10-2014 |
20140195600 | Network Connection Hand-off Using State Transformations - Some embodiments provide a director agent, a server agent, and a specialized hand-off protocol for improving scalability and resource usage within a server farm. A first network connection is established between a client and the director agent in order to receive a content request from the client from which to select a server from a set of servers that is responsible for hosting the requested content. A second network connection is established between the server agent that is associated with the selected server and a protocol stack of the selected server. The first network connection is handed-off to the server agent using the specialized hand-off protocol. The server agent performs network connection state parameter transformations between the two connections to create a network connection through which content can be passed from the selected server to the client without passing through the director. | 07-10-2014 |
20140143415 | Optimized Content Distribution Based on Metrics Derived from the End User - Some embodiments provide systems and methods for determining a server of a distributed hosting system to optimally distribute content to an end user. The method includes identifying an IP address of the end user. Based on the IP address, a set of servers send packets to the end user to derive performance metrics. The performance metrics are used to determine a server from the set of servers that optimally distributes content to the end user. The method modifies a configuration for resolving end user requests such that the optimal server is identified to the end user when the end user requests content from the hosting system. Some embodiments determine the optimal server by providing downloadable content that is embedded with a monitoring tool. The monitoring tool causes the end user to derive performance metrics for the hosting system when downloading a particular object from a set of servers. | 05-22-2014 |
20140136335 | Systems and Methods for Billing Content Providers for Designated Select Content Delivered Over a Data Network - Some embodiments provide multi-tenant billing systems and methods whereby data network usage fees that are associated with delivering select content of different content providers over a data network are passed to the content provider that originates or is the source for that select content instead of the end users that request and receive the select content. A deep packet inspection server or a billing server may be configured with a list of identifiers to distinguish between the network usage that is related to the delivery of the designated free content from the network usage that is not related to the delivery of the designated free content. The network usage can be distinguished by filtering or processing usage logs that track the network usage using the configured list of identifiers. Such filtering or processing can be done in real-time or in a delayed manner in a standalone or integrated system. | 05-15-2014 |
20140122711 | Systems and Methods for Performing Localized Server-Side Monitoring in a Content Delivery Network - Some embodiments provide systems and methods for performing localized and real-time server-side network performance monitoring. These systems and methods leverage the distributed architecture of a content delivery network (CDN) so as to perform distributed monitoring with each Point-of-Presence of the CDN responsible for monitoring performance to a localized set of end users. These systems and methods also leverage existing traffic flows from a server to a particular end user in order to perform real-time server-side network performance monitoring without the injection of specialized monitoring packets and without active involvement of the end user in deriving the performance measurements. The performance measurements are then used to optimize delivery of existing and future traffic flows to the end user. | 05-01-2014 |
20140095804 | Efficient Cache Validation and Content Retrieval in a Content Delivery Network - Some embodiments provide systems and methods for validating cached content based on changes in the content instead of an expiration interval. One method involves caching content and a first checksum in response to a first request for that content. The caching produces a cached instance of the content representative of a form of the content at the time of caching. The first checksum identifies the cached instance. In response to receiving a second request for the content, the method submits a request for a second checksum representing a current instance of the content and a request for the current instance. Upon receiving the second checksum, the method serves the cached instance of the content when the first checksum matches the second checksum and serves the current instance of the content upon completion of the transfer of the current instance when the first checksum does not match the second checksum. | 04-03-2014 |
20140095592 | Network Connection Hand-Off and Hand-Back - Some embodiments provide a director agent, a server agent, and a specialized hand-off protocol for improving scalability and resource usage within a server farm. A first network connection is established between a client and the director agent in order to receive a content request from the client from which to select a server from a set of servers that is responsible for hosting the requested content. A second network connection is established between the server agent that is associated with the selected server and a protocol stack of the selected server. The first network connection is handed-off to the server agent using the specialized hand-off protocol. The server agent performs network connection state parameter transformations between the two connections to create a network connection through which content can be passed from the selected server to the client without passing through the director. | 04-03-2014 |
20140082126 | Sandboxing Content Optimization at the Network Edge - Some embodiments provide systems and methods for sandboxing content optimization to occur entirely within a network edge or PoP of a CDN. Some embodiments pass a first request for a first URL to a first back-end at the network edge that is configured to cache an optimized instance of the particular object. When the optimized instance of the particular object is not cached at the first back-end, a second request is issued for a second URL identifying a non-optimized instance of the particular object. The second request resolves internally within the network edge to a second back-end that is configured to cache the non-optimized object. The non-optimized object from the second back-end is optimized and passed to the first back-end. The first back-end caches the optimized instance of the non-optimized object and serves the optimized instance to a requesting end user. | 03-20-2014 |
20140047085 | Configuration Management Repository for a Distributed Platform - Some embodiments provide a repository that manages configurations for a distributed platform and that automatedly configures servers of the distributed platform with different hierarchical sets of configurations while ensuring integrity and consistency across the servers and in the repository. In some embodiment, the repository includes a data store that stores configurations for a first set of servers that are operated by a first service provider and a second set of servers that are operated by a second service provider. The data store also identifies different sets of configurations to deploy to different sets of servers from the first and second sets of servers. The repository also includes a function processor to automatedly deploy the different sets of configurations to the different sets of servers and to perform functions for updating the configurations in a manner that ensures integrity and consistency. | 02-13-2014 |
20140043970 | Bandwiddth Modification for Transparent Capacity Management in a Carrier Network - Some embodiments provide a capacity management agent that modifies bandwidth that is allocated between an end user and a carrier network by caching requested content that is streamed at a first rate and then providing the cached content to the end user through the carrier network at a second rate. The agent performs a process that includes receiving data intended for a service region of the carrier network from an external data network. The process identifies resource availability at the service region. Next, the process passes the data to the service region at the first rate when the resource availability at the service region is not less than a threshold amount and caches the data for passing to the service region at the second rate that consumes fewer carrier network resource than the first rate when the resource availability at the service region is less than the threshold amount. | 02-13-2014 |
20130339519 | Systems and Methods for Performing Localized Server-Side Monitoring in a Content Delivery Network - Some embodiments provide systems and methods for performing localized and real-time server-side network performance monitoring. These systems and methods leverage the distributed architecture of a content delivery network (CDN) so as to perform distributed monitoring with each Point-of-Presence of the CDN responsible for monitoring performance to a localized set of end users. These systems and methods also leverage existing traffic flows from a server to a particular end user in order to perform real-time server-side network performance monitoring without the injection of specialized monitoring packets and without active involvement of the end user in deriving the performance measurements. The performance measurements are then used to optimize delivery of existing and future traffic flows to the end user. | 12-19-2013 |
20130265873 | Request Modification for Transparent Capacity Management in a Carrier Network - Some embodiments provide a capacity management agent that modifies content requests to adjust bandwidth consumption when streaming requested content from a content provider to a requesting user. The modifications include modifying a URL or header information of the request. The agent performs a process that receives a request for content of a content provider. The process identifies a parameter of the carrier network and modifies the request when the parameter satisfies a threshold. The process passes the request to the content provider and the content provider provides content that consumes a first set of resources in response to an unmodified request and a second set of resources in response to a modified request. When the parameter identifies congestion, the first set of resources is greater than the second set of resources. When the condition parameter identifies underutilization, the first set of resources is less than the second set of resources. | 10-10-2013 |
20130227051 | Multi-Layer Multi-Hit Caching for Long Tail Content - Some embodiments provide an optimized multi-hit caching technique that minimizes the performance impact associated with caching of long-tail content while retaining much of the efficiency and minimal overhead associated with first hit caching in determining when to cache content. The optimized multi-hit caching utilizes a modified bloom filter implementation that performs flushing and state rolling to delete indices representing stale content from a bit array used to track hit counts without affecting identification of other content that may be represented with indices overlapping with those representing the stale content. Specifically, a copy of the bit array is stored prior to flushing the bit array so as to avoid losing track of previously requested and cached content when flushing the bit arrays and the flushing is performed to remove the bit indices representing stale content from the bit array and to minimize the possibility of a false positive. | 08-29-2013 |
20130179529 | Optimizing Multi-Hit Caching for Long Tail Content - Some embodiments provide an optimized multi-hit caching technique that minimizes the performance impact associated with caching of long-tail content while retaining much of the efficiency and minimal overhead associated with first hit caching in determining when to cache content. The optimized multi-hit caching utilizes a modified bloom filter implementation that performs flushing and state rolling to delete indices representing stale content from a bit array used to track hit counts without affecting identification of other content that may be represented with indices overlapping with those representing the stale content. Specifically, a copy of the bit array is stored prior to flushing the bit array so as to avoid losing track of previously requested and cached content when flushing the bit arras and the flushing is performed to remove the bit indices representing stale content from the bit array and to minimize the possibility of a false positive. | 07-11-2013 |
20130046883 | End-to-End Content Delivery Network Incorporating Independently Operated Transparent Caches and Proxy Caches - Some embodiments provide an end-to-end federated CDN solution that assimilates a transparent caching server that is operated by a transparent caching server operator into a CDN that independently operates CDN caching servers. Specifically, the logs from the transparent caching server are assimilated into the CDN by aggregating the logs from the transparent caching server and processing the transparent caching server logs to identify network usage for content of a CDN content provider customer that is delivered by the transparent caching server. The network usage is then combined with the network usage that tracked by the CDN caching servers in order to provide comprehensive report metrics for the content provider customer and to bill the content provider customer for all network usage related to delivering the content provider customer's content irrespective of whether the content was delivered by a transparent caching server or a CDN caching server. | 02-21-2013 |
20130046807 | Systems and Methods for Invoking Commands Across a Federation - Some embodiments provide different frameworks for seamlessly issuing and executing commands across servers of different federation participants. Each framework facilitates issuance and execution of a command that originates from a first federation participant and that is intended for execution at servers of a second federation participant. In some embodiments, a framework implements a method for enabling command interoperability between distributed platforms that each operate a set of servers on behalf of content providers. The method involves receiving a command targeting a particular configuration that a first distributed platform deploys to a server that is operated by a second distributed platform. The method identifies the server of the second distributed platform that is deployed with the particular configuration. The method communicably couples to a command invocation system of the second distributed platform and issues the command to the command invocation system for issuance of the command to the identified server. | 02-21-2013 |
20130046664 | End-to-End Content Delivery Network Incorporating Independently Operated Transparent Caches and Proxy Caches - Some embodiments provide an end-to-end federated CDN solution that assimilates a transparent caching server that is operated by a transparent caching server operator into a CDN that independently operates CDN caching servers. Specifically, the logs from the transparent caching server are assimilated into the CDN by aggregating the logs from the transparent caching server and processing the transparent caching server logs to identify network usage for content of a CDN content provider customer that is delivered by the transparent caching server. The network usage is then combined with the network usage that tracked by the CDN caching servers in order to provide comprehensive report metrics for the content provider customer and to bill the content provider customer for all network usage related to delivering the content provider customer's content irrespective of whether the content was delivered by a transparent caching server or a CDN caching server. | 02-21-2013 |
20120215915 | Global Load Balancing on a Content Delivery Network - The invention relates to systems and methods of global load balancing in a content delivery network having a plurality of edge servers which may be distributed across multiple geographic locations. According to one aspect of the invention, a global load balancing system includes a first load balancing server for receiving a packet requesting content to be delivered to a client, selecting one of the plurality of edge servers to deliver the requested content to the client, and forwarding the packet across a network connection to a second load balancing server, which forwards the packet to the selected edge server. The selected edge server, in response to receiving the packet, sends across a network connection the requested content with an address for direct delivery to the client, thereby allowing the requested content to be delivered to the client while bypassing a return path through the first load balancing server. | 08-23-2012 |
20120198045 | GLOBAL LOAD BALANCING ON A CONTENT DELIVERY NETWORK - The invention relates to systems and methods of global load balancing in a content delivery network having a plurality of edge servers which may be distributed across multiple geographic locations. According to one aspect of the invention, a global load balancing system includes a first load balancing server for receiving a packet requesting content to be delivered to a client, selecting one of the plurality of edge servers to deliver the requested content to the client, and forwarding the packet across a network connection to a second load balancing server, which forwards the packet to the selected edge server. The selected edge server, in response to receiving the packet, sends across a network connection the requested content with an address for direct delivery to the client, thereby allowing the requested content to be delivered to the client while bypassing a return path through the first load balancing server. | 08-02-2012 |
20110087844 | CONTENT NETWORK GLOBAL REPLACEMENT POLICY - This invention is related to content delivery systems and methods. In one aspect of the invention, a content provider controls a replacement process operating at an edge server. The edge server services content providers and has a data store for storing content associated with respective ones of the content providers. A content provider sets a replacement policy at the edge server that controls the movement of content associated with the content provider, into and out of the data store. In another aspect of the invention, a content delivery system includes a content server storing content files, an edge server having cache memory for storing content files, and a replacement policy module for managing content stored within the cache memory. The replacement policy module can store portions of the content files at the content server within the cache memory, as a function of a replacement policy set by a content owner. | 04-14-2011 |
20100275125 | CONTENT NETWORK GLOBAL REPLACEMENT POLICY - This invention is related to content delivery systems and methods. In one aspect of the invention, a content provider controls a replacement process operating at an edge server. The edge server services content providers and has a data store for storing content associated with respective ones of the content providers. A content provider sets a replacement policy at the edge server that controls the movement of content associated with the content provider, into and out of the data store. In another aspect of the invention, a content delivery system includes a content server storing content files, an edge server having cache memory for storing content files, and a replacement policy module for managing content stored within the cache memory. The replacement policy module can store portions of the content files at the content server within the cache memory, as a function of a replacement policy set by a content owner. | 10-28-2010 |
20100036954 | Global load balancing on a content delivery network - The invention relates to systems and methods of global load balancing in a content delivery network having a plurality of edge servers which may be distributed across multiple geographic locations. According to one aspect of the invention, a global load balancing system includes a first load balancing server for receiving a packet requesting content to be delivered to a client, selecting one of the plurality of edge servers to deliver the requested content to the client, and forwarding the packet across a network connection to a second load balancing server, which forwards the packet to the selected edge server. The selected edge server, in response to receiving the packet, sends across a network connection the requested content with an address for direct delivery to the client, thereby allowing the requested content to be delivered to the client while bypassing a return path through the first load balancing server. | 02-11-2010 |
20090070533 | Content network global replacement policy - This invention is related to content delivery systems and methods. In one aspect of the invention, a content provider controls a replacement process operating at an edge server. The edge server services content providers and has a data store for storing content associated with respective ones of the content providers. A content provider sets a replacement policy at the edge server that controls the movement of content associated with the content provider, into and out of the data store. In another aspect of the invention, a content delivery system includes a content server storing content files, an edge server having cache memory for storing content files, and a replacement policy module for managing content stored within the cache memory. The replacement policy module can store portions of the content files at the content server within the cache memory, as a function of a replacement policy set by a content owner. | 03-12-2009 |