Patent application number | Description | Published |
20100131487 | HTTP CACHE WITH URL REWRITING - URL rewriting is a common technique for allowing users to interact with internet resources using easy to remember and search engine friendly URLs. When URL rewriting involves conditions derived for sources other than the URL, inconsistencies in HTTP kernel cache and HTTP user output cache may arise. Methods and a system for rewriting a URL while preserving cache integrity are disclosed herein. Conditions used by a rule set to rewrite a URL may be determined as cache friendly conditions or cache unfriendly conditions. If cache unfriendly conditions exist, the HTTP kernel cache is disabled and the HTTP user output cache is varied based upon a key. If no cache unfriendly conditions exist, then the HTTP kernel cache is not disabled and the HTTP user output cache is not varied. A rule set is applied to the URL and a URL rewrite is performed to create a rewritten URL. | 05-27-2010 |
20110178973 | Web Content Rewriting, Including Responses - A content rewriting system is described herein that allows web site administrators to setup rewriting of web responses in an easy and efficient manner. The system provides a configuration schema and an efficient workflow that enables web administrators to easily setup rules to modify HTML or other content without having a high performance penalty or losing flexibility. The content rewriting system applies regular expressions or wildcard patterns to a response to locate and replace the content parts based on the rewriting logic expressed by outbound rewrite rules. The system parses an initial response generated by a web application, applies one or more outbound rules to rewrite the response, and provides the rewritten response to a client that submitted a request for the response. | 07-21-2011 |
20120185557 | INTER-CACHE COMMUNICATION USING HTTP RESOURCE - Servicing resource requests. A method includes at a first caching node, receiving a request for a first resource. The method further includes at the first caching node, determining at least one of health or load information of the first caching node. The method further includes generating a response to the request for the first resource, including generating a header to the response. The header is a custom header. The custom header includes at least one of health or load information. The method further includes sending the response to the request including sending the custom header with the at least one of health or load information. | 07-19-2012 |
20120203866 | PROXY-BASED CACHE CONTENT DISTRIBUTION AND AFFINITY - A distributed caching hierarchy that includes multiple edge routing servers, at least some of which receiving content requests from client computing systems via a load balancer. When receiving a content request, an edge routing server identifies which of the edge caching servers the requested content would be in if the requested content were to be cached within the edge caching servers, and distributes the content request to the identified edge caching server in a deterministic and predictable manner to increase the likelihood of increasing a cache-hit ratio. | 08-09-2012 |
20120323872 | TWO-PHASE EVICTION PROCESS FOR FILE HANDLE CACHES - Systems and methods are provided for evicting entries from a file handle cache. In accordance with certain embodiments, a two-stage eviction process is utilized. In a first stage of the eviction process, entries in the file entry cache are analyzed and marked for eviction while a shared lock is maintained on the file handle cache. The shared lock enables the file handle cache to be concurrently accessed by a content serving system to service content requests. In a second stage of the eviction process, entries in the file handle cache that are marked for eviction are removed while an exclusive lock is maintained on the file handle cache. The exclusive lock prevents the content serving system from concurrently accessing the file handle cache to service content requests. | 12-20-2012 |
20140095649 | PROXY-BASED CACHE CONTENT DISTRIBUTION AND AFFINITY - A distributed caching hierarchy that includes multiple edge routing servers, at least some of which receiving content requests from client computing systems via a load balancer. When receiving a content request, an edge routing server identifies which of the edge caching servers the requested content would be in if the requested content were to be cached within the edge caching servers, and distributes the content request to the identified edge caching server in a deterministic and predictable manner to increase the likelihood of increasing a cache-hit ratio. | 04-03-2014 |
20140115444 | Web Content Rewriting, Including Responses - A content rewriting system is described herein that allows web site administrators to setup rewriting of web responses in an easy and efficient manner. The system provides a configuration schema and an efficient workflow that enables web administrators to easily setup rules to modify HTML or other content without having a high performance penalty or losing flexibility. The content rewriting system applies regular expressions or wildcard patterns to a response to locate and replace the content parts based on the rewriting logic expressed by outbound rewrite rules. The system parses an initial response generated by a web application, applies one or more outbound rules to rewrite the response, and provides the rewritten response to a client that submitted a request for the response. | 04-24-2014 |
20150381757 | PROXY-BASED CACHE CONTENT DISTRIBUTION AND AFFINITY - A distributed caching hierarchy that includes multiple edge routing servers, at least some of which receiving content requests from client computing systems via a load balancer. When receiving a content request, an edge routing server identifies which of the edge caching servers the requested content would be in if the requested content were to be cached within the edge caching servers, and distributes the content request to the identified edge caching server in a deterministic and predictable manner to increase the likelihood of increasing a cache-hit ratio. | 12-31-2015 |