Patent application number | Description | Published |
20110213764 | Dynamic Search Health Monitoring - A method for monitoring search performance on a server computer includes determining the processing time for a plurality of operations related to a search on the server computer. The determined processing time for each of the plurality of operations is stored in a database. Aggregate processing times are determined for the plurality of operations and the aggregate processing times are stored in the database. | 09-01-2011 |
20110314127 | QUICK DEPLOY OF CONTENT - Non-administrators selectively move content from a source network to a destination network. The source network includes a deployment server that can deploy an item and its dependencies (if any) listed in a quick deploy list without having to get specific administrator authorization for the deployment. The deployment server can be configured to check the quick deploy list relatively frequently so that any items listed in the quick deploy list are quickly deployed. The listing of items in the quick deploy list is organized into “jobs” and identifies a path that indicates the source and destination for each job. A job is associated with one path and specifies one or more items of the source to be deployed and a schedule to deploy the specified item(s). The deployment server is configured by the administrator with permissions for specific authors to list items in the quick deploy list. | 12-22-2011 |
20120102024 | Result Types For Conditional Data Display - A set of user interfaces is provided to an administrator of a website. The user interfaces enable the administrator to configure search functionality of the website to present different types of search results differently. For instance, the administrator can use the user interfaces to configure the search functionality to present different properties for search results belonging to different result types. | 04-26-2012 |
20120254118 | RECOVERY OF TENANT DATA ACROSS TENANT MOVES - A history of locations of tenant data is maintained. The tenant data comprises data that is currently being used by the tenant and the corresponding backup data. When a tenant's data is changed from one location to another, a location and a time is stored within the history that may be accessed to determine a location of the tenant's data at a specified time. Different operations trigger a storing of a location/time within the history. Generally, an operation that changes a location of the tenant's data triggers the storing of the location within the history (e.g. upgrade of farm, move of tenant, adding a tenant, load balancing of the data, and the like). When tenant data is needed for an operation (e.g. restore), the history may be accessed to determine the location of the data. | 10-04-2012 |
20120310887 | AUTOMATIC CONFIGURATION OF A RECOVERY SERVICE - A secondary location is configured as a recovery service for a primary location of the service. The secondary location is maintained in a warm state that is configured to replace the primary location in a case of a failover. During normal operation, the secondary location is automatically updated to reflect a current state of the primary location that is actively servicing user load. Content changes to the primary location are automatically reflected to the secondary location. System changes applied to the primary location are automatically applied to the secondary location. For example, removing/adding machines, updating machine/role assignments, removing adding/database are automatically applied to the secondary location such that the secondary location substantially mirrors the primary location. After a failover to the secondary location, the secondary location becomes the primary location and begins to actively service the user load. | 12-06-2012 |
20120310912 | CRAWL FRESHNESS IN DISASTER DATA CENTER - Content that is stored at a secondary location for a service is crawled before it is placed in operation to assist in maintaining an up to date search index. The content that is crawled at the secondary location includes content that is obtained from the primary location of the service. When a crawler at the secondary location attempts to access content that is stored at the primary location, the crawler is directed to access the corresponding copy of the content that is stored at the secondary location instead of accessing the content at the primary location. The content may be crawled at the secondary location at different times, such as when the information is updated, according to a schedule, and the like. | 12-06-2012 |
20120311375 | REDIRECTING REQUESTS TO SECONDARY LOCATION DURING TEMPORARY OUTAGE - During an outage at a primary location for an online service that is temporary in duration (e.g. a “temporary outage”), requests are temporarily switched from the primary location to a secondary location for the online service. The temporary outage may be caused by many different reasons (e.g. power outage, planned maintenance, and the like). The secondary location may be configured as read only during the temporary outage such that users are still able to access their data during the temporary without causing changes to be made to the data. The requests to the primary location of the online service are automatically redirected to be handled by the secondary location. When the temporary outage ends, the requests are automatically switched back to the primary location. | 12-06-2012 |
20120311376 | RECOVERY SERVICE LOCATION FOR A SERVICE - A secondary location of a network acts as a recovery network for a primary location of the service. The secondary location is maintained in a warm state that is configured to replace the primary location in a case of a failover. During normal operation, the primary location actively services user load and performs backups that include full backups, incremental backups and transaction logs that are automatically replicated to the secondary location. Information is stored (e.g. time, retry count) that may be used to assist in determining when the backups are restored correctly at the secondary location. The backups are restored and the transaction logs are replayed at the secondary location to reflect changes (content and administrative) that are made to the primary location. After failover to the secondary location, the secondary location becomes the primary location and begins to actively service the user load. | 12-06-2012 |
20120311377 | REPLAYING JOBS AT A SECONDARY LOCATION OF A SERVICE - Jobs submitted to a primary location of a service within a period of time before and/or after a fail-over event are determined and are resubmitted to a secondary location of the service. For example, jobs that are submitted fifteen minutes before the fail-over event and jobs that are submitted to the primary network before the fail-over to the second location is completed are resubmitted at the secondary location. After the fail-over event occurs, the jobs are updated with the secondary network that is taking the place of the primary location of the service. A mapping of job input parameters (e.g. identifiers and/or secrets) from the primary location to the secondary location are used by the jobs when they are resubmitted to the secondary location. Each job determines what changes are to be made to the job request based on the job being resubmitted. | 12-06-2012 |
20130110795 | INHERITANCE OF RULES ACROSS HIERARCHICAL LEVELS | 05-02-2013 |
20130110813 | Routing Query Results | 05-02-2013 |
20130110860 | USER PIPELINE CONFIGURATION FOR RULE-BASED QUERY TRANSFORMATION, GENERATION AND RESULT DISPLAY | 05-02-2013 |
20150039931 | REPLAYING JOBS AT A SECONDARY LOCATION OF A SERVICE - Jobs submitted to a primary location of a service within a period of time before and/or after a fail-over event are determined and are resubmitted to a secondary location of the service. For example, jobs that are submitted fifteen minutes before the fail-over event and jobs that are submitted to the primary network before the fail-over to the second location is completed are resubmitted at the secondary location. After the fail-over event occurs, the jobs are updated with the secondary network that is taking the place of the primary location of the service. A mapping of job input parameters (e.g. identifiers and/or secrets) from the primary location to the secondary location are used by the jobs when they are resubmitted to the secondary location. Each job determines what changes are to be made to the job request based on the job being resubmitted. | 02-05-2015 |