Patent application title: Methods, Systems, and Software for Identifying Functional Bio-Molecules
Richard John Fox (Redwood City, CA, US)
IPC8 Class: AC40B3002FI
Class name: Combinatorial chemistry technology: method, library, apparatus method of screening a library in silico screening
Publication date: 2010-01-07
Patent application number: 20100004135
Patent application title: Methods, Systems, and Software for Identifying Functional Bio-Molecules
Richard John Fox
Weaver Austin Villeneuve & Sampson LLP
Origin: OAKLAND, CA US
IPC8 Class: AC40B3002FI
Patent application number: 20100004135
The present invention generally relates to methods of rapidly and
efficiently searching biologically-related data space. More specifically,
the invention includes methods of identifying bio-molecules with desired
properties, or which are most suitable for acquiring such properties,
from complex bio-molecule libraries or sets of such libraries. The
invention also provides methods of modeling sequence-activity
relationships. As many of the methods are computer-implemented, the
invention additionally provides digital systems and software for
performing these methods.
1. A method for identifying amino acid residues for variation in a protein
variant library in order to affect a desired activity, said method
comprising:(a) receiving data characterizing a training set of a protein
variant library,wherein the data provides activity and sequence
information for each protein variant in the training set;(b) from the
data, developing a sequence-activity model that predicts activity as a
function of amino acid residue type and corresponding position in a
protein sequence,wherein the sequence-activity model includes one or more
non-linear terms, each representing an interaction between two or more
amino acid residues in the protein sequence; and(c) using the
sequence-activity model to identify one or more amino acid residues at
specific positions for variation to impact the desired activity.
CROSS-REFERENCE TO RELATED APPLICATIONS
This application is a continuation of U.S. patent application Ser. No. 11/981,578 [Atty Ref: MXGNP004X2C1], filed Oct. 30, 2007, naming Richard John Fox as inventor, which is a continuation of U.S. patent application Ser. No. 10/874,802 [Atty Ref: MXGNP004X2], filed Jun. 22, 2004, naming Richard John Fox as inventor, which is a continuation-in-part of U.S. patent application Ser. No. 10/629,351 [Atty Ref: MXGNP004X1], filed Jul. 29, 2003, naming Richard John Fox as inventor, which is a continuation-in-part of U.S. patent application Ser. No. 10/379,378 [Atty Ref: MXGNP004], filed Mar. 3, 2003, naming Gustaffson et al. as inventors which in turn claims priority under 35 USC 119(e) from U.S. Provisional Application No. 60/360,982, filed Mar. 1, 2002. Each of these documents is incorporated herein by reference in its entirety and for all purposes.
The present invention relates to the fields of molecular biology, molecular evolution, bioinformatics, and digital systems. More specifically, the invention relates to methods for computationally predicting the activity of a biomolecule. Systems, including digital systems, and system software for performing these methods are also provided. Methods of the present invention have utility in the optimization of proteins for industrial and therapeutic use.
Protein design has long been known to be a difficult task if for no other reason than the combinatorial explosion of possible molecules that constitute searchable sequence space. The sequence space of proteins is immense and is impossible to explore exhaustively. Because of this complexity, many approximate methods have been used to design better proteins; chief among them is the method of directed evolution. Directed evolution of proteins is today dominated by various high throughput screening and recombination formats, often performed iteratively.
In parallel, various computational techniques have been proposed for exploring sequence-activity space. Relatively speaking, these techniques are in their infancy and significant advances are still needed. Accordingly, new ways to efficiently search sequence space to identify functional proteins would be highly desirable.
The present invention provides techniques for generating and using models that employ non-linear terms, particularly terms that account for interactions between two or more residues in the sequence. These non-linear terms may be "cross product" terms that involve multiplication of two or more variables, each representing the presence (or absence) of the residues participating in the interaction. In some embodiments, the invention involves techniques for selecting the non-linear terms that best describe the activity of the sequence. Note that there are often far more possible non-linear interaction terms than there are true interactions between residues. Hence, to avoid overfitting, only a limited number of non-linear are typically employed and those employed should reflect interactions that affect activity.
One aspect of the invention provides a method for identifying amino acid residues for variation in a protein variant library. This method may be characterized by the following operations: (a) receiving data characterizing a training set of a protein variant library, (b) from the data, developing a sequence-activity model that predicts activity as a function of amino acid residue type and corresponding position in a protein sequence; and (c) using the sequence-activity model to identify one or more amino acid residues at specific positions for variation to impact the desired activity. The sequence-activity model includes one or more non-linear terms, each of which represents an interaction between two or more amino acid residues in the protein sequence. The training set data provides activity and sequence information for each protein variant in the training set.
The protein variant library may include proteins from various sources. In one example, the members include naturally occurring proteins such as those encoded by members of a single gene family. In another example, the members include proteins obtained by using a recombination-based diversity generation mechanism. For example, DNA fragmentation-mediated recombination, synthetic oligonucleotide-mediated recombination or a combination thereof may be performed on nucleic acids encoding all or part of one or more naturally occurring parent proteins for this purpose. In still another example, the members are obtained by performing DOE to identify the systematically varied sequences.
In some embodiments, at least one non-linear term is a cross-product term containing a product of one variable representing the presence of one interacting residue and another variable representing the presence of another interacting residue. The form of the sequence-activity model may be a sum of at least one cross-product term and one or more linear terms, with each of the linear terms representing the presence of a variable residue in the training set. The cross-product terms may be selected from a group of potential cross-product terms by various techniques including, for example, running a genetic algorithm to select the cross-product terms based upon the predictive ability of various models employing different cross-product terms.
The sequence activity model may be produced from the training set by many different techniques. In a preferred embodiment, the model is a regression model such as a partial least squares model or a principal component regression model. In another example, the model is a neural network.
In some embodiments, the method also includes (d) using the sequence activity model to identify one or more amino acid residues that are to remain fixed (as opposed to being varied) in new protein variant library.
Using the sequence activity model to identify residues for fixing or variation may involve any of many different possible analytical techniques. In some cases, a "reference sequence" is used to define the variations. Such sequence may be one predicted by the model to have a highest value (or one of the highest values) of the desired activity. In another case, the reference sequence may be that of a member of the original protein variant library. From the reference sequence, the method may select subsequences for effecting the variations. In addition or alternatively, the sequence activity model ranks residue positions (or specific residues at certain positions) in order of impact on the desired activity.
One goal of the method may be to generate a new protein variant library. As part of this process, the method may identify sequences that are to be used for generating this new library. Such sequences include variations on the residues identified in (c) above or are precursors used to subsequently introduce such variations. The sequences may be modified by performing mutagenesis or a recombination-based diversity generation mechanism to generate the new library of protein variants. This may form part of a directed evolution procedure. The new library may also be used in developing a new sequence activity model. The new protein variant library is analyzed to assess effects on a particular activity such as stability, catalytic activity, therapeutic activity, resistance to a pathogen or toxin, toxicity, etc.
In some embodiments, the method involves selecting one or more members of the new protein variant library for production. One or more of these may then be synthesized and/or expressed in an expression system. In a specific embodiment, the method continues in the following manner: (i) providing an expression system from which a selected member of the new protein variant library can be expressed; and (ii) expressing the selected member of the new protein variant library.
In some embodiments, rather than use amino acid sequences, the methods employ nucleotide sequences to generate the models and predict activity. Variations in groups of nucleotides, e.g., codons, affect the activity of peptides encoded by the nucleotide sequences. In some embodiments, the model may provide a bias for codons that are preferentially expressed (compared to other codons encoding the same amino acid) depending upon the host employed to express the peptide.
Yet another aspect of the invention pertains to apparatus and computer program products including machine-readable media on which are provided program instructions and/or arrangements of data for implementing the methods and software systems described above. Frequently, the program instructions are provided as code for performing certain method operations. Data, if employed to implement features of this invention, may be provided as data structures, database tables, data objects, or other appropriate arrangements of specified information. Any of the methods or systems of this invention may be represented, in whole or in part, as such program instructions and/or data provided on machine-readable media.
These and other features of the present invention will be described in more detail below in the detailed description of the invention and in conjunction with the following figures.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a flow chart depicting a sequence of operations, including identifying particular residues for variation, that may be used to generate one or more generations of protein variant libraries.
FIG. 2 is a flow chart depicting a genetic algorithm for selecting non-linear cross product terms in accordance with an embodiment of this invention.
FIGS. 3A-3F are graphs showing examples of this invention in which the predictive capabilities of certain linear and non-linear models are compared.
FIG. 4 is a flow chart depicting a bootstrap p-value method of generating protein variant libraries in accordance with an embodiment of this invention.
FIG. 5 is a schematic of an example digital device.
DETAILED DISCUSSION OF THE INVENTION
Before describing the present invention in detail, it is to be understood that this invention is not limited to particular sequences, compositions, algorithms, or systems, which can, of course vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting. As used in this specification and appended claims, the singular forms "a", "an", and "the" include plural referents unless the content and context clearly dictates otherwise. Thus, for example, reference to "a device" includes a combination of two or more such devices, and the like. Unless indicated otherwise, an "or" conjunction is intended to be used in its correct sense as a Boolean logical operator, encompassing both the selection of features in the alternative (A or B, where the selection of A is mutually exclusive from B) and the selection of features in conjunction (A or B, where both A and B are selected).
The following definitions and those included throughout this disclosure supplement those known to persons of skill in the art.
A "bio-molecule" refers to a molecule that is generally found in a biological organism. Preferred biological molecules include biological macromolecules that are typically polymeric in nature being composed of multiple subunits (i.e., "biopolymers"). Typical bio-molecules include, but are not limited to, molecules that share some structural features with naturally occurring polymers such as RNAs (formed from nucleotide subunits), DNAs (formed from nucleotide subunits), and polypeptides (formed from amino acid subunits), including, e.g., RNAs, RNA analogues, DNAs, DNA analogues, polypeptides, polypeptide analogues, peptide nucleic acids (PNAs), combinations of RNA and DNA (e.g., chimeraplasts), or the like. Bio-molecules also include, e.g., lipids, carbohydrates, or other organic molecules that are made by one or more genetically encodable molecules (e.g., one or more enzymes or enzyme pathways) or the like.
The term "nucleic acid" refers to deoxyribonucleotides or ribonucleotides and polymers (e.g., oligonucleotides, polynucleotides, etc.) thereof in either single- or double-stranded form. Unless specifically limited, the term encompasses nucleic acids containing known analogs of natural nucleotides that have similar binding properties as the reference nucleic acid and are metabolized in a manner similar to naturally occurring nucleotides. Unless otherwise indicated, a particular nucleic acid sequence also implicitly encompasses conservatively modified variants thereof (e.g., degenerate codon substitutions) and complementary sequences as well as the sequence explicitly indicated. Specifically, degenerate codon substitutions may be achieved by generating sequences in which the third position of one or more selected (or all) codons is substituted with mixed-base and/or deoxyinosine residues (Batzer et al. (1991) Nucleic Acid Res. 19:5081; Ohtsuka et al. (1985) J. Biol. Chem. 260:2605-2608; Rossolini et al. (1994) Mol. Cell. Probes 8:91-98). The term nucleic acid is used interchangeably with, e.g., oligonucleotide, polynucleotide, cDNA, and mRNA.
A "nucleic acid sequence" refers to the order and identity of the nucleotides comprising a nucleic acid.
A "polynucleotide" is a polymer of nucleotides (A, C, T, U, G, etc. or naturally occurring or artificial nucleotide analogues) or a character string representing a polymer of nucleotides, depending on context. Either the given nucleic acid or the complementary nucleic acid can be determined from any specified polynucleotide sequence.
The terms "polypeptide" and "protein" are used interchangeably herein to refer to a polymer of amino acid residues. Typically, the polymer has at least about 30 amino acid residues, and usually at least about 50 amino acid residues. More typically, they contain at least about 100 amino acid residues. The terms apply to amino acid polymers in which one or more amino acid residues are analogs, derivatives or mimetics of corresponding naturally occurring amino acids, as well as to naturally occurring amino acid polymers. For example, polypeptides can be modified or derivatized, e.g., by the addition of carbohydrate residues to form glycoproteins. The terms "polypeptide," and "protein" include glycoproteins, as well as non-glycoproteins.
A "motif" refers to a pattern of subunits in or among biological molecules. For example, the motif can refer to a subunit pattern of the unencoded biological molecule or to a subunit pattern of an encoded representation of a biological molecule.
"Screening" refers to the process in which one or more properties of one or more bio-molecule is determined. For example, typical screening processes include those in which one or more properties of one or more members of one or more libraries is/are determined.
The term "covariation" refers to the correlated variation of two or more variables (e.g., amino acids in a polypeptide, etc.).
"Directed evolution" or "artificial evolution" refers to a process of artificially changing a character string by artificial selection, recombination, or other manipulation, i.e., which occurs in a reproductive population in which there are (1) varieties of individuals, with some varieties being (2) heritable, of which some varieties (3) differ in fitness (reproductive success determined by outcome of selection for a predetermined property (desired characteristic). The reproductive population can be, e.g., a physical population or a virtual population in a computer system.
A "data structure" refers to the organization and optionally associated device for the storage of information, typically multiple "pieces" of information. The data structure can be a simple recordation of the information (e.g., a list) or the data structure can contain additional information (e.g., annotations) regarding the information contained therein, can establish relationships between the various "members" (i.e., information "pieces") of the data structure, and can provide pointers or links to resources external to the data structure. The data structure can be intangible but is rendered tangible when stored or represented in a tangible medium (e.g., paper, computer readable medium, etc.). The data structure can represent various information architectures including, but not limited to simple lists, linked lists, indexed lists, data tables, indexes, hash indices, flat file databases, relational databases, local databases, distributed databases, thin client databases, and the like. In preferred embodiments, the data structure provides fields sufficient for the storage of one or more character strings. The data structure is optionally organized to permit alignment of the character strings and, optionally, to store information regarding the alignment and/or string similarities and/or string differences. In one embodiment, this information is in the form of alignment "scores" (e.g., similarity indices) and/or alignment maps showing individual subunit (e.g., nucleotide in the case of nucleic acid) alignments. The term "encoded character string" refers to a representation of a biological molecule that preserves desired sequence/structural information regarding that molecule. As noted throughout, non-sequence properties of bio-molecules can be stored in a data structure and alignments of such non-sequence properties, in a manner analogous to sequence based alignment can be practiced.
A "library" or "population" refers to a collection of at least two different molecules, character strings, and/or models, such as nucleic acid sequences (e.g., genes, oligonucleotides, etc.) or expression products (e.g., enzymes) therefrom. A library or population generally includes a number of different molecules. For example, a library or population typically includes at least about 10 different molecules. Large libraries typically include at least about 100 different molecules, more typically at least about 1000 different molecules. For some applications, the library includes at least about 10000 or more different molecules.
"Systematic variance" refers to different descriptors of an item or set of items being changed in different combinations.
"Systematically varied data" refers to data produced, derived, or resulting from different descriptors of an item or set of items being changed in different combinations. Many different descriptors can be changed at the same time, but in different combinations. For example, activity data gathered from polypeptides in which combinations of amino acids have been changed is systematically varied data.
The terms "sequence" and "character strings" are used interchangeably herein to refer to the order and identity of amino acid residues in a protein (i.e., a protein sequence or protein character string) or to the order and identity of nucleotides in a nucleic acid (i.e., a nucleic acid sequence or nucleic acid character string).
II. Generating Improved Protein Variant Libraries
In accordance with the present invention, various methods are provided for generating new protein variant libraries that can be used to explore protein sequence and activity space. A feature of many such methods is a procedure for identifying amino acid residues in a protein sequence that are predicted to impact a desired activity. As one example, such procedure includes the following operations: (a) receiving data characterizing a training set of protein variants, wherein the data provides activity and sequence information for each protein variant in the training set; (b) from the data, developing a sequence activity model that predicts activity as a function of amino acid residue type and corresponding position in the sequence (preferably the model includes one or more non-linear terms, each representing an interaction between two or more amino acid residues); and (c) using the sequence activity model to identify one or more amino acid residues at specific positions in one or more protein variants that are to be varied in order to impact the desired activity.
FIG. 1 presents a flow chart showing one application of the present invention. It presents various operations that may be performed in the order depicted or in some other order. As shown, a process 01 begins at a block 03 with receipt of data describing a training set comprising residue sequences for a protein variant library. In other words, the training set data is derived from a protein variant library. Typically that data will include, for each protein in the library, a complete or partial residue sequence together with an activity value. In some cases, multiple types of activities (e.g., rate constant data and thermal stability data) are provided together in the training set.
In many embodiments, the individual members of the protein variant library represent a wide range of sequences and activities. This allows one to generate a sequence-activity model having applicability over a broad region of sequence space. Techniques for generating such diverse libraries include systematic variation of protein sequences and directed evolution techniques. Both of these are described in more detail elsewhere herein. Note however that it is often desirable to generate models from gene sequences representing a particular gene family; e.g., a particular kinase found in multiple species. As most residues will be identical across all members of the family, the model describes only those residues that vary. Thus statistical models based on such relatively small training sets, compared to the set of all possible variants, are valid in a local sense. The goal is not to find a global fitness function since that is beyond the capacity (and often the need) of the systems under consideration.
Activity data may be obtained by assays or screens appropriately designed to measure activity magnitudes. Such techniques are well known and are not central to this invention. The principles for designing appropriate assays or screens are widely understood. Techniques for obtaining protein sequences are also well known and are not central to this invention. The activity used with this invention may be protein stability (e.g., thermal stability). However, many important embodiments consider other activities such as catalytic activity, resistance to pathogens and/or toxins, therapeutic activity, toxicity, and the like.
After the training set data has been generated or acquired, the process uses it to generate a sequence-activity model that predicts activity as a function of sequence information. See block 05. Such model is a non-linear expression, algorithm or other tool that predicts the relative activity of a particular protein when provided with sequence information for that protein. In other words, protein sequence information is input and an activity prediction is output. For many embodiments of this invention, the model can also rank the contribution of various residues to activity. Methods of generating such models, which all fall under the rubric of machine learning, (e.g., partial least squares regression (PLS), principal component regression (PCR), and multiple linear regression (MLR)) will be discussed below, along with the format of the independent variables (sequence information), the format of the dependent variable(s) (activity), and the form of the model itself (e.g., a linear first order expression).
A model generated at block 05 is employed to identify multiple residue positions (e.g., position 35) or specific residue values (e.g. glutamine at position 35) that are predicted to impact activity. See block 07. In addition to identifying such positions, it may "rank" the residue positions or residue values based on their contributions to activity. For example, the model may predict that glutamine at position 35 has the most pronounced, positive effect on activity, phenylalanine at position 208 has the second most pronounced, positive effect, and so on. In a specific approach described below, PLS or PCR regression coefficients are employed to rank the importance of specific residues. In another specific approach, a PLS load matrix is employed to rank the importance of specific residue positions.
After the process has identified residues that impact activity, some of them are selected for variation as indicated at a block 09. This is done for the purpose of exploring sequence space. Residues are selected using any of a number of different selection protocols, some of which will be described below. In one example, specific residues predicted to have the most beneficial impact on activity are preserved (i.e., not varied). A certain number of other residues predicted to have a lesser impact are, however, selected for variation. In another example, the residue positions found to have the biggest impact on activity are selected for variation, but only if they are found to vary in high performing members of the training set. For example, if the model predicts that residue position 197 has the biggest impact on activity, but all or most of the proteins with high activity have leucine at this position, then position 197 would not be selected for variation in this approach. In other words, all or most proteins in a next generation library would have leucine at position 197. However, if some "good" proteins had valine at this position and others had leucine, then the process would choose to vary the amino acid at this position. In some cases, it will be found that a combination of two or more interacting residues have the biggest impact on activity. Hence, in some strategies, these residues are co-varied.
After the residues for variation have been identified, the method next generates a new variant library having the specified residue variation. See block 11. Various methodologies are available for this purpose. In one example, an in vitro or in vivo recombination-based diversity generation mechanism is performed to generate the new variant library. Such procedures may employ oligonucleotides containing sequences or subsequences for encoding the proteins of the parental variant library. Some of the oligonucleotides will be closely related, differing only in the choice of codons for alternate amino acids selected for variation at 09. The recombination-based diversity generation mechanism may be performed for one or multiple cycles. If multiple cycles are used, each involves a screening step to identify which variants have acceptable performance to be used in a next recombination cycle. This is a form of directed evolution.
In a different example, a "reference" protein sequence is chosen and the residues selected at 09 are "toggled" to identify individual members of the variant library. The new proteins so identified are synthesized by an appropriate technique to generate the new library. In one example, the reference sequence may be a top-performing member of the training set or a "best" sequence predicted by a PLS or PCR model.
In another approach, the sequence activity model is used as a "fitness function" in a genetic algorithm for exploring sequence space. After one or more rounds of the genetic algorithm (with each round using the fitness function to select one or more possible sequences for a genetic operation), a next generation library is identified for use as described in this flow chart. In a very real sense this strategy can be viewed as in silico directed evolution. In an ideal case, if one had in hand an accurate, precise global or local fitness function one could perform all the evolution in silico and synthesize a single best variant for use in the final commercial or research application. Though this is likely to be impossible to achieve in most cases such a view of the process lends clarity to the goals and approach of using machine learning techniques for directed evolution.
After the new library has been produced, it is screened for activity, as indicated in a block 13. Ideally, the new library will present one or more members with better activity than was observed in the previous library. However, even without such advantage, the new library can provide beneficial information. Its members may be employed for generating improved models that account for the effects of the variations selected in 09, and thereby more accurately predict activity across wider regions of sequence space. Further, the library may represent a passage in sequence space from a local maximum toward a global maximum (in activity).
Depending on the goal of process 01, it may be desirable to generate a series of new protein variant libraries, with each one providing new members of a training set. The updated training set is then used to generate an improved model. To this end, process 01 is shown with a decision operation 15, which determines whether yet another protein variant library should be produced. Various criteria can be used to make this decision. Examples include the number of protein variant libraries generated so far, the activity of top proteins from the current library, the magnitude of activity desired, and the level of improvement observed in recent new libraries.
Assuming that the process is to continue with a new library, the process returns to operation 05 where a new sequence-activity model is generated from sequence and activity data obtained for the current protein variant library. In other words, the sequence and activity data for the current protein variant library serves as part of the training set for the new model (or it may serve as the entire training set). Thereafter, operations 07, 09, 11, 13, and 15 are performed as described above, but with the new model.
At some point, in process 01, this cycle will end and no new library will be generated. At that point, the process may simply terminate or one or more sequences from one or more of the libraries may be selected for development and/or manufacture. See block 17.
A. Choosing Protein Variant Libraries
Protein variant libraries are groups of multiple proteins having one or more residues that vary from member to member in a library. They may be generated by methods of this invention. They may provide the data for training sets used to generate sequence-activity models in accordance with this invention. The number of proteins included in a protein variant library depends on the application and the cost.
In one example, the protein variant library is generated from one or more naturally occurring proteins, which may be protein members encoded by a single gene family. Other starting points for the library may be used. From these seed or starting proteins, the library may be generated by various techniques. In one case, the library is generated by DNA fragmentation-mediated recombination as described in Stemmer (1994) Proc. Natl. Acad. Sci. USA 10747-10751 and WO 95/22625, synthetic oligonucleotide-mediated recombination as described in Ness et al. (2002) Nature Biotechnology 20:1251-1255 and WO 00/42561) on nucleic acids encoding part or all of one or more parent proteins. A combination of these methods may be used as well (i.e., recombination of DNA fragments and synthetic oligonucleotides) as well as other recombination-based methods described in, for example, WO97/20078 and WO98/27230.
In another case, a single starting sequence is modified in various ways to generate the library. Preferably the library is generated by systematically varying the individual residues of the starting sequence. In one example, a design of experiment (DOE) methodology is employed to identify the systematically varied sequences. In another example, a "wet lab" procedure such as oligonucleotide-mediated recombination is used to introduce some level of systematic variation.
As used herein, the term "systematically varied sequences" refers to a set of sequences in which each residue is seen in multiple contexts. In principle, the level of systematic variation can be quantified by the degree to which the sequences are orthogonal from one another (maximally different compared to the mean). In practice, the process does not depend on having maximally orthogonal sequences, however, the quality of the model will be improved in direct relation to the orthogonality of the sequence space tested. In a simple example, a peptide sequence is systematically varied by identifying two residue positions, each of which can have one of two different amino acids. A maximally diverse library includes all four possible sequences. Such maximal systematic variation increases exponentially with the number of variable positions; e.g., by 2N, when there are 2 options at each of N residue positions. Those having ordinary skill in the art will readily recognize that maximal systematic variation, however, is not required by the invention methods. Systematic variation provides a mechanism for identifying a relatively small set of sequences for testing that provides a good sampling of sequence space.
Protein variants having systematically varied sequences can be obtained in a number of ways using techniques that are well known to those having ordinary skill in the art. As indicated, suitable methods include recombination-based methods that generate variants based on one or more "parental" polynucleotide sequences. Polynucleotide sequences can be recombined using a variety of techniques, including, for example, DNAse digestion of polynucleotides to be recombined followed by ligation and/or PCR reassembly of the nucleic acids. These methods include those described in, for example, Stemmer (1994) Proc. Natl. Acad. Sci. USA, 91:10747-10751, U.S. Pat. No. 5,605,793, "Methods for In Vitro Recombination," U.S. Pat. No. 5,811,238, "Methods for Generating Polynucleotides having Desired Characteristics by Iterative Selection and Recombination," U.S. Pat. No. 5,830,721, "DNA Mutagenesis by Random Fragmentation and Reassembly," U.S. Pat. No. 5,834,252, "End Complementary Polymerase Reaction," U.S. Pat. No. 5,837,458, "Methods and Compositions for Cellular and Metabolic Engineering," "WO98/42832, "Recombination of Polynucleotide Sequences Using Random or Defined Primers," WO 98/27230, "Methods and Compositions for Polypeptide Engineering," WO 99/29902, "Method for Creating Polynucleotide and Polypeptide Sequences," and the like.
Synthetic recombination methods are also particularly well suited for generating protein variant libraries with systematic variation. In synthetic recombination methods, a plurality of oligonucleotides are synthesized which collectively encode a plurality of the genes to be recombined. Typically the oligonucleotides collectively encode sequences derived from homologous parental genes. For example, homologous genes of interest are aligned using a sequence alignment program such as BLAST (Atschul, et al., J. Mol. Biol., 215:403-410 (1990). Nucleotides corresponding to amino acid variations between the homologues are noted. These variations are optionally further restricted to a subset of the total possible variations based on covariation analysis of the parental sequences, functional information for the parental sequences, selection of conservative or non-conservative changes between the parental sequences, or other like criteria. Variations are optionally further increased to encode additional amino acid diversity at positions identified by, for example, covariation analysis of the parental sequences, functional information for the parental sequences, selection of conservative or non-conservative changes between the parental sequences, or apparent tolerance of a position for variation. The result is a degenerate gene sequence encoding a consensus amino acid sequence derived from the parental gene sequences, with degenerate nucleotides at positions encoding amino acid variations. Oligonucleotides are designed which contain the nucleotides required to assemble the diversity present in the degenerate gene. Details regarding such approaches can be found in, for example, Ness et al. (2002), Nature Biotechnology 20:1251-1255, WO 00/42561, "Oligonucleotide Mediated Nucleic Acid Recombination," WO 00/42560, "Methods for Making Character Strings, Polynucleotides and Polypeptides having Desired Characteristics," WO 01/75767, "In Silico Cross-Over Site Selection," and WO 01/64864, "Single-Stranded Nucleic Acid Template-Mediated Recombination and Nucleic Acid Fragment Isolation." The identified polynucleotide variant sequences may be transcribed and translated, either in vitro or in vivo, to create a set or library of protein variant sequences.
The set of systematically varied sequences can also be designed a priori using design of experiment (DOE) methods to define the sequences in the data set. A description of DOE methods can be found in Diamond, W. J. (2001) Practical Experiment Designs: for Engineers and Scientists, John Wiley & Sons and in "Practical Experimental Design for engineers and scientists" by William J Drummond (1981) Van Nostrand Reinhold Co New York, "Statistics for experimenters" George E. P. Box, William G Hunter and J. Stuart Hunter (1978) John Wiley and Sons, New York, or, e.g., on the world wide web at itl.nist.gov/div898/handbook/. There are several computational packages available to perform the relevant mathematics, including Statistics Toolbox (MatLab), JMP, Statistica and Statease Design expert. The result is a systematically varied and orthogonal dispersed data set of sequences that is suitable for building the sequence activity model of the present invention. DOE-based data sets can be readily generated using either Plackett-Burman or Fractional Factorial designs. Id.
In engineering and chemical sciences, fractional factorial designs, for example, are used to define fewer experiments (than in full factorial designs) in which a factor is varied (toggled) between two or more levels. Optimization techniques are used to ensure that the experiments chosen are maximally informative in accounting for factor space variance. The same design approaches (e.g., fractional factorial, D-optimal design) can be applied in protein engineering to construct fewer sequences where a given number of positions are toggled between two or more residues. This set of sequences would be an optimal description of systematic variance present in the protein sequence space in question.
An example of the DOE approach applied to protein engineering includes the following operations: 1) Identify positions to toggle based on the principles described earlier (present in parental sequences, level of conservation, etc.) 2) Create a DOE experiment using one of the commonly available statistical packages by defining the number of factors (variable positions), the number of levels (choices at each position), and the number of experiments to run. The information content of the output matrix (typically consisting of 1s and 0s that represent residue choices at each position) depends directly on the number of experiments to run (the more the better). 3) Use the output matrix to construct a protein alignment that codes the 1s and 0s back to specific residue choices at each position. 4) Synthesize the genes encoding the proteins represented in the protein alignment. 5) Test the proteins encoded by the synthesized genes in relevant assay(s). 6) Build a model on the tested genes/proteins. 7) Follow the steps described before to identify positions of importance and to build a subsequent library with improved fitness.
For example purposes, consider a protein in which the functionally best amino acid residues at 20 positions are to be determined, e.g., where there are 2 possible amino acids available at each position. In this case, a resolution IV factorial design would be appropriate. A resolution IV design is defined as one which is capable of elucidating the effects of all single variables, with no two-factor effects overlapping them. The design would then specify a set of 40 specific amino acid sequences that would cover the total diversity of 220 (˜1 million) possible sequences. These sequences are then generated by a standard gene synthesis protocol and the function and fitness of these clones is determined.
An alternative to the above approaches is to employ all available sequences, e.g., the GenBank® database and other public sources, to provide the protein variant library. Although this entails massive computational power, current technologies make the approach feasible. Mapping all available sequences provides an indication of sequence space regions of interest.
B. Generating a Sequence Activity Model & Using That Model to Identify Residue Positions for Variation
As indicated above, a sequence-activity model used with the present invention relates protein sequence information to protein activity. The protein sequence information used by the model may take many forms. Frequently, it is a complete sequence of the amino acid residues in a protein; e.g., HGPVFSTGGA . . . . In some cases, however, it may be unnecessary to provide the complete amino acid sequence. For example, it may be sufficient to provide only those residues that are to be varied in a particular research effort. At later stages in research, for example, many residues may be fixed and only limited regions of sequence space remain to be explored. In such situations, it may be convenient to provide sequence activity models that require, as inputs, only the identification of those residues in the regions of the protein where the exploration continues. Still further, some models may not require exact identities of residues at the residue positions, but instead identify one or more physical or chemical properties that characterize the amino acid at a particular residue position. For example, the model may require specification of residue positions by bulk, hydrophobicity, acidity, etc. In some models, combinations of such properties are employed.
The form of the sequence-activity model can vary widely, so long as it provides a vehicle for correctly approximating the relative activity of proteins based on sequence information. Generally, it will treat activity as a dependent variable and sequence/residue values as independent variables. Examples of the mathematical/logical form of models include linear and non-linear mathematical expressions of various orders, neural networks, classification and regression trees/graphs, clustering approaches, recursive partitioning, support vector machines, and the like. In one preferred embodiment, the model form is a linear additive model in which the products of coefficients and residue values are summed. In another preferred embodiment, the model form is a non-linear product of various sequence/residue terms, including certain residue cross products (which represent interaction terms between residues).
Models are developed from a training set of activity versus sequence information to provide the mathematical/logical relationship between activity and sequence. This relationship is typically validated prior to use for predicting activity of new sequences or residue importance.
Various techniques for generating models are available. Frequently, such techniques are optimization or minimization techniques. Specific examples include partial least squares, various other regression techniques, as well as genetic programming optimization techniques, neural network techniques, recursive partitioning, support vector machine techniques, CART (classification and regression trees), and/or the like. Generally, the technique should produce a model that can distinguish residues that have a significant impact on activity from those that do not. Preferably, the model should also rank individual residues or residue positions based on their impact on activity.
In one important class of techniques, models are generated by a regression technique that identifies covariation of independent and dependent variables in a training set. Various regression techniques are known and widely used. Examples include multiple linear regression (MLR), principal component regression (PCR) and partial least squares regression (PLS).
MLR is the most basic of these techniques. It simply solves a set of coefficient equations for members of a training set. Each equation relates to the activity of a training set member (dependent variable) with the presence or absence of a particular residue at a particular position (independent variables). Depending upon the number of residue options in the training set, these expressions can be quite large.
Like MLR, PLS and PCR generate models from equations relating sequence activity to residue values. However, these techniques do so in a different manner. They first perform a coordinate transformation to reduce the number of independent variables. They then perform the regression on the transformed variables. In MLR, there are a potentially very large number of independent variables: two or more for each residue position that varies within the training set. Given that proteins and peptides of interest are often quite large and the training set may provide many different sequences, the number of independent variables can quickly become very large. By reducing the number of variables to focus on those that provide the most variation in the data set, PLS and PCR generally require fewer samples and simplify the problem of generating a model.
PCR is similar to PLS regression in that the actual regression is done on a relatively small number of latent variables obtained by coordinate transformation of the raw independent variables (residue values). The difference between PLS and PCR is that the latent variables in PCR are constructed by maximizing covariation between the independent variables (residue values). In PLS regression, the latent variables are constructed in such a way as to maximize the covariation between the independent variables and the dependent variables (activity values). Partial Least Squares regression is described in Hand, D. J., et al. (2001) Principles of Data Mining (Adaptive Computation and Machine Learning), Boston, Mass., MIT Press, and in in Geladi, et al. (1986) "Partial Least-Squares Regression: a Tutorial," Anal. Chim. Acta, 198:1-17. Both of these references are incorporated herein by reference for all purposes.
In PCR and PLS, the direct result of the regression is an expression for activity that is a function of the weighted latent variables. This expression can be transformed to an expression for activity as a function of the original independent variables by performing a coordinate transformation that converts the latent variables back to the original independent variables.
In essence, both PCR and PLS first reduce the dimensionality of the information contained in the training set and then perform a regression analysis on a transformed data set; which has been transformed to produce new independent variables, but preserves the original dependent variable values. The transformed versions of the data sets may result in only a relatively few expressions for performing the regression analysis. Compare this with a situation where no dimension reduction is performed. In that situation, each separate residue for which there can be a variation must be considered. This can be a very large set of coefficients; 2N coefficients, where N is the number of residue positions that may vary in the training set. In a typical principal component analysis, only 3, 4, 5, 6 principal components are employed.
The ability of machine learning techniques to fit the training data is often referred to as the "model fit" and in regression techniques such as MLR, PCR and PLS is typically measured by the sum squared difference between measured and predicted values. For a given training set, the optimal model fit will always be achieved using MLR, with PCR and PLS often having a worse model fit (higher sum squared error between measurements and predictions). However, the chief advantage of using latent variable regression techniques such as PCR and PLS lies in the predictive ability of such models. Obtaining a model fit with very small sum squared error in no way guarantees the model will be able to accurately predicted new samples not seen in the training set--in fact, it is often the opposite case, particularly when there are many variables and only a few observations (samples). Thus latent variable regression techniques (PCR, PLS) while often having worse model fits on the training data are usually more robust and are able to predict new samples outside the training set more accurately.
Another class of tools that can be used to generate models in accordance with this invention is the support vector machines. These mathematical tools take as inputs training sets of sequences that have been classified into two or more groups based on activity. Support vector machines operate by weighting different members of a training set differently depending upon how close they are to a hyperplane interface separating "active" and "inactive" members of the training set. This technique requires that the scientist first decide which training set members to place in the active group and which training set members to place in the inactive group. This may be accomplished by choosing an appropriate numerical value of activity to serve as the boundary between active and inactive members of the training set. From this classification, the support vector machine will generate a vector, W, that can provide coefficient values for individual ones of the independent variables defining the sequences of the active and inactive group members in the training set. These coefficients can be used to "rank" individual residues as described elsewhere herein. The technique attempts to identify a hyperplane that maximizes the distance between the closest training set members on opposite sides of that plane. In another variation, support vector regression modeling is carried out. In this case, the dependent variable is a vector of continuous activity values. The support vector regression model will generate a coefficient vector, W, which can be used to rank individual residues.
SVMs have been used to look at large data sets in many studies and have been quite popular in the DNA microarray field. Their potential strengths include the ability to finely discriminate (by weighting) which factors separate samples from each other. To the extent that an SVM can tease out precisely which residues contribute to function, it can be a particularly useful tool for ranking residues in accordance with this invention. SVMs are described in S. Gunn (1998) "Support Vector Machines for Classification and Regressions," Technical Report, Faculty of Engineering and Applied Science, Department of Electronics and Computer Science, University of Southampton, which is incorporated herein by reference for all purposes.
Another model generation technique of interest is genetic programming. This technique employs a Darwinian style evolution to discover the formulae and rules that characterize the data of a training set. It can be used in regression problems of the types described herein. The underlying effect can be linear or non-linear. Genetic programming is described in R. Goodacre et al. (2000) "Detection of the Dipicolinic Acid Biomarker in Bacillus Spores Using Curie-Point Pyrolysis Mass Spectrometry and Fourier Transform Infrared Spectroscopy," Anal. Chem., 72, 119-127, which is incorporated herein by reference for all purposes. Examples of software tools for performing genetic programming include the "GMAX" and the "GMAX-Bio" available from Aber Genomic Computing Ltd of Wales, UK.
i) Linear Model Examples
While the present invention is directed to non-linear models, these may be more easily understood in the context of linear models of sequence versus activity. Thus, the form and development of linear models will now be described. In general, a linear regression model of activity versus sequence has the following form:
y = c 0 + i = 1 N j = 1 M c ij x ij ( 1 ) ##EQU00001##
In this linear expression, y is predicted response, while cij and xij are the regression coefficient and bit value or dummy variable used to represent residue choice, respectively at position i in the sequence. There are N residue positions in the sequences of the protein variant library and each of these may be occupied by one or more residues. At any given position, there may be j=1 through M separate residue types. This model assumes a linear (additive) relationship between the residues at every position. An expanded version of equation 1 follows:
y=c0+c11x11+c12x12+ . . . c1Mx1M+c2lx21+c22x22+ . . . c2Mx2M+ . . . +cNMxNM
As indicated, data in the form of activity and sequence information is derived from the initial protein variant library and used to determine the regression coefficients of the model. The dummy variables are first identified from an alignment of the protein variant sequences. Amino acid residue positions are identified from among the protein variant sequences in which the amino acid residues in those positions differ between sequences. Amino acid residue information in some or all of these variable residue positions may be incorporated in the sequence activity model.
Table I contains sequence information in the form of variable residue positions and residue types for 10 illustrative variant proteins, along with activity values corresponding to each variant protein. Understand that these are representative members of a larger set that is required to generate enough equations to solve for all the coefficients. Thus, for example, for the illustrative protein variant sequences in Table I, positions 10, 166, 175, and 340 are variable residue positions and all other positions, i.e., those not indicated in the Table, contain residues that are identical between Variants 1-10.
TABLE-US-00001 TABLE I Illustrative Sequence and Activity Data Variable Positions: 10 166 175 340 y (activity) Variant 1 Ala Ser Gly Phe y1 Variant 2 Asp Phe Val Ala y2 Variant 3 Lys Leu Gly Ala y3 Variant 4 Asp Ile Val Phe y4 Variant 5 Ala Ile Val Ala y5 Variant 6 Asp Ser Gly Phe y6 Variant 7 Lys Phe Gly Phe y7 Variant 8 Ala Phe Val Ala y8 Variant 9 Lys Ser Gly Phe y9 Variant 10 Asp Leu Val Ala y10 and so on.
Thus, based on equation 1, a regression model can be derived from the systematically varied library in Table I, i.e.:
y=c0+c10Alax10Ala+c10Aspx10Asp+c10Lysx1- 0Lys+c166Serx166Ser+c166Phex166Phe+c166Leux1- 66Leu+c166Ilex166Ile+c175Glyx175Gly+c175Valx175Val+c340Phex340Phe+c340Alax340Ala (2)
The bit values (x dummy variables) can be represented as either 1 or 0 reflecting the presence or absence of the designated amino acid residue or alternatively, 1 or -1, or some other surrogate representation. For example, using the 1 or 0 designation, x10Ala would be "1" for Variant 1 and "0" for Variant 2. Using the 1 or -1 designation, x10Ala would be "1" for Variant 1 and "-1" for Variant 2. The regression coefficients can thus be derived from regression equations based on the sequence activity information for all variants in library. Examples of such equations for Variants 1-10 (using the 1 or 0 designation for x) follow:
y1=c0+c10Ala(1)+c10Asp(0)+c10Lys(0)+c166Ser(- 1)+c166Phe(0)+c166Leu(0)+c166Ile(0)+c175Gly(1)+c1- 75Val(0)+c340Phe(1)+c340Ala(0)
y2=c0+c10Ala(0)+c10Asp(1)+c10Lys(0)+c166Ser(- 0)+c166Phe(1)+c166Leu(0)+c166Ile(0)+c175Gly(0)+c1- 75Val(1)+c340Phe(0)+c340Ala(1)
y3=c0+c10Ala(0)+c10Asp(0)+c10Lys(1)+c166Ser(- 0)+c166Phe(0)+c166Leu(1)+c166Ile(0)+c175Gly(1)+c1- 75Val(0)+c340Phe(0)+c340Ala(1)
y4=c0+c10Ala(0)+c10Asp(1)+c10Lys(0)+c166Ser(- 0)+c166Phe(0)+c166Leu(0)+c166Ile(1)+c175Gly(0)+c1- 75Val(1)+c340Phe(1)+c340Ala(0)
y5=c0+c10Ala(1)+c10Asp(0)+c10Lys(0)+c166Ser(- 0)+c166Phe(0)+c166Leu(0)+c166Ile(1)+c175Gly(0)+c1- 75Val(1)+c340Phe(0)+c340Ala(1)
y6=c0+c10Ala(0)+c10Asp(1)+c10Lys(0)+c166Ser(- 1)+c166Phe(0)+c166Leu(0)+c166Ile(0)+c175Gly(1)+c1- 75Val(0)+c340Phe(1)+c340Ala(0)
y7=c0+c10Ala(0)+c10Asp(0)+c10Lys(1)+c166Ser(- 0)+c166Phe(1)+c166Leu(0)+c166Ile(0)+c175Gly(1)+c1- 75Val(0)+c340Phe(1)+c340Ala(0)
y8=c0+c10Ala(1)+c10Asp(0)+c10Lys(0)+c166Ser(- 0)+c166Phe(1)+c166Leu(0)+c166Ile(0)+c175Gly(0)+c1- 75Val(1)+c340Phe(0)+c340Ala(1)
y9=c0+c10Ala(0)+c10Asp(0)+c10Lys(1)+c166Ser(- 1)+c166Phe(0)+c166Leu(0)+c166Ile(0)+c175Gly(1)+c1- 75Val(0)+c340Phe(1)+c340Ala(0)
The complete set of equations can be readily solved using a regression technique (e.g., PCR, PLS, or MLR) to determine the value for regression coefficients corresponding to each residue and position of interest. In this example, the relative magnitude of the regression coefficient correlates to the relative magnitude of contribution of that particular residue at the particular position to activity. The regression coefficients may then be ranked or otherwise categorized to determine which residues are more likely to favorably contribute to the desired activity. Table II provides illustrative regression coefficient values corresponding to the systematically varied library exemplified in Table I:
TABLE-US-00002 TABLE II Illustrative Rank Ordering of Regression Coefficients REGRESSION COEFFICIENT VALUE c166Ile 62.15s c175Gly 61.89 c10Asp 60.23 c340 Ala 57.45 c10 Ala 50.12 c166 Phe 49.65 c166Leu 49.42 c340 Phe 47.16 c166Ser 45.34 c175 Val 43.65 c10 Lys 40.15
The rank ordered list of regression coefficients can be used to construct a new library of protein variants that is optimized with respect to a desired activity (i.e., improved fitness). This can be done in various ways. In one case, it is accomplished by retaining the amino acid residues having coefficients with the highest observed values. These are the residues indicated by the regression model to contribute the most to desired activity. If negative descriptors are employed to identify residues (e.g., 1 for leucine and -1 for glycine), it becomes necessary to rank residue positions based on absolute value of the coefficient. Note that in such situations, there is typically only a single coefficient for each residue. The absolute value of the coefficient magnitude gives the ranking of the corresponding residue position. Then, it becomes necessary to consider the signs of the individual residues to determine whether each of them is detrimental or beneficial in terms of the desired activity.
ii) Non-Linear Models
Non-linear modeling is employed to account for residue-residue interactions that contribute to activity in proteins. An N-K landscape describes this problem. The parameter N refers to the number of variable residues in a collection of related polypeptides sequences. The parameter K represents the interaction between individual residues within anyone of these polypeptides. Interaction is usually a result of close physical proximity between various residues whether in the primary, secondary, or tertiary structure of the polypeptide. The interaction may be due to direct interactions, indirect interactions, physicochemical interactions, interactions due to folding intermediates, translational effects, and the like.
The parameter K is defined such that for value K=1, each variable residue (e.g., there are 20 of them) interacts with exactly one other residue in its sequence. In the case where all residues are physically and chemically separate from the effects of all other residues, the value of K is zero. Obviously, depending upon the structure of the polypeptide, K can have a wide range of different values. With a rigorously solved structure of the polypeptide in question, a value for K may be estimated. Often, however, this is not the case.
A purely linear, additive model of polypeptide activity (as described above) can be improved by including one or more non-linear interaction terms representing specific interactions between 2 or more residues. In the context of the model form presented above, these terms are depicted as "cross-products" containing two or more dummy variables representing the two or more particular residues (each associated with a particular position in the sequence) that interact to have a significant positive or negative impact on activity. For example, a cross-product term may have the form cabxaxb, where xa is a dummy variable representing the presence of a particular residue at a particular position on the sequence and the variable xb represents the presence of a particular residue at a different position (that interacts with the first position) in the polypeptide sequence. A detailed example form of the model is shown below.
The presence of all residues represented in the cross-product term (each of two or more specific types of residue at specifically identified positions) impacts the overall activity of the polypeptide. The impact can be manifest in many different ways. For example, each of the individual interacting residues when present alone in a polypeptide may have a negative impact on activity, but when each of them is present together in the polypeptide, the overall effect is positive. The opposite may be true in other cases. In addition, there may be a synergistic effect produced in which each of the individual residues alone has a relatively limited impact on activity, but when all of them are present the effect on activity is greater than the cumulative effects of all the individual residues.
It is possible that a non-linear model could include a cross-product term for every possible combination of interacting variable residues in the sequence. However, this would not represent physical reality, as only a subset of the variable residues actually interact with one another. In addition, it would result in "overfitting" to produce a model that provides spurious results that are manifestations of the particular polypeptides used to create the model and do not represent real interactions within the polypeptide. The correct number of cross-product terms for a model that represents physical reality, and avoids overfitting, is dictated by the value of K. For example, if K=1, the number of cross-product interaction terms equals N.
Note that in general, it may be more preferable to have too few cross-product terms than too many. If the relatively few cross-product terms included in the non-linear model are actually those having the biggest influence on activity, than it is definitely preferable to have too few. As should be apparent, in constructing a non-linear model, it is important to identify those cross-product interaction terms representing true structural interactions that have a significant impact on activity. This can be accomplished in various ways. These include forward addition, where candidate cross-product terms (starting with the term with largest regression coefficient and progressing to terms with smaller regression coefficients) are added to the initial linear only model one at a time until the addition of terms is no longer statistically significant (as measured by an F-test or some other appropriate statistical test); reverse elimination, where all possible cross product terms are added at the beginning and removed one at a time (starting with the term with smallest regression coefficient and progressing to terms with larger regression coefficients) until removal the least important remaining term is statistically significant. One example presented below involves the use of a genetic algorithm to identify the useful non-linear terms.
Generally, the approach to generating a non-linear model containing such interaction terms is the same as the approach described above for generating a linear model. In other words, a training set is employed to "fit" the data to a model. However, one or more non-linear terms, preferably the cross-product terms discussed above, are added to the form of the model. Further, the resulting non-linear model, like the linear models described above, can be employed to rank the importance of various residues on the overall activity of a polypeptide. Various techniques can identify the best combination of variable residues as predicted by the non-linear equation. Unfortunately, unlike the linear case, it is often impossible to accomplish this by simple inspection of the additive model. Approaches to ranking the residues are described below.
Note that there are a very large number of possible cross-product terms for variable residues, even when limited to interactions caused by only two residues. As more interactions occur, the number of potential interactions to consider for a non-linear model grows in an exponential manner. If the model includes the possibility of interactions that include three or more residues, the number of potential terms grows even more rapidly.
In a simple case where there are 20 variable residues and K=1 (assumes that each variable residue interacts with one other variable residue), there should be 20 interaction terms (cross-products) in the model. If there are any fewer, the model will not fully describe the interactions (although some of the interactions may not have a significant impact on activity), and any more and the model may overfit the data set. There are N*(N-1)/2 or 190 possible pairs of interactions. Finding the combination of 20 unique pairs that describe the 20 interactions in the sequence is a significant computational problem. There are approximately 5.48×1026 possible combinations.
Numerous techniques can be employed to identify the relevant cross-product terms. Depending upon the size of the problem and the computational power available, one might be able to explore all possible combinations and thereby identify the one model that best fits the data (the numbers of the training set). However, often the problem will be too large for the available computational resources and so one must resort to an efficient search algorithm or an approximation. As mentioned, one suitable search technique is a genetic algorithm.
In a genetic algorithm, an appropriate fitness function and an appropriate mating procedure are defined. The fitness function provides a criterion for determining which models (combinations of cross-product terms) are "most fit" (i.e., likely to provide the best results). The mating procedure provides a mechanism for introducing new combinations of cross-product terms from successful "parental" models in a previous generation. One example of a genetic algorithm for identifying a combination of cross-product terms will now be described with reference to FIG. 2. This algorithm begins with a first generation comprising multiple possible models, some of which do a better job of representing physical reality than others. See block 201. The first and each successive generation is represented as a number of models in a "population". Each "model" is a combination of linear terms (fixed across all models) and non-linear cross-product terms. The "model" in this genetic algorithm does not intrinsically include coefficients for the individual linear and non-linear terms, only an identification of a combination of non-linear terms (e.g., the cross-product terms). The genetic algorithm proceeds towards convergence by marching through successive generations of models, each characterized by a different combination of non-linear interaction terms.
Each model in a generation is used to fit a training set of polypeptides (having known sequences and associated activities). The training set is used to fit the individual models of the current generation. See block 203, 205, 207, and 209 of FIG. 2. In one example, a partial least squares technique or similar regression technique is used to perform the fit.
The predictive power of the resulting model (including coefficients obtained by the regression on the training set data) is used as a fitting function. To provide a detailed assessment of the predictive power, many different fits of a model may be provided for a given training set. See blocks 205, 207, and 209. Each fit provides its own unique set of coefficient values for the linear and non-linear terms of the model under consideration. In one approach, a "leave one out" approach is employed in which all but one member of the training set is used to fit the model. This left out member is then used to test the predictive power of the resulting instance of the model. The model instance (model terms together coefficient values identified by fitting) is expected to do a good job of predicting the activities of the training set members employed to produce it. However, it may not do so well at predicting the activity of a polypeptide from outside the utilized members of the training set. In a specific embodiment, multiple "leave one out" model instances are generated and each is assessed for its ability to predict the activity of the left out member. The resulting set of predictions is combined to get an aggregate measure of the predictive capability (see block 211). In one example, the aggregate measure is a predicted residual sum of squares (PRESS) for the various leave one out model instances of the current model. The PRESS is, in effect, the fitting function of the genetic algorithm.
After each combination of non-linear cross-product terms (model) in a particular generation is evaluated for its predictive power (i.e., decision 213 is answered in the negative), the genetic algorithm is checked for convergence. See block 215. Assuming that the genetic algorithm has not yet converged, the models of the current generation are ranked. Those that do the best job of predicting activity may be preserved and used in the next generation. See block 217. For example, an elitism rate of 10% may be employed. In other words, the top 10% of models (as determined using the fitting function and measured by, e.g., PRESS scores) are set aside to become members of the next generation. The remaining 90% of the members in the next generation are obtained by mating "parents" from the previous generation. See blocks 219, 221, and 223.
The "parents" are models selected randomly from the previous generation. See block 219. However, the random selection is typically weighted toward more fit numbers of the previous generation. For example, the parent models may be selected using a linear weighting (e.g., a model that performs 1.2 times better than another model is 20% more likely to be selected) or a geometric weighting (i.e., the predictive differences in models are raised to a power in order to obtain a probability of selection).
After a set of parent models has been selected, pairs of such models are mated (block 221) to produce children models by providing some non-linear terms from one parent and other non-linear terms form the other parent. In one approach, the non-linear terms (cross-products) of the two parents are aligned and each term is considered in succession to determine whether the child should take the term from parent A or from parent B. In one implementation, the mating process begins with parent A and randomly determines whether a "cross over" event should occur at the first non-linear term encountered. If so, the term is taken from parent B. If not, the term is taken from parent A. The next term in succession is considered for cross over, etc. The terms continue to come from the parent donating the previous term under consideration until a cross over event occurs. At that point, the next term is donated from the other parent and all successive terms are donated from that parent until another cross over event occurs. To ensure that the same non-linear cross-product terms is not selected at two different locations in the child model, various techniques may be employed, e.g., a partially matched cross over technique.
After each non-linear term has been considered, a child "model" is defined for the next generation. Then another two parents are chosen to produce another child model, and so on. Eventually, after a complete generation has been selected in this manner (block 223), the next generation is ready for evaluation, and process control then returns to block 203, where the members of the next generation are evaluated as described above.
The process continues generation-by-generation until convergence, (i.e., decision block 215 is answered in the negative. At that point, the top ranked model is selected from the current generation as the overall best model. See block 225. Convergence can be tested by many conventional techniques. Generally, it involves determining that the performance of the best model from a number of successive generations does not change appreciably.
At this point, an example will be presented to show the value of incorporating non-linear cross-product terms in a model predicting activity from sequence. Consider the following non-linear model in which it is assumed there are only two residue options at each variable position in the sequence. In this example, the protein sequence is cast into a coded sequence by using dummy variables that correspond to choice A or choice B, using +1 and -1 respectively. The model is immune to the arbitrary choice of which numerical value used to assign each residue choice.
TABLE-US-00003 TABLE III Variable Position 1 2 3 4 5 6 7 8 9 10 Choice A I L L M G W K C S F Choice B V A I P H N R T A Y Protein V A L P G W K T S F Sequence Model -1 -1 +1 -1 +1 +1 +1 -1 +1 +1 Sequence
With this coding scheme, the linear model used to associate protein sequences with activity can be written as follows:
y=c1x1+c2x2+c3x3+ . . . +cnxn+cNxN+c0 (3)
where y is the response (activity), cn the regression coefficient for the residue choice at position n, x the dummy variable coding for the residue choice (+1/-1) at position n, and c0 the mean value of the response. This form of the model assumes there are no interactions between the variable residues--each residue choice contributes independently to the overall fitness of the protein.
The non-linear model includes a certain number of (as yet undetermined) cross-product terms to account for interactions between residues:
y=1x1+c2x2+c3x3+ . . . +cnxn+c1,2x1x3+c1,3x1x3+c2,3- x2x3+c0 (4)
where the variables are the same as those in Eq. (3) but now there are non-linear terms, e.g., c1,2 is the regression coefficient for the interaction between variable positions 1 and 2.
In order to assess the performance of the linear and non-linear models a synthetic data source known as the NK landscape (Kauffman, 1993) was used. As mentioned, N is the number of variable positions in a simulated protein and K is the epistatic coupling between residues. The synthetic data set was generated only in silico.
This data set was used to generate an initial training set with S=40 synthetic samples, N=20 variable positions and K=1 (to reiterate, for K=1 each variable position is functionally coupled to one other variable position). In generating the randomized proteins, each variable position had an equal probability of containing the dummy variable +1 or -1. The residue-residue interactions (represented by cross-products) and actual activities are known for each member of the synthetic training set. Another V=100 samples were generated for use in validation. Again, the residue-residue interactions and activities are known for each member of the validation set.
The training sets were used to construct both linear and non-linear models using the methods described above. Some non-linear models were generated with selection of the cross-product terms (using a genetic algorithm as described above) and other non-linear models were generated without selection of such terms. For the training set size of S=40, the linear model was capable of correlating the measured and predicted values reasonably well, but demonstrated weaker correlation when validated against data not seen in the training set (see FIG. 3A). As shown, the dark data points represent the cross-validated predictions made by a linear model based on the other 39 data points in the training set and predicting the single held out data point. Thus there are actually 40 slightly different models represented by the dark data points. The light data points represent the predictions made by a single model constructed from the 40 training samples and used to predict the validation samples V, none of which were seen in the original training set. The use of the validation set then gives a good measure of the true predictive capacity of the model, as opposed to the cross validated training set, which can suffer from the model overfit problem especially for the non-linear cases described below.
This result for S=40 is interesting considering the linear model was used to model a non-linear fitness landscape. In this case, the linear model could, at best, capture the average contribution to fitness for the choice of a given residue. With enough of these average contributions taken together, the linear model could roughly predict the measured response. The validation results for the linear model were marginally better when the training size was increased to S=100 (see FIG. 3B). Then tendency of relatively simple models to underfit data is known as bias.
When the non-linear model was trained using only S=40 samples (and 20 non-linear cross-product terms were selected using a genetic algorithm as described above), the correlation with training set members was excellent (see FIG. 3C). Unfortunately, the model contained limited predictive power outside the training set, as evidenced by its limited correlation with measured values in the validation set. This non-linear model, with many potential variables (210 possible), and limited training data to facilitate identification of the proper cross-product terms, was able to essentially just memorize the data set it was trained on. This tendency of high complexity models to overfit the data is known as the variance. The bias-variance tradeoff represents a fundamental problem in machine learning and some form of validation is almost always required to address it when dealing with new or uncharacterized machine learning problems. Satisfyingly, for the larger training set (S=100) the non-linear model performed exceedingly well for both the training prediction and, more importantly, the validation prediction (see FIG. 3D). The validation predictions were so good that most of the data points are obscured by the dark circles use to plot the training set.
For comparison, FIGS. 3E and 3F show the performance of non-linear models prepared without careful selection of the cross-product terms. Unlike the models in FIGS. 3C and 3D, every possible cross-product term was chosen (i.e., 190 cross-product terms for N=20). As can be seen, the ability to predict validation set activity is relatively poor compared that of the non-linear models generated with selection of cross-product terms. This is a manifestation of overfitting.
iii) Generating an Optimized Protein Variant Library by Modifying Model-Predicted Sequences
Rather than simply synthesizing the single best-predicted protein one may generate a combinatorial library of proteins based on a sensitivity analysis of the best protein to changes in the residue choices at each location. The more sensitive a given residue choice is for the predicted protein, the greater the predicted fitness change will be. One can rank these sensitivities from highest to lowest and use the sensitivity scores to create combinatorial protein libraries in subsequent rounds by incorporating those residues based on sensitivity. For a linear model the sensitivity may be identified by simply considering the size of the coefficient associated with a given residue term in the model. For the non-linear model, this will not be possible. Instead the residue sensitivity may be determined by using the model to calculate changes in activity when a single residue is varied in the "best" predicted sequence.
Residues may be considered in the order in which they are ranked. For each residue under consideration, the process determines whether to "toggle" that residue. The term "toggling" refers to the introduction of multiple amino acid residue types into a specific position in the sequences of protein variants in the optimized library. For example, serine may appear in position 166 in one protein variant, whereas phenylalanine may appear in position 166 in another protein variant in the same library. Amino acid residues that did not vary between protein variant sequences in the training set typically remain fixed in the optimized library.
An optimized protein variant library can be designed such that all of the identified "high" ranking regression coefficient residues are fixed, and the remaining lower ranking regression coefficient residues are toggled. The rationale for this being that one should search the local space surrounding the `best` predicted protein. Note that the starting point "backbone" in which the toggles are introduced may be the best protein predicted by a model or an already validated `best` protein from a screened library.
In an alternative approach, at least one or more, but not all of the high-ranking regression coefficient residues identified may be fixed in the optimized library, and the others toggled. This approach is recommended if it is desired not to drastically change the context of the other amino acid residues by incorporating too many changes at one time. Again, the starting point for toggling may be the best set of residues as predicted by the model or a best validated protein from an existing library. Or the starting point may be an "average" clone that models well. In this case, it may be desirable to toggle the residues predicted to be of higher importance. The rationale for this being that one should explore a larger space in search for activity hills previously omitted from the sampling. This type of library is typically more relevant in early rounds as it generates a more refined picture for subsequent rounds.
Alternatives to the above methodology involve different procedures for using residue importance (rankings) in determining which residues to toggle. In one such alternative, higher ranked residue positions are more aggressively favored for toggling. The information needed in this approach includes the sequence of a best protein from the training set, a PLS or PCR predicted best sequence, and a ranking of residues from the PLS or PCR model. The "best" protein is a wet-lab validated "best" clone in the dataset (clone with the highest measured function that still models well, i.e., falls relatively close to the predicted value in cross validation). The method compares each residue from this protein with the corresponding residue from a "best predicted" sequence having the highest value of the desired activity. If the residue with the highest load or regression coefficient is not present in the `best` clone, the method introduces that position as a toggle position for the subsequent library. If the residue is present in the best clone, the method will not treat the position as a toggle position, and it will move the next position in succession. The process is repeated for various residues, moving through successively lower load values, until the library is of sufficient size is generated.
The number of regression coefficient residues to retain, and number of regression coefficient residues to toggle, can be varied. Factors to consider include the desired library size, the magnitude of difference between regression coefficients, and the degree to which nonlinearity is thought to exist--retaining residues with small (neutral) coefficients may uncover important nonlinearities in subsequent rounds of evolution. Typical optimized protein variant libraries of the present invention contain about 2N protein variants, where N represents the number of positions that are toggled between two residues. Stated another way, the diversity added by each additional toggle doubles the size of the library such that 10 toggle positions produces ˜1,000 clones (1,024), 13 positions ˜10,000 clones (8,192) and 20 positions ˜1,000,000 clones (1,048,576). The appropriate size of library depends on factors such as cost of screen, ruggedness of landscape, preferred percentage sampling of space etc. In some cases, it has been found that a relatively large number of changed residues produces a library in which an inordinately large percentage of the clones are non-functional. Therefore for some applications, it may be recommended that the number of residues for toggling ranges from about 2 to about 30; i.e., the library size ranges from between about 4 and 230˜109 clones.
In practice, one can pursue various subsequent round library strategies at the same time, with some strategies being more aggressive (fixing more "beneficial" residues) and other strategies being more conservative (fixing fewer "beneficial" residues in the hopes of exploring the space more thoroughly).
It may be desirable to identify and preserve groups or residues or "motifs" that occur in most naturally occurring or otherwise successful peptides. For example, it may be found that Ile at variable position 3 is always coupled with Val at variable position 11 in naturally occurring peptides. It has been found that such residue groups can be important to preserving activity in the peptide. Hence, in one embodiment, preservation of such groups is required in any toggling strategy. In other words, the only accepted toggles are those that preserve a particular grouping in the base protein or those that generate a different grouping that is also found in active proteins. In the latter case it will be necessary to toggle two or more residues.
In varied approaches, a wet-lab validated `best` (or one of the few best) protein in the current optimized library (i.e., a protein with the highest, or one of the few highest, measured function that still models well, i.e., falls relatively close to the predicted value in cross validation) may serve as a backbone where various schemes of changes are incorporated. In another approach, a wet-lab validated `best` (or one of the few best) protein in the current library that may not model well may serve as a backbone where various schemes of changes are incorporated. In other approaches, a sequence predicted by the sequence activity model to have the highest value (or one of the highest values) of the desired activity may serve as the backbone. In these approaches, the dataset for the "next generation" library (and possibly a corresponding model) is obtained by changing residues in one or a few of the best proteins. In one embodiment, these changes comprise a systematic variation of the residues in the backbone. In some cases, the changes comprise various mutagenesis, recombination and/or subsequence selection techniques. Each of these may be performed in vitro, in vivo, or in silico.
While the optimal sequence is predicted by a linear model can be identified by inspection as described above, the same is not true of a non-linear model. Certain residues appear in both linear and cross product terms and their overall affect on activity in the context of many possible combinations of other residues can present a daunting problem.
As with selection of cross product terms for a non-linear model, the optimal sequence predicted by a non-linear model can be identified by testing all possible sequences with the model (assuming sufficient computational resources) or a searching algorithm such as a genetic algorithm. One exemplary genetic algorithm will now be described.
In this algorithm, the fitness function is simply the non-linear model's prediction of activity. In a specific example, an elitism rate of about 5-10% is employed. Selection of parents for mating involves a linear weighted fitness operation. The selected parents provide ordered sets of residues and a uniform cross over operation is employed. The best computer generated protein is picked after no improvement is seen for at least 15 generations.
The information contained in the computer-evolved proteins identified as described above can be used to synthesize novel proteins in the lab and test them on physical assays. An accurate in silico representation of the real wet lab fitness function, allows researchers to reduce the number of cycles of evolution or the number variants needed to screen in the lab. Optimized protein variant libraries can be generated using the recombination methods described herein, or alternatively, by gene synthesis methods, followed by in vivo or in vitro expression. After, the optimized protein variant libraries are screened for desired activity, they may be sequenced. As indicated above in the discussion of FIG. 1, the activity and sequence information from the optimized protein variant library can be employed to generate another sequence activity model from which a further optimized library can be designed, using the methods described herein. In one approach, all proteins from this new library are used as part of the dataset.
iv) Alternative Modelling Options
Multiple other variations on the above approach are within the scope of this invention. As one example, the xij variables are representations of the physical or chemical properties of amino acids--rather than the exact identities of the amino acids themselves (leucine versus valine versus proline, . . . ). Examples of such properties include lipophilicity, bulk, and electronic properties (e.g., formal charge, van der Waals surface area associated a partial charge, etc.). To implement this approach, the xij values representing amino acid residues can be presented in terms of their properties or principal components constructed from the properties.
In another variation, the xij variables represent nucleotides, rather than amino acid residues. The goal is to identify nucleic acid sequences that encode proteins for a protein variant library. By using nucleotides rather than amino acids, one can optimize on parameters other than merely specific activity. For example, protein expression in a particular host or vector may be a function of nucleotide sequence. Two different nucleotide sequences may encode a protein having one amino acid sequence, but one of the nucleotide sequences expresses greater quantities of protein and/or expresses the protein in a more active state. By using nucleotide sequences rather than amino acid sequences, the methods of this invention can optimize for expression properties, for example, as well as specific activity.
In a specific embodiment, the nucleotide sequence is represented as codons. Models may employ codons as the atomic unit of a nucleotide sequence such that the predicted activities are a function of various codons in the nucleotide sequence. Each codon together with its position in the overall nucleotide sequence serves as an independent variable for generating sequence activity models. Note that different codons for given amino acid express differently in a given organism. More specifically, each organism has a preferred codon, or distribution of codon frequencies, for a given amino acid. By using codons as the independent variables, the invention accounts for these preferences. Thus, the invention can be used to generate a library of expression variants (e.g., where "activity" includes expression level from a particular host organism).
An outline of a particular method includes the following operations: (a) receiving data characterizing a training set of a protein variant library; (b) from the data, developing a non-linear sequence activity model that predicts activity as a function of nucleotide types and corresponding position in the nucleotide sequence; (c) using the sequence activity model to rank positions in a nucleotide sequence and/or nucleotide types at specific positions in the nucleotide sequence in order of impact on the desired activity; and (d) using the ranking to identify one or more nucleotides, in the nucleotide sequence, that are to be varied or fixed in order to impact the desired activity. As indicated, the nucleotides to be varied are preferably codons encoding particular amino acids.
Other variations of the above approach involve use of different techniques for ranking residues or otherwise characterizing them in terms of importance. With linear models, the magnitudes of regression coefficients were used to rank residues. Residues having coefficients with large magnitudes (e.g., 166 Ile) were viewed as high-ranking residues. This characterization was used to decide whether or not to vary a particular residue in the generation of a new, optimized library of protein variants. For non-linear models, the sensitivity analysis was more complex.
PLS and other techniques provide other information, beyond regression coefficient magnitude, that can be used to rank specific residues or residue positions. Techniques such as PLS and Principle Component Analysis (PCA) or PCR provide information in the form of principle components or latent vectors. These represent directions or vectors of maximum variation through multi-dimensional data sets such as the protein sequence-activity space employed in this invention. These latent vectors are functions of the various sequence dimensions; i.e., the individual residues or residue positions that comprise the protein sequences of the variant library used to construct the training set. A latent vector will therefore comprise a sum of contributions from each of the residue positions in the training set. Some positions will contribute more strongly to the direction of the vector. These will be manifest by relatively large "loads," i.e., the coefficients used to describe the vector. As a simple example, a training set may be comprised of tripeptides. The first latent vector will typically have contributions from all three residues.
Vector 1=a1(residue position 1)+a2(residue position 2)+a3(residue position 3)
The coefficients, a1, a2, and a3, are the loads. Because these reflect the importance of the corresponding residue positions to variation in the dataset, they can be used to rank the importance of individual residue positions for purposes of "toggling" decisions, as described above. Loads, like regression coefficients, may be used to rank residues at each toggled position. Various parameters describe the importance of these loads. Some such Variable Importance in Projection (VIP) make use of a load matrix, which is comprised of the loads for multiple latent vectors taken from a training set. In Variable Importance for PLS Projection, the importance of the ith variable (e.g., residue position) is computed by calculating VIP (variable importance in projection). For a given PLS dimension, a, (VIN)ak2 is equal to the squared PLS weight (wak)2 of a variable multiplied by the percent explained variability in y (dependent variable, e.g., certain function) by that PLS dimension. (VIN)ak2 is summed over all PLS dimensions (components). VIP is then calculated by dividing the sum by the total percent variability in y explained by the PLS model and multiplying by the number of variables in the model. Variables with large VIP, larger than 1, are the most relevant for correlating with a certain function (y) and hence highest ranked for purposes of making toggling decisions.
Another embodiment of the invention employs techniques that rank residues not simply by the magnitudes of their predicted contributions to activity, but by the confidence in those predicted contributions as well. In some cases the researcher will be concerned with spurious values of the coefficients or principal components.
In a more statistically rigorous approach, the ranking is based on a combination of magnitude and distribution. Coefficients with both high magnitudes and tight distributions give the highest ranking. In some cases, one coefficient with a lower magnitude than another may be given a higher ranking by virtue of having less variation. Thus, some embodiments of the invention rank residues or nucleotides based on both magnitude and standard deviation or variance. Various techniques can be used to accomplish this. One of these, a bootstrap p-value approach, will now be described.
An example of a method that employs a bootstrap method is depicted in FIG. 4. As shown there, a method 125 begins at a block 127 where an original data set S is provided. This may be a training set as described above. For example, it may be generated by systematically varying the individual residues of a starting sequence in any one of the manners described above. In the example of method 125, the data set S has M different data points (activity and sequence information collected from amino acid or nucleotide sequences) for use in the analysis.
From data set S, various bootstrap sets B are created. Each of these is obtained by sampling, with replacement, from set S to create a new set of M members--all taken from original set S. See block 129. The "with replacement" condition produces variations on the original set S. The new bootstrap set, B, will sometimes contain replicate samples from S. And, it may also lack certain samples originally contained in S.
As an example, consider a set S of 100 sequences. Each bootstrap set B used in the method contains itself 100 sequences. A bootstrap set B is created by randomly selecting each of the 100 member sequences from the 100 sequences in the original set S. Thus, it is possible that some sequences will be selected more than once and others will not be selected at all.
Using the bootstrap set B currently under consideration, the method next builds a model. See block 131. The model may be built as described above, using PLS, PCR, a SVM, genetic programming, etc. This model will provide coefficients or other indicia of ranking for the residues or nucleotides found in the various samples from set B. As shown at a block 133, these coefficients or other indicia are recorded for subsequent use.
Next, at a decision block 135, the method determines whether another bootstrap set should be created. If yes, the method returns to block 129 where a new bootstrap set B is created as described above. If no, the method proceeds to a block 137 discussed below. The decision at block 135 turns on how many different sets of coefficient values are to be used in assessing the distributions of those values. The number of sets B should be sufficient to generate accurate statistics. As an example, 100 to 1000 bootstrap sets are prepared and analyzed. This is represented as about 100 to 1000 passes through blocks 129, 131, and 133 of method 125.
After a sufficient number bootstrap sets B have been prepared and analyzed as described, decision 135 is answered in the negative. As indicated, the method then proceeds to block 137. There a mean and standard deviation of a coefficient (or other indicator generated by the model) is calculated for each residue or nucleotide (including codons) using the coefficient values (e.g., 100 to 1000 of them, one from each bootstrap set). From this information, the method can calculate the t-statistic and determine the confidence interval that the measured value is different from zero. From the t-statistic it calculates the p-value for the confidence interval. In this case, the smaller p-value the more confidence that the measured regression coefficient is different from zero.
Note that the p-value is but one of many different types of characterization that can account for the statistical variation in a coefficient or other indicator of residue importance. Examples include calculating 95 percent confidence intervals for regression coefficients and excluding any regression coefficient for consideration for which 95 percent confidence interval crosses zero line. Basically, any characterization that accounts for standard deviation, variance, or other statistically relevant measure of data distribution can be used. Such characterization preferably also accounts for the magnitude of the coefficients.
A large standard deviation can result from various sources. One source is poor measurements in the data set. Another is a limited representation of a particular residue or nucleotide in the original data set. In this latter case, some bootstrap sets will contain no occurrences of a particular residue or nucleotide. In such cases, the value of the coefficient for that residue will be zero. Other bootstrap sets will contain at least some occurrences of the residue or nucleotide and give a non-zero value of the corresponding coefficient. But the sets giving a zero value will cause the standard deviation of the coefficient to become relatively large. This reduces the confidence in the coefficient value and results in a lower rank. But this is to be expected, given that there is relatively little data on the residue or nucleotide in question.
Next, at a block 139, the method ranks the regression coefficients (or other indicators) from lower (best) p-value to highest (worst) p-value. This ranking correlates highly with the absolute value of the regression coefficients themselves, owing to the fact that the larger the absolute value, the more standard deviations removed from zero. Thus, for a given standard deviation, the p-value becomes smaller as the regression coefficient becomes larger. However, the absolute ranking will not always be the same with both p-value and pure magnitude methods, especially when relatively few data points are available to begin with in set S.
Finally, as shown at a block 141, the method fixes and toggles certain residues based on the rankings observed in the operation of block 139. This is essentially the same use of rankings described above for other embodiments. In one approach, the method fixes the best residues (now those with the lowest p-values) and toggles the others (those with highest p-values).
This method 125 has been shown in silico to perform well. Moreover, the p-value ranking approach naturally deals with single or few instance residues: the p-values will generally be higher (worse) because in the bootstrap process, those residues that did not appear often in the original data set will be less likely to get picked up at random. Even if their coefficients are large, their variability (measured in standard deviations) will be quite high as well. Intuitively, this is the desired result, since those residues that are not well represented (either have not seen with sufficient frequency or have lower regression coefficients) may be good candidates for toggling in the next round of library design.
III. Digital Apparatus and Systems
As should be apparent, embodiments of the present invention employ processes acting under control of instructions and/or data stored in or transferred through one or more computer systems. Embodiments of the present invention also relate to apparatus for performing these operations. Such apparatus may be specially designed and/or constructed for the required purposes, or it may be a general-purpose computer selectively activated or reconfigured by a computer program and/or data structure stored in the computer. The processes presented herein are not inherently related to any particular computer or other apparatus. In particular, various general-purpose machines may be used with programs written in accordance with the teachings herein. In some cases, however, it may be more convenient to construct a specialized apparatus to perform the required method operations. A particular structure for a variety of these machines will appear from the description given below.
In addition, embodiments of the present invention relate to computer readable media or computer program products that include program instructions and/or data (including data structures) for performing various computer-implemented operations. Examples of computer-readable media include, but are not limited to, magnetic media such as hard disks, floppy disks, magnetic tape; optical media such as CD-ROM devices and holographic devices; magneto-optical media; semiconductor memory devices, and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM) and random access memory (RAM), and sometimes application-specific integrated circuits (ASICs), programmable logic devices (PLDs) and signal transmission media for delivering computer-readable instructions, such as local area networks, wide area networks, and the Internet. The data and program instructions of this invention may also be embodied on a carrier wave or other transport medium (e.g., optical lines, electrical lines, and/or airwaves).
Examples of program instructions include both low-level code such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter. Further, the program instructions include machine code, source code and any other code that directly or indirectly controls operation of a computing machine in accordance with this invention. The code may specify input, output, calculations, conditionals, branches, iterative loops, etc.
In one example, code embodying methods of the invention are embodied in a fixed media or transmissible program component containing logic instructions and/or data that when loaded into an appropriately configured computing device causes the device to perform a genetic operator on one or more character string. FIG. 5 shows an example digital device 500 that should be understood to be a logical apparatus that can read instructions from media 517, network port 519, user input keyboard 509, user input 511 or other inputting means. Apparatus 500 can thereafter use those instructions to direct statistical operations in data space, e.g., to construct one or more data set (e.g., to determine a plurality of representative members of the data space). One type of logical apparatus that can embody the invention is a computer system as in computer system 500 comprising CPU 507, optional user input devices keyboard 509, and GUI pointing device 511, as well as peripheral components such as disk drives 515 and monitor 505 (which displays GO modified character strings and provides for simplified selection of subsets of such character strings by a user. Fixed media 517 is optionally used to program the overall system and can include, e.g., a disk-type optical or magnetic media or other electronic memory storage element. Communication port 519 can be used to program the system and can represent any type of communication connection.
The invention can also be embodied within the circuitry of an application specific integrated circuit (ASIC) or programmable logic device (PLD). In such a case, the invention is embodied in a computer readable descriptor language that can be used to create an ASIC or PLD. The invention can also be embodied within the circuitry or logic processors of a variety of other digital apparatus, such as PDAs, laptop computer systems, displays, image editing equipment, etc.
IV. Other Embodiments
While the foregoing invention has been described in some detail for purposes of clarity and understanding, it will be clear to one skilled in the art from a reading of this disclosure that various changes in form and detail can be made without departing from the true scope of the invention. For example, all the techniques and apparatus described above may be used in various combinations. All publications, patents, patent applications, or other documents cited in this application are incorporated by reference in their entirety for all purposes to the same extent as if each individual publication, patent, patent application, or other document were individually indicated to be incorporated by reference for all purposes.
Patent applications by Richard John Fox, Redwood City, CA US
Patent applications by Maxygen, Inc.
Patent applications in class In silico screening
Patent applications in all subclasses In silico screening