 Methodology article
 Open Access
 Published:
Bayesian model accounting for withinclass biological variability in Serial Analysis of Gene Expression (SAGE)
BMC Bioinformatics volume 5, Article number: 119 (2004)
Abstract
Background
An important challenge for transcript counting methods such as Serial Analysis of Gene Expression (SAGE), "Digital Northern" or Massively Parallel Signature Sequencing (MPSS), is to carry out statistical analyses that account for the withinclass variability, i.e., variability due to the intrinsic biological differences among sampled individuals of the same class, and not only variability due to technical sampling error.
Results
We introduce a Bayesian model that accounts for the withinclass variability by means of mixture distribution. We show that the previously available approaches of aggregation in pools ("pseudolibraries") and the BetaBinomial model, are particular cases of the mixture model. We illustrate our method with a brain tumor vs. normal comparison using SAGE data from public databases. We show examples of tags regarded as differentially expressed with high significance if the withinclass variability is ignored, but clearly not so significant if one accounts for it.
Conclusion
Using available information about biological replicates, one can transform a list of candidate transcripts showing differential expression to a more reliable one. Our method is freely available, under GPL/GNU copyleft, through a user friendly webbased online tool or as R language scripts at supplemental website.
Background
An important challenge in Serial Analysis of Gene Expression (SAGE) [1] analysis is the decision whether a gene is differentially expressed between two classes, for example tumoral vs. normal classes. In statistical terms, this essential step is to test the null hypothesis H_{0}: "gene has no differential expression between the two probed classes". A much more usual approach is to assign an index (Pvalue or Bayes factor, for example) that measures the confidence/significance of the hypothesis and let the biologists themselves to establish a cutoff of what they call significant.
This necessity arises because counting sequenced SAGE tags is a process prone to random and systematic errors that affect gene expression abundance estimates. Systematic errors may come from various sources such as GC content bias [2], sequencing errors [3, 4] as well as the possibility of non unique tags. This kind of error can be detected/corrected using some bioinformatics procedures such as quality control of automatic sequencing pipeline [5], or statistical estimation procedures such as "denoising" [6, 7]. Random errors are due to the inherent stochastic characteristic of SAGE data acquisition: sampling from automatic sequencing. Like colored balls in an urn, sampling and counting SAGE tags from a library is commonly modeled by a Bernoulli Process relying on an infinite population sampling approximation.
If an Expressed Sequence Tag (EST) library is nonnormalized, its counting data, also known as "DigitalNorthern", reflects the abundance of genes. Likewise, the Massively Parallel Signature Sequencing (MPSS) [8] technique counts tags to infer the transcriptome, but using a completely different strategy from traditional DNA sequencing methods, that allows augmented highthroughput capability. Therefore, all the results discussed here are readily applicable to "DigitalNorthern" or MPSS context since, from a mathematical viewpoint, all represent the same bioinformatics problem: counting transcripts (as balls in urns).
Nowadays, the variability in SAGE abundance data is modeled only as due to sampling from sequencing, since almost all statistical procedures are performed after aggregation of observations from various libraries of the same class, creating a "pseudolibrary". See [9–11] for good reviews on statistical techniques used in SAGE analysis. This extensively used trick tacitly ignores the withinclass variability, i.e., the biological variability among individuals within a class (different patients having the same cancer diagnosis, for example), and could lead to overconfident conclusions.
Results
Here we propose a Bayesian model of mixtures to account for withinclass variability as a generalization of the BetaBinomial model [12]. We also show that the usual "pseudolibrary" construction is a particular case of our mixture model. Finally, we propose the use of the Bayes Error Rate to intuitively rank the differential expression hypothesis under a Bayesian framework, avoiding several technicalities and difficulties such as: typeI and typeII error analysis, Bonferronilike multiple testing correction, asymptotic results evocation, imposition of a test statistic and null probability density function (pdf), and so on.
Statistical model
The counting process from automatic sequencing of one single ith library is often modeled as a Bernoulli Process and a fixed unknown tag abundance π_{ i }is implicitly assumed. The pdf of the random variable of interest, "expression abundance" π ∈ [0;1] among all n libraries is unknown, thus each library could be regarded as being created by a realization of π. These features lead naturally to mixture models [13, 14]:
where: f(·) is the unknown pdf of the abundance among sameclass libraries parameterized by a vector θ, X = (x_{1},..., x_{ n }) is the vector of counts in all n libraries of same class, M = (m_{1},..., m_{ n }) is the vector of library sizes and L is the likelihood of each ith observation.
The common procedure of merging all observations from libraries of the same class, constructing a "pseudolibrary" before statistical inference, is recognized as a particular case of this mixture model: just assume that all libraries have strictly the same abundance, with no biological variability. Mathematically, this is a function with infinite probability density over one single abundance value π = θ and zero over every other π ≠ θ, or a Dirac's Delta function. Using f(·) as a Dirac's Delta function constrained to [0;1], turn Eq.1 into the familiar and commonly used binomial distribution (see derivation in the Methods section).
We believe that Dirac's Delta is a naive description of reallife SAGE libraries. The Beta distribution is an alternative with nonzero withinclass variance to account for intuitively expected biological differences among them. Using f(·) as a Beta in Eq.1, yields the socalled BetaBinomial model (see derivation in the Methods section).
Given the parameter vector θ that describes the random variable π of some fixed gene G, we must decide if there is a difference between A and B classes (e.g, tumor vs. normal classes). We propose to consider genes as being differentially expressed based on nonsuperposition of the predictive Beta pdfs of both A and B classes. By "predictive" we mean that we use the a posteriori mode in the Beta pdfs. The "nonsuperposition" intuitive feature is mathematically written as the Bayes Error Rate E [15]:
where f(·) is the Beta pdf and "hat" over the parameters indicate the values that lead to an a posteriori pdf maximum. The a posteriori distribution is obtained as usual from Bayesian Statistical Theory (a priori pdf choice and detailed derivation are in the Methods section).
Intuitively, if the pdfs are "far apart", the gene probably has reproducible differential expression between classes. In this case, rarely could one misclassify class A as B and viceversa. Figure 1 gives some insight about this fact. Using our proposed approach, the "far apart" notion means a small Bayes Error Rate E. For adepts of the Frequentist Statistics, this evidence measure could resemble a typeI and typeII errors sum, however it is just an illustrative analogy.
As in any significance test method, the experimenter must define what is a high significance E value. This cutoff should be guided by external and independent confirmatory assays. To avoid crude decision boundaries, one could rank their significance results but there is no way to avoid some arbitrariness in any kind of statistical test.
In the classical Frequentist Statistics framework, it is common to call a result as significant if it presents a Pvalue ≤ 0.01 in a tliketest, hoping that this could control the error at this level. However, due to technical difficulties such as lack of sensitivity of posterior confirmatory methods or high absolute expression (not differential expression) necessity, this apparent statistically sound results could be not useful in a pragmatic sense. That is why we prefer to rank the differential expression results and allow researchers to establish a cutoff compatible with their subsequent application for the selected genes, rather than split them based in assumptionderived errorrate cutoffs. People familiar with the Frequentist Statistics framework could miss multiple testing considerations, typeI/typeII error studies, and so on. However, in the Bayesian framework, several of these concerns are meaningless since we work with parameters space and not with sample space. The bayesians avoid statements about "data that could be observed but was not" and work only with available information (prior and experimental), extracting all possible information from data effectively observed.
For those genes classified as differentially expressed, one should aggregate intuitive information adding "errorbars" to expression ratios. Recently we have developed a method to add credibility intervals to gene expression ratio [16], which could improve posterior analyses such as clustering [17] or comparison with microarray data.
Comparison with available methods using publically available data
To show the model is usefulness, we applied it to a tumor vs. normal twoclasses comparison problem. We chose a subset of brain tumor SAGE data from The Cancer Genome Anatomy Project's SAGE Genie public database website [18]. The SAGE Genie performs several bioinformatics protocols to assure the quality of its data with systematic errors cleaning/correction [19].
We used all 4 available libraries in SAGE Genie until Jan/2004 from astrocytoma grade III tumors and almost all (except the fetal library) normal brain libraries (see Methods section for details about libraries).
We want to stress 3 typical and important cases: (i) when our measure agreed with other evidence measures accepting null hypothesis H_{0}, i.e., there is no evidence of differential behavior between tumor and normal classes; (ii) when our method agreed with others rejecting H_{0}, i.e., there is evidence of differential expression; and (iii) when our method showed evidence in favor but other evidence measures showed evidence against the H_{0}. Case (iii) is the main motivation of our method since it reveals situations that researchers may call a gene differentially expressed and, in fact, it could be not so significant if biological replicates are taken into account. The other evidence measures used were: the AudicClaverie bayesian evidence [20], the classical Fisher Exact Test Pvalue, and the classical χ^{2} Pvalue, all obtained using the IDEG6 webinterface [21, 22] (see Methods section).
A case (i) prototype is the TTTCAATAGA tag with X_{T} = (0, 2, 5, 8) and X_{N} = (1, 1, 0, 0, 0, 7, 2). The AudicClaverie, Fisher and χ^{2} methods yield Pvalues of 0.06, 0.44, 0.41, respectively, indicating low evidence against H_{0} for all mystical significance level cutoffs ≤ 0.01, ≤ 0.05 or ≤ 0.1. The Bayes Error Rate evidence is E = 0.61, an intuitively unacceptable superposition level between the normal and tumoral predictive Beta pdfs, showing that there is no separable behavior between classes. Figure 2a shows an obvious superposition between pdf and observations of this two classes.
A case (ii) prototype is the AAAAGAAACT tag with X_{T} = (7, 11, 18, 10) and X_{N} = (7, 1, 2, 1, 2, 0, 3). All Pvalues are 0.00 (zero), significant at any cutoff level. Our evidence is E = 0.03, showing safely that this gene behaves differentially between normal brain and astrocytoma grade III patients. Figure 2b shows that two Betas are apart from each other and, even observing clear withinclass variability, the expression is different.
A case (iii) prototype is the TTGGAGATCT tag with X_{T} = (7, 239, 244, 123) and X_{N} = (54, 27, 33, 21, 40, 196, 28). All Pvalues are 0.00 (zero), indicating significant difference between classes. On the other hand, our evidence E = 0.73 indicates high superposition between tumor and normal classes. Figure 2c shows that withinclass variability for tumor class is not negligible. It is obvious that individual libraries confound their results with normal brain libraries, and the Betas have a relatively high intersection. Using a common "pseudolibrary" approach, one is lead to call this gene as a strong discriminator between classes. We believe that this is a suspect conclusion.
There are several other obvious case (iii) examples, such as tag TACAGTATGT in Figure 2d, that received Pvalues < 0.01 from all other methods, and they are the main concern of our method since they may lead to waste of resources in clinical validation efforts of genes that, by SAGE itself, could be left behind in favor of other promising genes. All tag results are available as additional file and graphics for all tags are at the supplemental website [23].
One could think about a case (iv) when considering withinclass variability leads one to H_{0} rejection, but considering "pseudolibraries" leads to H_{0} acceptance. This seems to be inconsistent since one expects that, once H_{0} is accepted in a simplified model, it should also be accepted in the complete model. In fact, we do not observe such a situation, except by tags with Pvalues or Bayes Error Rate very close to arbitrarily defined cutoff values. We believe that these occurrences are just "edge effect" manifestations.
Discussion
In order to assure that we are dealing with a fundamental question in SAGE analysis, we show more insights analyzing the method's robustness using the same data but excluding "small" libraries. Also, we draw some parallels between our proposed method and the only available published solution for dealing with withinclass variability, a ttest approximation [12].
We used our method with all available libraries but some of them are smaller than 50,000 tags (see Table 1). In the SAGE community, libraries smaller than this arbitrary limit are considered "small". Several researchers claim that these are nonrepresentative and should be excluded from analysis. We observed several case (iii) tag examples which remain as case (iii) if we use libraries with size > 40,000 and > 50,000 (shown at the supplemental website only). Figure 3 shows a tag example analyzed in these tree setups and it is clear that inclusion of "small" libraries gave pretty much the same result, indicating robustness of our method against small class size variations and against "small" sized libraries. Moreover, these libraries are not always outliers from biological sampling but seem to be samples like any other. These results suggest that one can use the "small" libraries, jointly with non"small" ones, because biological variability seems to be greater than binomial sampling variability.
Obviously, we are not recommending to use only "small" libraries in SAGE analysis, but suggesting that our method is relatively robust. For low expression genes, the binomial sampling variability should become more relevant as the library size decreases. Also, the results obtained using two/three libraries could be very different from using just one. These proprieties could be tag dependent since some tags could be much noisier than others for biological reasons. Some "denoising" procedure could be used before application of our method [7]. Therefore, our findings should be carefully interpreted.
To prove that the incoherence of using "pseudolibraries" methods is not a prerogative of tags showing small foldchanges, we analyzed another three very illustrative examples: ATGGCAACAG, GGATGTGAAA, and GTATGGGCCC; which are case (iii) tags. These tags present high fold changes: 7.59, 8.15 and 25.80 foldchange respectively, augmented in pooled tumor libraries. Using the wellknown Fisher Exact test, χ^{2} classical test and the AudicClaverie's method, we get 0.00 (zero!) for all Pvalues of the no differential expression null hypothesis. Using the conceptually different Bayesian Pvalue implemented at SAGE Genie [24, 25] we obtain 0.01, 0.00 and 0.00 respectively for posterior probabilities of foldchanges greater than 4fold. Finally, using our own proposed measure, applied to the pool, we get E = 0.00 meaning no superposition between the two classes pdfs. All these results indicate strong significance in differential expression of these tags.
However, if we consider withinclass variability, the test proposed by Baggerly et al. [12] yields 0.08, 0.07 and 0.15 respectively for ttest Pvalues, and our method yields Bayes Error Rates of 0.38, 0.37, 0.43 respectively; indicating not so significant evidence in favor of the differentially expressed hypothesis. A closer look at the graphics of these tags induces one to believe that there is no reproducible differential expression because several observations of tumor and normal are superimposed (all graphics available at supplemental website [23]).
Since we show clearly that methods that use "pseudolibrary" aggregation could be incoherent in some cases, a natural question is how our proposed method performs compared to the only published solution that accounts for withinclass variability, the Baggerly et al. [12]ttest approximation. Without knowing the true state of all tags, it is impossible to carry out a serious benchmark. Since the interpretation of evidence measures is very different, the performance could be subjected to an arbitrary cutoff selection for each method. Figure 4 shows a scatterplot of evidence measures obtained for each of the two methods.
It is clear from this graphic that there are many more tags considered as differentially expressed using our method than the ttest approximation, considering E ≤ 0.1 and Pvalue ≤ 0.01. There are also some tags selected by ttest and ignored by ours. It is impossible to know which method perform better without the true unknown status of those tags. Looking at individual libraries results, constructed as depicted in Figure 2 for example, could help in this analysis but this is a subjective procedure.
It is important to bear in mind that a difficulty is hidden in the Beta modeling imposed in the very first beginning. If Beta is not a good model for an unknown biological behavior, then some apparent inconsistency could appear in both Baggerly et al. [12] and our approaches. However, our general mixture model allows another propositions. Other simplex constrained pdfs, different from Beta, exist but the tractability is much more difficult [26]. We believe that to build a fully nonparametric approach to this problem is a very hard issue, but should be considered as a future challenge.
Conclusion
Until now, almost all statistical methods for SAGE data analysis tacitly ignore the withinclass variability. To our knowledge, the firsts to formally address this issue was Baggerly et al. [12] who introduced the BetaBinomial model as the correct way to model the probability of counting tags instead of Binomial models. They also proposed a tlike statistics, outlined a possible hypothesis test using the classical Frequentist Statistics framework and evoked some asymptotic results for t pdf justification.
In this work we presented the Bayesian alternative for this problem and defined a theoretical model that views Baggerly's BetaBinomial approach or even the common Binomial approach as particular cases of mixture models. Other models are possible modifying the mixing distribution, such as BetaPoisson [14], or using other simplex constrained pdf [26] to model expression abundance. At last, but not at least, we proposed a method for ranking differentially expressed genes between two classes using the Bayes Error Rate as an intuitive measure of separation between the classes pdfs, avoiding statistical test formalism and its conceptual/practical difficulties.
We show that there are cases in which approaches that ignore withinclass variability will lead to high significance in differences between tumor and normal classes, but looking carefully at individual observations jointly, one should not attribute such high significance to them since abundance probability density functions have considerable superposition.
In conclusion, we recommend that withinclass variability must be taken into account in any statistical analysis of SAGE data if replicates are available. We suggest that biological replication should be considered in planning new SAGE experiments.
Methods
General bayesian model
To start generically, suppose that the probability density function (pdf) for the random variable of interest "expression abundance" π ∈ [0;1] of some gene G is indexed in a model family by means of a parameter vector θ. Therefore, following the usual Bayesian framework, the a posteriori pdf that describes the class is:
where: X = (x_{1},..., x_{ n }) is the vector of counts in all n libraries of same class, M = (m_{1},..., m_{ n }) is the vector of total observations in all n libraries of same class, g(·) is the a priori pdf, and L is the likelihood of each ith observation. Note that the product of all likelihood functions over all observations is the socalled Likelihood Function.
The counting process from automatic sequencing is often modeled as a Binomial. Since the sample size and the stopping rule are not known in advance the model is not strictly Binomial. We do not need the combinatorial constant in the model, but we write it just because it is commonly used and will vanish in posteriori expression anyway.
"PseudoLibrary" method as particular case
Merging all observations from the same class libraries and constructing "pseudolibraries", with the sum of their components, is the standard procedure to use replicates. Our general model is reduced to this (unrealistic) one if one uses f(·) as a Dirac's Delta in Eq.1:
where: 1{·} is the indicator function.
Using Eq.4 in Eq.3 yield:
where: g(θ) = 1, the noninformative uniform a priori distribution.
The expert recognizes that θ ~ Beta(1 + Σ x_{ i }, 1 + Σ m_{ i } Σ x_{ i }), and the sum of observations is the mathematical translation of "pseudolibraries" construction.
BetaBinomial method as particular case
The only published solution that allows nonzero withinclass variance in SAGE analysis is the BetaBinomial model [12]. Using f(·) as a Beta in Eq.1 we get the BetaBinomial model as a particular case of general model:
where: B(·) is the beta special function, and:
Again, an expert recognizes θ = (θ_{1}, θ_{2}) as the mean and standard deviation (stdv) of a Beta random variable. We prefer this parameterization of Beta distributions instead of the common (α, β) one because: (i) it is much more intuitive to biologists to deal with mean and stdv than with abstract α and β, and (ii) as α, β > 0, the domain Θ = {(θ_{1}, θ_{2}): 0 ≤ θ_{1} ≤ 1, 0 ≤ θ_{2}^{2} <θ_{1} (1 θ_{1}) ≤ 1/4} is bounded and much more amenable to perform the necessary numerical computations.
Using Eq.6 in Eq.3 yield:
where: g(θ_{1}, θ_{2}) is the priori pdf.
A Priori distribution definition
To complete a Bayesian model, it is necessary to choose the a priori pdf. We use an uniform distribution over Θ. On the other hand, we know in advance that variance of this model cannot be smaller than the variance eventually obtained if we do not consider withinclass variability. Even if the withinclass variability is very small, it cannot be estimated as being smaller than the simple sampling error because they are inseparable, and sampling error is the lower limit [12]. As an illustration, the same situation could occur if one takes several diameter measurements of a folded paper ball and a perfect sphere using a common ruler. In the first case, the intrinsic nature of the measured object dominates the measurement variability but, in the second case, we cannot know the diameter of the perfect sphere with better precision than our ruler can measure.
This kind of knowledge is naturally incorporated in Bayesian statistics by means of a priori distributions. To match our desired features, it is sufficient to define an uniform over the Θ parameter space but constrained at a minimum stdv σ, obtained from the result of no withinclass variance model:
over domain Θ = {(θ_{1}, θ_{2}): 0 ≤ θ_{1} ≤ 1, 0 ≤ θ_{2}^{2} <θ_{1} (1  θ_{1}) ≤ 1/4}.
Since we showed (Eq.5) that the no withinclass variance model is θ ~ Beta(1 + Σx_{ i }, 1 + Σm_{ i } Σx_{ i }), it is easy to obtain our lower stdv boundary from Beta variance:
Therefore, using Eq.9 and Eq.10 in Eq.8, our posteriori is completely defined.
Differential expression detection
We detect a tag as differentially expressed using the Bayes Error Rate E [15] in both predictive Beta pdfs:
where:
Note that f(·) is the Beta pdf, as in Eq.6 development. The "hat" over θ = (θ_{1}, θ_{2}) indicates values that leads Eq.8 to maximum. As usual, the maximization, subject to constrain Θ defined previously, is made upon logarithm of posteriori's core since it gives the same estimates as the posteriori itself:
Figure 5 shows an example of this process. See Results section to get an intuitive notion of this evidence measure.
Implementation – numerical analysis
The method was implemented as R language [27, 28] scripts which are freely available under GPL/GNU copyleft license at supplemental web site [23]. At this web page there are details on how to run it locally.
Our method is computerintensive mainly because some numeric maximization and integration are needed. We used efficient R builtin routines to perform such numerical tasks. Remember that maximization needed in Eq.12 is constrained, thus we used simply auxiliary reparameterization to obtain linear constrains and used the constrOptim R routine. For numerical integration we used the 1dimensional gaussian quadrature integrate R builtin function. Although numerical integration of Eq.9 should be performed in all [0;1] support, the relevant contribution for this integral is concentrated in a much smaller region. Integrating over the formal limits will cause serious numerical errors, and to avoid this problem we approximate our integration region to an interval delimited by 0.005 and 0.995 quantile of each Beta pdf since the relevant density lie in there.
The credibility intervals ("errorbars") for the expression ratio of interesting tags were obtained as described in our recent work [16]. We chose arbitrarily 68% credibility intervals.
Implementation – Web based interface
We have also developed an easytouse webbased service that performs all calculations at our server and provides passwordprotected results. Although desirable, for the sake of automatic web hyperlink with SAGE Genie database, it is not necessary to explicitly identify the tags analyzed but rather any (custom) i.d. string. This could increase privacy or make our webinterface useful for "DigitalNorthern", MPSS or any mathematically related problem of mixtures from binomial sampling. Figure 6 shows snapshots of the interface.
Publically available data
The Table 1 list the SAGE Genie's library name, Gene Expression Omnibus (GEO) [29] accession code and size of all used libraries.
For our aims, it is sufficient to focus the analysis at the tag level. Thus, we process the tag counts and let the identification of tag's best gene match as a posterior question that could be carefully done only to really interesting tags. We choose not to process tags whose counts appear only in libraries of one class. It is important to note that all libraries are from bulk material, without celllines, and came from patients with similar disease description. The normal libraries came from different normal regions of the brain.
We think that this data set is very illustrative since there are biological replicates in the tumor class allowing clear verification of withinclass biological variability. On the other hand, taking only one kind of disease, astrocytoma grade III, instead of all brain tumors in the database, leads one to believe that the withinclass variability is in fact due to biological diversity of the patients and not due to very distinct molecular profile of distinct brain tumors stored in SAGE Genie's database.
Therefore, we believe that this in silico comparison is wellsuited to demonstrate the necessity of dealing with withinclass effect, although it is not our aim here to make a detailed or biological analysis of brain tumor data.
Comparison with other methods
To bring some intuition about our differential expression evidence measure, we tabulated evidence measure obtained from the famous Fisher Exact Test, the classical Pearson's χ^{2} proportion test and the bayesian AudicClaverie's method. All these tests were performed using the easytouse webinterface IDEG6 [21, 22].
The "Pvalues" are conceptually very different from our evidence measure but are the most used evidence measures. Although numbers cannot be compared, the conclusions obtained from these methods should be since graphical representation of each library observation gives clear indication of incoherence of "pseudolibrary" methods. The results of the significance measures for all tags are available as additional Excel^{©} or OpenOffice^{©} interactive files in which the user can set cutoffs for the significance measures, and explore the differences in conclusions.
We carry out a qualitative comparison of our method with Baggerly et al. [12]ttest approximation in a graphical way since it is impossible to judge them without the unknown true status of analyzed tags, given the too different interpretation of numeric values returned. In their Frequentist framework, the estimator p_{ i }= x_{ i }/m_{ i }is used for π_{ i }and a linear combination of these abundances is proposed as the correct way to combine results from different libraries:
where w_{ i }are the weights that yield an unbiased minimum variance estimator V_{ u }for weighted proportion's variance and θ = (α, β) are the Beta pdf parameters. However, this unbiased variance could be unrealistically small when it becomes smaller than the sampling variability. We know that the variance of this model cannot be smaller than the variance eventually obtained if we do not consider withinclass variability. Therefore, they propose the final ad hoc estimator:
V = max [V_{ u }; V_{ pseudolib }] (14)
where:
The max(·) function assure that V is not unrealistic small when V_{ u }is unrealistic small. To fit all these parameters, they used the computationally practical Method of Moments. Once p_{A}, p_{B}, V_{A} and V_{B} are found for classes A and B, these authors test if the proportions are significantly different proposing the use of a t_{ w }statistics as following a Student's t_{ df }pdf:
Abbreviations
 SAGE:

Serial Analysis of Gene Expression
 MPSS:

Massively Parallel Signature Sequencing
 EST:

Expressed Sequence Tag
 pdf:

probability density function
 GEO:

Gene Expression Omnibus
References
 1.
Velculescu VE, Zhang L, Vogelstein B, Kinzler KW: Serial analysis of gene expression. Science 1995, 270: 484–487.
 2.
Margulies EH, Kardia SL, Innis JW: Identification and prevention of a GC content bias in SAGE libraries. Nucleic Acids Res 2001, 29: e60. 10.1093/nar/29.12.e60
 3.
Stern MD, Anisimov SV, Boheler KR: Can transcriptome size be estimated from SAGE catalogs? Bioinformatics 2003, 19: 443–448. 10.1093/bioinformatics/btg018
 4.
Stollberg J, Urschitz J, Urban Z, Boyd CD: A Quantitative Evaluation of SAGE. Genome Research 2000, 10: 1241–1248. 10.1101/gr.10.8.1241
 5.
Akmaev VR, Wang CJ: Correction of sequence based artifacts in serial analysis of gene expression. Bioinformatics 2004, 20: 1254–1263. 10.1093/bioinformatics/bth077
 6.
Morris JS, Baggerly KA, Coombes KR: Bayesian shrinkage estimation of the relative abundance of mRNA transcipts using SAGE. Biometrics 2003, 59: 476–486. 10.1111/15410420.00057
 7.
Blades NJ, Jones JB, Kern SE, Parmigiani G: Denoising of data from serial analysis of gene expression. Bioinformatics, in press.
 8.
Brenner S, Johnson M, Bridgham J: Gene expression analysis by massively parallel signature sequencing (MPSS) on microbead arrays. Nature Biotechnology 2000, 18: 630–634. 10.1038/76469
 9.
Man MZ, Wang X, Wang Y: POWER_SAGE: comparing statistical tests for SAGE experiments. Bioinfo matics 2000, 16: 953–959. 10.1093/bioinformatics/16.11.953
 10.
Romualdi C, Bortoluzzi S, Danieli GA: Detecting differentially expressed genes in multiple tag sampling experiments: comparative evaluation of statistical tests. Hum Mol Genet 2001, 19: 2133–2141. 10.1093/hmg/10.19.2133
 11.
Ruijter JM, Kampen AHC, Baas F: Statistical evaluation of SAGE libraries: consequences for experimental design. Physiol Genomics 2002, 11: 37–44.
 12.
Baggerly KA, Deng L, Morris JS, Aldaz CM: Differential expression in SAGE: accounting for normal betweenlibrary variation. Bioinformatics 2003, 19: 1477–1483. 10.1093/bioinformatics/btg173
 13.
Aitchison J, Dunsmore IR: Statistical Prediction Analysis Cambridge: Cambridge University Press 1975.
 14.
Bueno AMS, Pereira CAB, RabelloGay MN, Stern JM: Environmental genotoxicity evaluation: Bayesian approach for a mixture statistical model. Stochastic Environmental Research and Risk Assessment 2002, 16: 267–278. 10.1007/s004770020100x
 15.
Duda RO, Hart PE, Stork DG: Pattern Classification 2 Edition New York: WileyInterscience Press 2000.
 16.
Vêncio RZN, Brentani H, Pereira CAB: Using credibility intervals instead of hypothesis tests in SAGE analysis. Bioinformatics 2003, 19: 2461–2464. 10.1093/bioinformatics/btg357
 17.
Yeung KY, Medvedociv M, Bumgarner RE: Clustering geneexpression data with repeated measurements. Genome Biology 2003, 4: R34.1R34.17.
 18.
SAGE Genie Home Page[http://cgap.nci.nih.gov/SAGE]
 19.
Boon K, Osório EC, Greenhut SF, Schaefer CF, Shoemaker J, Polyak K, Morin PJ, Buetow KH, Strausberg RL, Souza SJ, Riggins GJ: An anatomy of normal and malignant gene expression. Proc Natl Acad Sci USA 2002, 99: 11287–11292. 10.1073/pnas.152324199
 20.
Audic S, Claverie J: The significance of digital gene expression profiles. Genome Research 1997, 7: 986–995.
 21.
Romualdi C, Bortoluzzi S, D'Alessi F, Danieli GA: IDEG6: a web tool for detection of differentially expressed genes in multiple tag sampling experiments. Physiol Genomics 2003, 12: 159–162.
 22.
IDEG6 Home Page[http://telethon.bio.unipd.it/bioinfo/IDEG6_form/]
 23.
SAGEbetaBin Home Page[http://www.vision.ime.usp.br/~rvencio/SAGEbetaBin]
 24.
Lal A, Lash AE, Altschul SF, Velculescu V, Zhang L, McLendon RE, Marra MA, Prange C, Morin PJ, Polyak K, Papadopoulos N, Vogelstein B, Kinzler KW, Strausberg RL, Riggins GJ: A public database for gene expression in human cancers. Cancer Research 1999, 21: 5403–5407.
 25.
Chen H, Centola M, Altschul SF, Metzger H: Characterization of gene expression in resting and activated mast cells. J Exp Med 1998, 188: 1657–1668. 10.1084/jem.188.9.1657
 26.
Aitchison J: A General class of distributions on the simplex. J R Statistical Soc B 1985, 47: 136–146.
 27.
Ihaka R, Gentleman R: R: A language for data analysis and graphics. Journal of Computational and Graphical Statistics 1996, 5: 299–314.
 28.
R project Home Page[http://rproject.org]
 29.
Gene Expression Omnibus Home Page[http://0www.ncbi.nlm.nih.gov.brum.beds.ac.uk/geo]
Acknowledgements
RV is supported by FAPESP 02/046988 fellowship. We thank Tie Koide for critical reading of the manuscript and BIOINFOUSP/RedeVision for computational support.
Author information
Additional information
Authors' Contributions
RV conceived and executed this work. HB helped with all biological issues. DFCP helped in differential expression detection methods and implemented the online webbased tool. CABP helped with Bayesian statistics and proposed the mixture ideas.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
About this article
Cite this article
Vêncio, R.Z., Brentani, H., Patrão, D.F. et al. Bayesian model accounting for withinclass biological variability in Serial Analysis of Gene Expression (SAGE). BMC Bioinformatics 5, 119 (2004) doi:10.1186/147121055119
Received:
Accepted:
Published:
DOI: https://0doiorg.brum.beds.ac.uk/10.1186/147121055119
Keywords
 Massively Parallel Signature Sequencing
 Binomial Sampling
 Evidence Measure