 Methodology article
 Open Access
 Published:
Testing allele homogeneity: the problem of nested hypotheses
BMC Genetics volume 13, Article number: 103 (2012)
Abstract
Background
The evaluation of associations between genotypes and diseases in a casecontrol framework plays an important role in genetic epidemiology. This paper focuses on the evaluation of the homogeneity of both genotypic and allelic frequencies. The traditional test that is used to check allelic homogeneity is known to be valid only under HardyWeinberg equilibrium, a property that may not hold in practice.
Results
We first describe the flaws of the traditional (chisquared) tests for both allelic and genotypic homogeneity. Besides the known problem of the allelic procedure, we show that whenever these tests are used, an incoherence may arise: sometimes the genotypic homogeneity hypothesis is not rejected, but the allelic hypothesis is. As we argue, this is logically impossible. Some methods that were recently proposed implicitly rely on the idea that this does not happen. In an attempt to correct this incoherence, we describe an alternative frequentist approach that is appropriate even when HardyWeinberg equilibrium does not hold. It is then shown that the problem remains and is intrinsic of frequentist procedures. Finally, we introduce the Full Bayesian Significance Test to test both hypotheses and prove that the incoherence cannot happen with these new tests. To illustrate this, all five tests are applied to real and simulated datasets. Using the celebrated power analysis, we show that the Bayesian method is comparable to the frequentist one and has the advantage of being coherent.
Conclusions
Contrary to more traditional approaches, the Full Bayesian Significance Test for association studies provides a simple, coherent and powerful tool for detecting associations.
Background
One of the main goals in genetic epidemiology is the evaluation of associations between specific genotypes or alleles and a certain disease. Association studies are usually performed in a casecontrol framework in which one or several polymorphisms of candidate genes are evaluated in a group of cases (that is, patients that have a disease) and in a group of controls from the same population (that is, healthy individuals) [1]. The frequencies of each of the genotypes are then computed so that statistical tests that aim at checking for associations between genes and the disease can be performed. The population studied usually must be homogeneous regarding ethnicity, gender distribution and other factors that may bias the results rendering falsepositive associations. See [2] for nontechnical summary of reasons that may render false discoveries in casecontrol studies and [3] for a theoretical analysis of the consequences of population stratification. For more on casecontrol studies, the reader is referred to [4].
Several statistical tests are usually employed for this scenario. Among them, CochranArmitage test for trends [5], homogeneity chisquare tests for contingency tables of both genotypic and allelic frequencies [6], likelihood ratio tests and Wald tests [7] are performed. See, for example, [8] and [9] for a summary of these tests. Some of these statistics are specifically designed to work under assumptions such as dominance models, recessive models or HardyWeinberg Equilibrium (HWE). However, a big importance is being given on new methods that are robust to model misspecification, mainly because power is usually small when the model is wrong and type 1 error rates are usually incorrect (see e.g. [7, 10, 11]).
HWE plays an important role in genetic studies, in particular when testing for allelic homogeneity [12]. The main reason is that the traditional test for allelic homogeneity fails when HWE does not hold, a point to which we will get back later. In words, HWE is a constrain on the genotypic proportions that implies, under some assumptions, stability of the different genotypes over the generations of the population (see e.g. [13] and [14]). These assumptions include, for example, random mating between individuals. For many diseases, random mating is not expected to be satisfied. The same holds for other conditions required for HWE, that in practice may be unrealistic in some situations. In fact, as stated by [15], “a population will never be exactly in HWE”. Hence, the need to design tests that are robust to departures from HWE is evident. A common practice in such problems is to first test HWE, discarding genes that are not in equilibrium. This is done in an attempt of identifying genotyping errors. Such an approach should be avoided, as discussed by [12]. The main reason is that the 2 steps procedure alters type1 errors. They also emphasize that the correct way to deal with this problem is to inherently account for deviations from HWE with adjusted tests, the approach we take here.
In the present paper, we focus on two hypotheses: 1. homogeneity of the genotypic frequencies; and 2. homogeneity of the allelic frequencies. Usually, data in such studies are summarized in two different ways [9]. The first one consists of a table with the genotypic frequencies of case and control groups. The second, a table with the allelic frequencies. Tables 1 and 2 illustrate this representation using data presented in [16], which was also considered by [12]. Their study was designed to test the hypothesis that GABA_{ A } sub genes would contribute to a disorder due to methamphetamine use. It is worth noting that Table 2 contains twice as many observations as Table 1. [9] discusses in details the problem of doubling the sample size. In particular, it is shown that methods that “treat alleles as individual entities” [9] have wrong nominal type1 errors when HWE does not hold. We must recall that the power of a test can increase considerably by increasing the sample size, a nominal increase that can be misleading when it is not reasonable to “treat alleles as individual entities”. This issue we will be discussed in further details later.
The aims of this paper are fourfold: 1  to describe how the analysis of such data is usually conducted and to emphasize its known flaw (namely lack of robustness to departures from HWE); 2  to describe one exact frequentist approach which is correct from a classical point of view; 3  to present a Bayesian method to deal with the problem, and 4  to advocate the use of the Bayesian solution by demonstrating why this is the best solution compared to the others. The main argument is based on an undesirable logical inconsistency that can happen whenever pvalues are used to test nested hypotheses. We prove that this does not happen when using the Bayesian method proposed. We also show that the Bayesian and the correct frequentist solutions have comparable power. Simulations and analyses of real data are shown in order to illustrate the problem.
The paper is organized as follows. Section Methods contains three subsections: Usual Procedures, which introduces the notation that is used throughout the paper, discusses the usual methods to deal with the problem and argues why the test for allelic homogeneity is wrong when there are departures from HWE; A Different Frequentist Test, which introduces a frequentist test that works even when departures of HWE happen and Bayesian Solution, which introduces the FBST approach to solve the problem. Section Results and Discussion first focuses on the issue of the logical incoherence that happens when using the frequentist procedures discussed in the paper and also shows that the same does not happen to the Bayesian method FBST. A brief discussion on Bayes factors is also provided. Finally, we address the question of whether the Bayesian method has good frequentist properties. Section Conclusions summarizes the findings of the paper.
Methods
Here, we formally describe three different approaches to deal with the problem described: the usual procedure, a correct frequentist proposal and a Bayesian solution.
Usual Procedures
We begin by describing the statistical model that is used to deal with the problem approached in this paper (namely, product of multinomials) and also how the hypotheses of interest are usually tested in genetic literature. For more details, see [9].
Let \left(\right)close="">\n \n G\n =\n {\n AA\n ,\n AB\n ,\n BB\n }\n \n be the set of all possible genotypes for the locus of interest. As in Table 3, denote by X= (X_{ AA },X_{ AB },X_{ BB }) and Y= (Y_{ AA },Y_{ AB },Y_{ BB }) the random vectors with the genotypic frequencies from the case and control groups sample, with \sum _{i\in \mathbb{G}}{X}_{i}=n and \sum _{i\in \mathbb{G}}{Y}_{i}=m being the total of individuals observed in each group. Also, let γ= (γ_{ AA },γ_{ AB },γ_{ BB }), where γ_{ i }is the probability that an individual from the case group has genotype i, and Π= (Π_{ AA },Π_{ AB },Π_{ BB }), where Π_{ i }is the probability that an individual from the control group has genotype i, i\in \mathbb{G}. The parametric space is
Considering observations from different individuals to be statistically independent, we have Xθ ∼ Multinomial(n,γ) and Yθ ∼ Multinomial(m,Π) with it X and it Y being conditionally independent as well. The likelihood function is then given by
which is the product of two multinomial distributions.
The first hypothesis to be tested (null hypothesis), namely that there is no difference in genotypic frequencies between the groups, may be formally expressed as
The usual procedure to test {H}_{0}^{G} is the chisquare test, i.e., the test based on the statistic
where {\widehat{\theta}}_{i}^{G} is the maximum likelihood estimator for the genotypic frequency i under the hypothesis {H}_{0}^{G}. Under {H}_{0}^{G}, Q^{G} has asymptotic distribution {\chi}_{2}^{2} (chisquare distribution with 2 degrees of freedom). Using this fact, it is possible to calculate an asymptotic pvalue. If one prefers exact tests, Monte Carlo methods can also be used. To sum up, in order to test the first hypothesis, one usually performs a traditional chisquare test of homogeneity to Table 3.
The second hypothesis states that there is no difference in allelic frequencies between the groups. This hypothesis  which will be made formal in the next section  is usually tested by considering the allelic frequencies in both samples, X_{ A }= 2X_{ AA } + X_{ AB } and Y_{ A }= 2Y_{ AA } + Y_{ AB }, as in Table 4 and applying the chisquare test of homogeneity to that table, which has twice as many observations as Table 4.
More formally, the statistic considered is
where {\widehat{\lambda}}_{i}^{A} is the maximum likelihood estimator for the allelic frequency i under the hypothesis that allelic frequencies are the same in both groups. This statistic is then compared to a {\chi}_{1}^{2} distribution, or sampled using a Monte Carlo method to calculate the pvalue. However, in this scenario, the distribution of the test statistic under the null hypothesis is not chisquare unless alleles are statistically independent. In other words, the distribution is chisquare only if a product multinomial model can be applied to Table 4. Essentially, this independence corresponds to the HWE. In fact, [9] formally proves that this is a valid test if, and only if, both groups, case and control, are under HWE. Otherwise, this test is biased: nominal level of significance is different from the real one [17]. [17] also shows how deviations from HWE alter typeI error rates, a point that will also be illustrated in Section Results and Discussion. Therefore, this test should not be used. It is important to note that despite being wrong, it is still widely used in genetic literature nowadays (see e.g. [18], that also discusses some aspects of the lack of robustness of this test). This leads to a larger number of false conclusions than the nominal errors of the procedures.
Applying the traditional tests to data from Table 1, one gets a pvalue of 0.152 for genotypic association and of 0.049 for allelic association. This means that the evidence we have that the two groups are in genotypic homogeneity is larger than the evidence we have that they are in allelic homogeneity. However, if genotypic proportions are the same, allelic proportions must also be the same. This implication will be made formal in Section Results and Discussion. In practice, the first pvalue being larger than the second implies that one can accept the hypothesis of genotypic homogeneity while rejecting allelic homogeneity, which is a contradiction. For instance, this is the case when the level of significance is 10%, as 0.152 > 0.1 but 0.049 < 0.1. To summarize: we are testing two nested hypotheses, that is, the nature of the problem is such that the first hypothesis implies the second. However, even though we reject the second, we do not reject the first. Does the contradiction happen because the allelic test is wrong? Next Section answers this by presenting an exact test for allelic homogeneity that is valid even if HWE does not hold.
A Different Frequentist Test
Some attempts to correct the abovementioned allelic test so that it works even when HWE assumption is not met are considered by [8, 12, 17, 19]. See [18] for a summary of these. Also, see [20], that proofs that the test proposed by [17] to correct for departures of HWE is equivalent to the one proposed by [9], which is the Armitage’s trend test [5].
Here we show another solution that has the advantages of being exact, unconditional, and that it can also be calculated in a computationally efficient way, even for large data sets. Moreover, it is defined in the same parametric space Θ as the genotypic test. Essentially, this test is derived by noticing that the hypothesis that allele frequencies are the same in both groups can be written in terms of the original parametric space as
Note that this formulation is always true independent of the HardyWeinberg equilibrium restriction and does not involve changing neither the sample space nor the parametric space.
The chisquare statistic may be used to test this hypothesis:
Here, {\widehat{\gamma}}_{i}^{A\ast} and {\widehat{\Pi}}_{i}^{A\ast} are the maximum likelihood estimators of genotypic frequencies under {H}_{0}^{A}. They can be found by maximizing
and then using the relations
Maximization of Equation (3) can be efficiently done by using numerical methods such as Newton’s method [21], which are already implemented in most statistical and mathematical softwares such as R and MATLAB. To calculate pvalues, the statistic Q^{A∗} can then be compared to a {\chi}_{1}^{2} distribution or, if one wishes to perform an exact test (the approach we take here), sampled using Monte Carlo methods. That is, one can generate several values of Q^{A∗}under the null hypothesis and compute the proportion of these that are larger than the observed statistic on the sample. This is the (estimate of the) exact pvalue. Confidence intervals can be obtained for it by using a normal approximation to the binomial distribution. Note that the dimension of the parametric space is 4 and under the null hypothesis it becomes 3. Hence, the number of degrees of freedom of the distribution of the chisquare statistic is \text{dim}(\mathrm{\Theta})\text{dim}({H}_{0}^{A})=43=1. This is also the number of degrees of freedom of the chisquare for the allelic test described before.
This test is very similar to the ones recently introduced by [7], except that the statistics used are different (Wald statistic, score statistic and maximum profile likelihood ratio), and results are asymptotic: chi squared approximation is used. Even though these tests are asymptotically equivalent, in order to illustrate our points it is important to have exact tests here.
The allelic pvalue for data from Table 3 is 0.069. It is surprising that despite the fact this test is correct, this pvalue is still smaller than 0.152  the pvalue for genotypic association. Hence, incoherence remains even when correcting the traditional allelic test. We note that the pvalue found by [12] for this same data set using corrected allelic test is 0.066, which also does not remove the contradiction. Note that here we use the exact test, hence this is not a problem of using an approximation. In the Section Results and Discussion we present other data sets in which this incoherence happens, showing that this problem is not unique to the particular data we chose to illustrate the point. Next Section is devoted to present a framework where this kind of contradiction does not happen.
Bayesian Solution
Bayesian methods are the alternative inductive way to deal with such a problem. These methods are widely used nowadays because they allow prior knowledge from the researcher and scientific community to be incorporated into the analysis (see [22] for applied examples of these methods in genetics) and, contrary to usual classical procedures, they do not require large samples for the analysis to be correct. That is, optimality of the procedure does not rely on asymptotic considerations. Many Bayesian methods designed to deal with precise hypotheses, i.e., hypotheses which have lower dimension than the parametric space, have been developed. Precise hypotheses must have a different treatment in Bayesian statistics: in general, they have zero posterior probability, so that they would always be rejected when using traditional methods. One way to deal with this problem is to assign a positive prior probability to null hypothesis [23], but this may seem a rather ad hoc solution and may lead to some inconsistencies [24]. Another approach is to use Bayes factors [25], a point to which we will get back later in the paper (see Section Results and Discussion).
In this paper, we choose to use the FBST (Full Bayesian Significance Test), a procedure introduced by [26]. This method was also used by [27]. The test is based on the evalue statistic, a Bayesian measure of evidence designed to evaluate sharp null hypotheses. In order to apply this method, we begin by specifying a prior density in the complete parametric space Θ, f(θ). We note that it is not necessary to attribute different probability to each of the hypothesis: it is only necessary to specify f(θ). This is not the case for Bayes factors, where specification of different probability distribution inside each of the hypothesis of interest is needed. After observing data x, let f(θx) be the posterior density of the parameter θ. The posterior density is given by
Suppose one is interested in testing the null hypothesis H: θ∈ Θ_{0}. Define the Tangential Set to the null hypothesis as
The measure of evidence proposed, the itevalue, is defined by
In words, evalue is the posterior probability of the subset of the parametric space consisting of points with lower posterior density than the maximum achieved under H. It is interesting to note the duality between pvalues and evalues: while the former are tails in the sample distribution from the observed values under the null hypothesis, the latter are tail areas in the posterior distribution from the sharp hypothesis. Evalues are easy to be calculated and successful papers that use FBST procedure in genetics include [14, 28, 29]. For more on evalues, the intuition behind it, asymptotic consistency results, and decisiontheoretic considerations see [30]. High evalues indicate high evidence in favor of the hypothesis, while low evalues indicate that the hypothesis is false.
Implementation of the FBST procedure requires two simple steps, which can be performed numerically:

Optimization  Finding the supremum of the posterior distribution under the null hypothesis, {\text{sup}}_{{\mathrm{\Theta}}_{0}}f(\mathit{\theta}\mathit{x}). This is usually done by using builtin functions from statistical packages such as R.

Integration  Integrating the posterior density over the Tangential Set, {T}_{x}^{H}. This step can be done by sampling from the posterior distribution by using methods such as MCMC. For the problem considered here, a usual Monte Carlo method is enough to efficiently sample from the posterior.
More details on the implementation of the FBST procedure can be found in [26]. To perform the complete FBST procedure one also needs to set a cutoff point, that is, one must say what a “small” evaluemeans. Several approaches are available:

Empirical power analysis [31]

Reference sensitivity analysis and paraconsistent logic [32].

[30] relate evalues to pvalues.

Bayesian decisiontheoretic approach [33], by the specification of a loss function that gives origin to FBST procedure.

An asymptotically consistent threshold for a given confidence level ([34] and [31]).
The prior distribution for γin the routine that was implemented and is available in the website is a Dirichlet distribution, as well as the prior distribution for Π. The family of the Dirichlet priors is widely used in this scenario once it is both broad enough to contemplate a huge number of different possible prior information and yet very easy to be dealt with both mathematically and computationally. Here, the two priors are considered to be independent, and in the implementation we provide online (see Conclusions) the (hyper)parameters (a_{ AA },a_{ AB },a_{ BB },b_{ AA },b_{ AB },b_{ BB }) are set by the user. That is,
Note that in this case the posterior distribution is also the product of two independent Dirichlet distributions (once they are conjugate with the multinomial distribution). Their parameters are (x_{ AA } + a_{ AA },x_{ AB } + a_{ AB },x_{ BB } + a_{ BB }) and (y_{ AA } + b_{ AA },y_{ AB } + b_{ AB },y_{ BB } + b_{ BB }) respectively. Simulation of the Dirchlet distribution can be efficiently done by sampling from Gamma distributions; see [35] for details. Note that the case where all (hyper)parameters (a_{ i } and b_{ i }) are equal to 1, θ is uniformly distributed a priori.
The FBST procedure can be used in general, not only for testing allelic homogeneity. In particular, it can be used to test HardyWeinberg equilibrium, as shown by [26] and [14]. In order to illustrate the procedure, Figure 1 shows the HardyWeinberg hypothesis line and the Tangential Sets for both case and control groups from the data set presented by [16]. Figure 2 does the same for simulated data (not under HardyWeinberg equilibrium). They also show the 99% HPD (Highest Posterior Density) sets. For the sake of neutrality, the prior distribution we use is the product of two independent Dirichlet distributions with parameters (1,1,1), i.e., the uniform distribution in Θ. Hence, the posterior distribution is proportional to the likelihood function, that is,
We see that while both groups from Figure 1 seem to be under HWE (in this case, tangential sets have small probabilities, and therefore evalues are large), the ones from Figure 2 seem to be far from the equilibrium (in this case, tangential sets have large probabilities, and therefore evalues are small).
When testing genotypic and allelic homogeneity using FBST and uniform priors (a_{ i }= b_{ i }= 1 for all i’s in the Dirichlet distribution), we obtain evalues of 0.434 and 0.493 respectively. Hence, contrary to what happens to pvalues, there is more evidence in favor of the allelic homogeneity hypothesis than there is in favor of the genotypic homogeneity hypothesis. Therefore the contradiction of not rejecting the first hypothesis while rejecting the second one cannot happen for any cutoff that is chosen. In fact, as we will show in next Section, this is a property of the FBST procedure: the undesirable contradiction can never happen.
Results and Discussion
We begin this Section by summarizing the results of the analyses for data presented in [16], which were presented during the exposition of the concepts throughout the paper. Results are shown in Table 5. The notation for this table is as follows: p^{G} is the traditional pvalue for genotypic association; e^{G} is the evalue for genotypic association; {p}_{U}^{A} is the usual (wrong) pvalue for allelic association; p^{A}is the pvalue for allelic association proposed in this paper; and e^{A} is the evalue for allelic association. This table also includes the results (both pvalues and evalues) for the test of the hypothesis of HWE for case and control group.
Hence, it is reasonable to expect that pvalues, as well as any other measure of evidence, should be such that p({H}_{0}^{A})\ge p({H}_{0}^{G}). To sum up, there should be more evidence in favor of {H}_{0}^{A} than in favor of {H}_{0}^{G}. In fact, this is what motivates the tests proposed by [7]. More generally, if we have two nested hypotheses, A ⊆ B ⊆ Θ, it would be desirable to have p(B) ≥ p(A). That is, one should always believe that B is at least as plausible as A. It is worth noting that this inequality must hold if one wants to guarantee that for any significance level α the rejection of B will imply the rejection of A. In other words, p(B) should always be greater than p(A) so that one will never conclude that A holds but B does not, which is, as we showed, logically impossible.
Even though this logical coherence is desirable, the analysis of data presented by [16] (Table 5) shows that this property is not achieved neither when using the traditional pvalue for allelic frequencies, nor when using the alternative test presented here. Hence, depending on the level of significance used (for example, 10%), one can conclude that genotypic homogeneity holds, but allelic homogeneity does not. This leads one to a logical contradiction that may be embarrassing for the researcher when showing his results to scientific community. Some authors (e.g. [36–39]) have already noticed that pvalues cannot be used as a measure of evidence because they do not respect this property. Attempts to correct frequentist tests so that they are coherent have been tried in some specific situations such as Analysis of Variance [40], but no general procedure could be obtained.
On the other hand, evalues are monotone in the set of all possible hypotheses. This can be seen by noting that
For the problem considered here, this means that e{v}_{x}({H}_{0}^{G})\le e{v}_{x}({H}_{0}^{A}) will hold for all datasets. Hence, one will always have at least as much evidence in favor of {H}_{0}^{A} as in favor of {H}_{0}^{G}, and therefore when performing the FBST procedure (that is, comparing the evalues with a given cutoff) one will never fall into the logical contradiction of rejecting {H}_{0}^{A} while not rejecting {H}_{0}^{G}. Equation 4 proves that the incoherence can never happen when using the FBST. Table 5 shows that this inequality indeed holds for the data presented. It is also interesting to note that in the case of nested hypotheses, FBST provides an intrinsic penalty that can be used for model selection [41].
In Table 6, one can find similar results on simulated data. Data was simulated in three different conditions: 1  under genotypic homogeneity (and, therefore, allelic homogeneity), 2  under only allelic homogeneity and 3  under neither allelic nor genotypic homogeneity. Bold pvalues indicate situations in which there is incoherence in the sense described here. Note that, as it was expected due to the proof that was given, none of the samples have incoherence when using analyses provided by evalues.
As mentioned before {H}_{0}^{G} implies {H}_{0}^{A} in the sense that if genotypic frequencies are the same in both groups then allelic frequencies must also be the same. In other words, it is impossible for the allelic frequencies to be different if the genotypic frequencies are equal. This can be formally seen by noting that
An important question is why we use FBST methodology rather then standard Bayes factors, the traditional Bayesian procedure to test sharp hypotheses [25]. The reason is that, contrary to evalues, Bayes factors are also not monotonic when dealing with sharp hypotheses as we will show here. In order to calculate Bayes factors, one must first assign a probability distribution for the parameters under each of the hypothesis of interest. In the problem we deal with, this means it is necessary to assign probabilities for θ under Θ, {H}_{0}^{G} and {H}_{0}^{A}. The Bayes factor for hypothesis H is then defined to be \frac{P(\mathit{\text{data}}\mathrm{H})}{P(\mathit{\text{data}}\mathrm{\Theta})}[38]. For the real dataset presented in [16] (Table 1), when using uniform probabilities for θ in Θ, {H}_{0}^{G} and {H}_{0}^{A} we have a Bayes factor of 6.63 in favor of {H}_{0}^{G}, while of 0.28 in favor of {H}_{0}^{A}, so that lack of monotonicity remains. The main reason for this is that it is not necessarily true that P(\mathit{\text{data}}{H}_{0}^{G})\le P(\mathit{\text{data}}\left{H}_{0}^{A}\right). See [38] for a different example where this happens. An informal explanation of the lack of monotonicity is given by [38]: “What the Bayes factor actually measures is the change in odds in favor of the hypothesis when going from the prior to the posterior”. Note that even though they are not monotonic, Bayes Factors provide a great tool for model selection [42], a point which we further discuss in the conclusions. One may also argue about the merits of using FBST as a genuine Bayesian procedure rather than traditional Bayes factors. We advocate that while Bayes factors are primarily motivated by the epistemological framework of Decision Theory and pvalues are supported by Popperian falsificationism, evalues and FBST are supported by the framework of Cognitive Constructivism. The reader is referred to [43–45] for more epistemological considerations and comparisons of these methods. It is also interesting that FBST can also be justified as a minimization procedure of a loss function, as shown by [33]. This makes evalue also compatible with standard Decision Theory and therefore traditional Bayesian statistics. We emphasize that whenever hypotheses are not sharp, posterior probabilities are usually more adequate.
We end up this Section by answering the question of whether FBST procedure has good power properties. Even though this is not of primary interested in this work and is not a relevant question for most orthodox Bayesians, we indicate that this Bayesian procedure has good frequency properties. In order to do this, we fix different values for γ_{ AA },γ_{ AB } and Π_{ AA }. We then set Π_{ AB } to be 2(γ_{ AA } + 1/2γ_{ AB }−Π_{ AA }−ε) for different values of ε.ε quantifies how far from allelic homogeneity the population is. The particular case ε = 0 corresponds to allelic homogeneity, that is, to a true hypothesis. For each value of ε, we simulate 100 data sets with 100 samples of cases and 100 samples of controls group. We then calculate the proportion of samples in which allelic homogeneity was rejected according to each criteria. We use levels of significance of 5% and 10%. The relationship provided by [30] is used to determine the cutoffs for evalues that make FBST have the desired level of significance. Results are shown in Figure 3. These graphs indicate that the usual test for allelic homogeneity has a larger power than the others. This conclusion is misleading, once the size of the test is not the nominal one, as we discuss in Section Usual Procedures. This can be seen by looking at the curve at ε = 0 and noting that the power (which for ε = 0 is the size of the test) is larger than 5% and 10% respectively. For more simulations regarding this test power, the reader is referred to [17]. This figure also shows that the power of the frequentist allelic test proposed here and the FBST test are virtually the same: even though FBST structure guarantees coherence in the results and frequentist tests do not have the property, their power are very close to each other. Hence, the FBST procedure also has good frequency properties.
Conclusions
Although the traditional approach of doubling the sample size to test allelic homogeneity hypothesis was already shown to be incorrect when HardyWeinberg equilibrium is not met, many recent articles in biology still use it. As Figure 3 illustrates by using power analysis functions, the nominal level of significance for the allelic usual test is not attained: at zero in the xaxis, the power is larger than 5%, contrary to the alternative ones. We have shown in this paper that a logical inconsistency that happens when using such procedure remains even when using adjusted frequentist tests. The main point of this inconsistency is the fact that if two vectors are equal any function of them must maintain the equality. The fact that even when using an exact approach incoherence remains hints that the problem is the change of dimension when going for global homogeneity to partial homogeneity: genotypic homogeneity is in dimension 2 (two degrees of freedom) and allelic homogeneity is in dimension 1 (one degree of freedom). As Wald Tests, Likelihood Ratio Tests, and ChiSquare tests are asymptotically equivalent, it is also expected that contradictions may happen to all of them.
Similar incoherences of pvalues in other situations have already been reported in the literature. As a simple ANOVAlike example, suppose we wish to compare the means of independent random variables from 3 different groups, μ_{1},μ_{2} and μ_{3}. If we assume their distribution is normal with variance 1 and the sample means in each group (sufficient statistics) are −0.192,0.015 and 0.017, the likelihood ratio pvalue for the hypothesis μ_{1} = μ_{2} is 0.037. On the other hand, when testing μ_{1} = μ_{2} = μ_{3} we get a pvalue of 0.054. Hence, at the level of 5%, the first hypothesis is rejected, but the second one is not. This makes it debatable whether it reasonable to use them as measures of evidence [37]. On the other hand, if we use the improper prior f(μ_{1},μ_{2},μ_{3})∝1, the evalues are 0.232 and 0.121, respectively. Hence the contradiction cannot happen for any cutoff.
As probabilities are monotonic, traditional Bayesian tests based on posterior probability calculations do enjoy monotonicity property, however using them here may be problematic because the hypotheses of interest are sharp. Mixed continuousdiscrete distributions are needed in this case. Bayes Factors, on the other hand, were shown to be not monotonic. This does not invalidate its use: in fact, as pointed out by [38] and [42], Bayes Factors provide a great tool for model selection. One of the reasons for this is that parsimonious models can have better predictive power than complex models [46].
The FBST computation always is performed in the full space that has dimension 4. Hence subhypotheses should coherently follow the orientation of the main hypothesis. Moreover, there is no need of specifying special priors for each of the null hypotheses, only for the whole parametric space Θ. It can also be easily implemented. The problem with the FBST is that the values of the significance index, “e”, are related to the dimension and increase as the dimension increases. However, in [47] it is shown how “e” relates with “p”. This allows one to look for the corresponding evalue for 5% of significance for instance. Another point in favor of the FBST is that its power is almost the same as the best frequentist test. Moreover, it is correct even when HWE does not hold. It is important to remember that evalues are probabilities of subsets of the parameter spaces although pvalues are probabilities of sets (tails) of the sample spaces. On the other hand one must understand that hypotheses are statements about points of the parameter space and not of the sample space: May this explain the reason why the evalues, contrary to pvalues, are coherent in all situations?
Using the R Software, a routine that performs all the tests considered in this paper can be downloaded on http://www.ime.usp.br/∼cpereira/programs/nested.r
References
Last JM: A Dictionary of Epidemiology. 2001, New York: Oxford University Press Inc.
Sullivan PF, Eaves LJ, Kendler KS, Neale MC: Genetic casecontrol association studies in neuropsychiatry. Arch Gen Psychiatry. 2001, 58: 101524. 10.1001/archpsyc.58.11.1015.
Gorroochurn P, Hodge SE, Heiman G, Greenberg DA: Effect of population stratification on casecontrol association studies. II. Falsepositive rates and their limiting behavior as number of subpopulations increase. Human Heredity. 2004, 58: 408. 10.1159/000081455.
Schlesselman JJ, Stolley PD: CaseControl Studies: Design, Conduct, Analysis. Ann Internal Med. 1983, 98: 122122.
Armitage P: Tests for Linear Trends in Proportions and Frequencies. Biometrics. 1955, 11 (3): 375386. 10.2307/3001775.
Cochran WG: Some Methods for Strengthening the Common Chisquare Tests. Biometrics. 1954, 10 (4): 417451. 10.2307/3001616.
Wang K: Statistical tests of genetic association for casecontrol study designs. Biostatistics. 2012, 13 (4): 724733. 10.1093/biostatistics/kxs002.
Guedj M, Wojcik J, DellaChiesa E, Nuel G, Forner K: A fast, unbiased and exact allelic test for casecontrol association studies. Human Heredity. 2006, 61: 210221. 10.1159/000094776.
Sasieni PD: From genotype to genes: doubling the sample size. Biometrics. 1997, 53: 12531261. 10.2307/2533494.
Freidlin B, Zheng G, Li Z, Gastwirth JL: Trend tests for casecontrol studies of genetic markers: power, sample size and robustness. Human Heredity. 2002, 53: 14652. 10.1159/000064976.
Tchetgen EJ, Kraft P: On the robustness of tests of genetic associations incorporating geneenvironment interaction when the environmental exposure is misspecified. Epidemiology. 2011, 22: 257261. 10.1097/EDE.0b013e31820877c5.
Zou GY, Donner A: The Merits of Testing HardyWeinberg Equilibrium in the Analysis of Unmatched CaseControl Data: A Cautionary Note. Ann Human Genet. 2006, 70: 923933. 10.1111/j.14691809.2006.00267.x.
Singer JM, Peres CA, Harle CE: A note on the HardyWeinberg model in generalized ABO systems. Stat & Probability Lett. 1991, 11: 173175. 10.1016/01677152(91)90138H.
Brentani H, Nakano EY, Martins CB, Izbicki R, Pereira CAB: Disequilibrium Coefficient: A Bayesian Perspective. Stat Appl Genet Mol Biol. 2011, 10: 317335.
Shoemaker J, Painter I, Weir BS: A Bayesian characterization of HardyWeinberg disequilibrium. Genetics. 1998, 149: 207988.
Lin SK, Chen CK, Ball D, Liu HC, Loh EW: Genderspecific contribution of the GABAA subunit genes on 5q33 in methamphetamine use disorder. Pharmacogenomics J. 2003, 3: 349355. 10.1038/sj.tpj.6500203.
Schaid DJ, Jacobsen SJ: Biased tests of association: comparison of allele frequencies when departing from HardyWeinberg proportions. Am J Epidemiol. 1999, 149: 706711. 10.1093/oxfordjournals.aje.a009878.
Zheng G: Can the Allelic Test be Retired from Analysis of CaseControl Association Studies?. Ann Human Genet. 2008, 72: 848851. 10.1111/j.14691809.2008.00466.x.
Guedj M, Nuel G, Prum B: A Note on Allelic Tests in CaseControl Association Studies. Ann Human Genet. 2008, 72: 407409. 10.1111/j.14691809.2008.00438.x.
Knapp M: Re: “Biased tests of association: comparisons of allele frequencies when departing from HardyWeinberg proportions”. Am J Epidemiol. 2001, 154 (3): 2878. 10.1093/aje/154.3.287.
Bonnans JF, Gilbert J, Lemarechal C, Sagastizabal C: Numerical optimization, theoretical and numerical aspects. 2006, Berlin: Springer Verlag
Stephens M, Balding DJ: Bayesian statistical methods for genetic association studies. Nat Rev Genet. 2009, 10: 681690.
Berger JO: Statistical Decision Theory and Bayesian Analysis. 1985, New York: Springer Science+Business Media
Lindley DV: A statistical paradox. Biometrika. 1957, 44: 187192.
Berger JO, Delampady M: Testing Precise Hypotheses. Stat Sci. 1987, 2: 317335. 10.1214/ss/1177013238.
Pereira CAB, Stern JM: Evidence and Credibility: A full Bayesian test of precise hypothesis. Entropy. 1999, 1: 99110. 10.3390/e1040099.
Rogatko A, Slifker MJ, Babb JS: HardyWeinberg equilibrium diagnostics. Theor Population Biol. 2002, 7: 62251.
Pereira CAB, Nakano F, Stern J, Whittle M: Genuine Bayesian multiallelic significance test for the HardyWeinberg equilibrium law. Genet Mol Res. 2006, 5: 619631.
Lauretto MS, Nakano F, Faria SRJ, Pereira CAB, Stern JM: A straightforward multiallelic significance test for the HardyWeinberg equilibrium law. Genet Mol Biol. 2009, 32: 619625. 10.1590/S141547572009000300028.
Pereira CAB, Stern JM, Wechsler S: Can a Significance Test be Genuinely Bayesian. Bayesian Anal. 2008, 3: 79100. 10.1214/08BA303.
Stern JM, Zacks S: Testing the independence of Poisson variates under the Holgate bivariate distribution: the power of a new evidence test. Stat & Probability Lett. 2002, 60: 313320. 10.1016/S01677152(02)003140.
Stern JM: Paraconsistent Sensitivity Analysis for Bayesian Signicance Tests. Advances in Artificial Intelligence  SBIA. Edited by: Bazzan A, Labidi S. 2004, New York: Springer, 1340143.
Madruga MR, Esteves L, Wechsler S: On the bayesianity of PereiraStern tests. Test. 2001, 10: 291299. 10.1007/BF02595698.
Stern JM: Cognitive Constructivism, EigenSolutions, and Sharp Statistical Hypotheses. Cybernetics and Human Knowing. 2007, 14: 936.
Gelman A, Carlin J, Stern H, Rubin D: Bayesian Data Analysis. Florida: Chapman and Hall/CRC Press. 2003
Gabriel K: Simultaneous test procedures  some theory of multiple comparisons. Ann Math Stat. 1969, 41: 224250.
Schervish M: P values: what they are and what they are not. Am Statistician. 1996, 50: 203206.
Lavine M, Schervish M: Bayes factors: what they are and what they are not. Am Statistician. 1999, 53: 119122.
Izbicki R, Esteves LG: Coherence in Hypotheses Testing. JSM Proc, Am Stat Assoc. 2011, 12201234.
Rom DM, Holland B: A new closed multiple testing procedure for hierarchical families of hypotheses. J Stat Planning and Inference. 1995, 46: 265275. 10.1016/03783758(94)00116D.
Pereira CAB, Stern JM: Model selection: Full Bayesian approach. Environmetrics. 2001, 12 (6): 559568. 10.1002/env.482.
Kass RE, Raftery AE: Bayes Factors. J Am Stat Assoc. 1995,, 90 (430): 773795. 10.1080/01621459.1995.10476572.
Borges W, Stern JM: The Rules of Logic Composition for the Bayesian Epistemic eValues. J IGPL. 2007, 15: 401420. 10.1093/jigpal/jzm032.
Stern JM: Constructive Verification, Empirical Induction, and Falibilist Deduction: A Threefold Contrast. Information. 2011, 2: 635650. 10.3390/info2040635.
Stern JM: Symmetry, Invariance and Ontology in Physics and Statistics. Symmetry. 2011, 3 (3): 611635.
Hastie T, Tibshirani R, Friedman JH: The elements of statistical learning: data mining, inference, and prediction: with 200 fullcolor illustrations. 2001, New York: SpringerVerlag
Diniz M, Pereira CAB, Polpo A, Stern JM, Wechsler S: Relationship Between Bayesian and Frequentist Significance Indices. Int J Uncertainty Quantification. 2012, 2 (2): 161172. 10.1615/Int.J.UncertaintyQuantification.v2.i2.60.
Acknowledgements
The authors are grateful for Luís Gustavo Esteves, Julio Stern, Marcelo Lauretto, Rafael Bassi Stern and Sergio Wechsler for having discussed all the methods of FBST used in this paper. We also thank them for all the patience and painstakingly reading. We thank the anonymous referees for their comments that much improved the quality of the paper. This work was supported by Coordenação de Aperfeiçoamento de Pessoal de Nível Superior; Conselho Nacional de Desenvolvimento Científico e Tecnológico; and Fundação de Amparo à Pesquisa do Estado de São Paulo.
Author information
Affiliations
Corresponding authors
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
RI wrote the original manuscript. AGH was responsible for finding real data for the problem, as well as discussing the methods from a biological point view. AGH and RI did the literature survey. EYN, RI and VF worked on the mathematics of the methods as well as implemented them. CAdeBP first noticed the lack of monotonicity of the previous approaches and introduced the FBST as a way of solving the problem. All authors revised, read and approved the final manuscript.
Authors’ original submitted files for images
Below are the links to the authors’ original submitted files for images.
Rights and permissions
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License(http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Izbicki, R., Fossaluza, V., Hounie, A.G. et al. Testing allele homogeneity: the problem of nested hypotheses. BMC Genet 13, 103 (2012). https://doi.org/10.1186/1471215613103
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/1471215613103
Keywords
 Allelic homogeneity test
 Bayesian methods
 Chisquared test
 HardyWeinberg equilibrium
 FBST
 Monotonicity