Parametric estimation of P(X > Y) for normal distributions in the context of probabilistic environmental risk assessment
 Published
 Accepted
 Received
 Academic Editor
 Budiman Minasny
 Subject Areas
 Environmental Impacts, Environmental Sciences, Statistics, Toxicology
 Keywords
 Bayesian estimator, Maximum likelihood estimator, Risk assessment, Simulation, Nanoparticle
 Copyright
 © 2015 Jacobs et al.
 Licence
 This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ) and either DOI or URL of the article must be cited.
 Cite this article
 2015) Parametric estimation of P(X > Y) for normal distributions in the context of probabilistic environmental risk assessment. PeerJ 3:e1164 https://doi.org/10.7717/peerj.1164 (
Abstract
Estimating the risk, P(X > Y), in probabilistic environmental risk assessment of nanoparticles is a problem when confronted by potentially small risks and small sample sizes of the exposure concentration X and/or the effect concentration Y. This is illustrated in the motivating case study of aquatic risk assessment of nanoAg. A nonparametric estimator based on data alone is not sufficient as it is limited by sample size. In this paper, we investigate the maximum gain possible when making strong parametric assumptions as opposed to making no parametric assumptions at all. We compare maximum likelihood and Bayesian estimators with the nonparametric estimator and study the influence of sample size and risk on the (interval) estimators via simulation. We found that the parametric estimators enable us to estimate and bound the risk for smaller sample sizes and small risks. Also, the Bayesian estimator outperforms the maximum likelihood estimators in terms of coverage and interval lengths and is, therefore, preferred in our motivating case study.
Introduction
Like all novel materials, engineered nanoparticle (ENPs) have no history of safe use. A risk assessment is important for the societal acceptance and safe use of ENPs. In order to perform a proper risk assessment, one needs knowledge and data on the properties of nanoparticles. These properties can be different in nanoparticles compared to conventional chemicals in areas such as physicochemical properties, life cycle, toxicokinetics and environmental fate. This information is hard to come by because of lack of knowledge and technical limitations, resulting in no or only small datasets for effect concentrations of ENPs. In the EU, environmental risk assessment is regulated by the European Chemicals Agency (ECHA) and probabilistic risk assessment is level 3 of their tiered risk assessment approach (ECHA, 2012).
In the motivating case study of aquatic risk assessment of nanoAg (Gottschalk, Kost & Nowack, 2013), we are confronted with such a small dataset of effect concentrations. Gottschalk, Kost & Nowack (2013) modeled the exposure of nanoAg from surface water with a probabilistic materialflow model (Gottschalk, Scholz & Nowack, 2010) to obtain a distribution of exposure concentration values. They collected the effect concentration data from available toxicity studies found in the literature. These effect concentration data consist of toxic endpoints (eq. LC_{50}, EC_{50}, NOEC) for 12 aquatic species. For some of these species there were more than one data point. We averaged these to obtain one value for each species. Histograms and normal density curves of the exposure and effect concentration data are given in Fig. 1.
In probabilistic risk assessment, the variability of environmental exposure due to natural variation in concentration values over various environments is modelled by an exposure concentration (X) distribution (ECD). Similarly, the variability in effect concentration values due to natural variation among species in their sensitivity to nanoparticles is modelled by a species sensitivity distribution (SSD) or effect concentration (Y) distribution. Probabilistic risk estimation is based on the overlapping of the ECD and the SSD (ECHA, 2012). The risk, R = P(X > Y), is the area under the curve obtained by multiplying the probability density function (pdf) of the ECD with the cumulative distribution function (cdf) of the SSD. Verdonck et al. (2003) critically discuss this approach to risk assessment.
In the ecotoxicological risk assessment literature, R = P(X > Y) as a definition for risk was first developed by Suter, Vaughan & Gardner (1983). This concept was further developed by Van Straalen (2002) as ecological risk (ER). A similar concept, known as expected total risk, was developed by the Water Environment Research Foundation (WERF) (Cardwell et al., 1993; WarrenHicks, Parkhurst & Butcher, 2002). For a visual representation of risk, the exceedance function (1cdf) of the ECD is plotted against the cdf of the SSD. Such a plot is called a joint probability curve (JPC) (ECOFRAM, 1999; Solomon, Giesy & Jones, 2000; Solomon & Takacs, 2002). The area under the JPC is the risk (Solomon & Takacs, 2002). Aldenberg, Jaworska & Traas (2002) showed that the area under the JPC, ER and expected total risk are mathematically identical. In this paper, we refer to the probability, P(X > Y), as the risk, R.
When we consider the case study data, there is no overlap between the effect and the exposure histograms (Fig. 1). There is no exposure concentration that is greater than an effect concentration, and, therefore, the empirical estimate of R = P(X > Y) is zero for these datasets. To conclude, however, that the true R is zero based on a small sample is rather imprudent. The denial of a probability of zero is referred to as Cromwell’s rule by Lindley (1971, p. 105–106; 2006, p. 90–91). Several possibilities exist to address the zero problem empirically. This is further discussed in Section ‘Nonparametric estimator’. The zero problem can also be solved by fitting a parametric distribution to the data. When considering the normal density curves in Fig. 1, we note that there is some overlap between the exposure and effect concentration distributions and, therefore, some nonzero probability of exposure values exceeding effect values.
It is common to assume independent lognormal distributions for the exposure distribution and the species sensitivity distribution (SSD) (Aldenberg, Jaworska & Traas, 2002; Verdonck et al., 2003; Wagner & Løkke, 1991). This is the same as assuming normal distributions on the logtransformed exposure and effect concentrations. This normal–normal model was developed in some detail by Aldenberg, Jaworska & Traas (2002) and allows an analytic expression for the risk when parameter values are known.
Estimation of R = P(X > Y) is also of interest in other areas such as engineering and medical applications. In engineering, R = P(X > Y) is known as the reliability in stress–strength models. This is a wellknown concept and has been studied extensively for the normal–normal model (Barbiero, 2011; Church & Harris, 1970; Downtown, 1973; Enis & Geisser, 1971; Govidarajulu, 1967; Nandi & Aich, 1996; Voinov, 1986; Weerahandi & Johnson, 1992) as well as for other distributions (Kundu & Gupta, 2006; Mokhlis, 2005; Nadar, Kızılaslan & Papadopoulos, 2014). None of these papers give sufficiently general theory for obtaining trustworthy interval estimates in the case study. In receiver operating characteristic (ROC) analysis such as used in medical applications, P(X > Y) is known as the area under the ROC curve (Li & Ma, 2011). Although usually used for categorical data, the area under the ROC curve can also be obtained for continuous data in both a nonparametric way and for the normal–normal model (Krzanowski & Hand, 2009).
In this paper, we will investigate the influence of sample size on the estimation of R = P(X > Y), with special attention to the sample size of effect concentrations. We also investigate the behaviour of the estimators of R = P(X > Y) for small risks. We consider one nonparametric estimator and three parametric estimators, namely, the maximum likelihood estimator (MLE), quasi maximum likelihood estimator (QMLE) and Bayesian estimator with noninformative prior for the normal–normal model. In comparing the parametric estimators with the nonparametric one, we investigate the maximum gain possible when making strong parametric assumptions as opposed to making no parametric assumptions at all. This is done in a simulation study in which we also assess the accuracy and precision of the estimators and compare them for various combinations of sample sizes and risks.
In ‘Theory and Methods’, we derive the estimators and provide the simulation structure. In ‘Simulation Results’, the simulation results will be given and discussed. In ‘Case Study’, the application is discussed in the context of the simulation results. ‘Discussion and Conclusion’ provides some general discussion, conclusions and recommendations for further study.
Theory and Methods
In this section, we describe the theory and methodology of our approach. We start by deriving the risk for the normal–normal model. Next we discuss the four estimation methods, provide the simulation structure and discuss the performance measures that we used.
Risk
Let X be the log_{10} exposure concentration random variable and Y be the log_{10} effect (or log_{10} sensitivity) concentration random variable.
In the normal–normal model, the distributions are given by $X\sim N\left({\mu}_{x},{\sigma}_{x}\right)Y\sim N\left({\mu}_{y},{\sigma}_{y}\right).$
Due to the additive property of the normal distribution we have $XY\sim N\left({\mu}_{x}{\mu}_{y},\sqrt{{\sigma}_{x}^{2}+{\sigma}_{y}^{2}}\right).$
The risk for the normal–normal model is given by (1)$R=P\left(X>Y\right)=P\left(XY>0\right)=1\Phi \left(\frac{\left({\mu}_{x}{\mu}_{y}\right)}{\sqrt{{\sigma}_{x}^{2}+{\sigma}_{y}^{2}}}\right)=\Phi \left(\frac{{\mu}_{x}{\mu}_{y}}{\sqrt{{\sigma}_{x}^{2}+{\sigma}_{y}^{2}}}\right)$ where Φ(⋅) denotes the standard normal distribution. Equation (1) is a wellknown result (Reiser & Guttman, 1986).
We note locationscale invariance in Eq. (1). The value of R is determined only by the difference of the expected values and the sum of the variances. The absolute value of the individual parameters is not relevant.
Point estimation
In the following sections, we derive the MLE, QMLE, Bayesian estimator and nonparametric estimator for the risk, R. We let (x_{1}, x_{2}, …, x_{nx}) be a random sample of size n_{x} of log exposure concentrations and (y_{1}, y_{2}, …, y_{ny}) an independent random sample of size n_{y} of log effect concentrations.
Maximum likelihood estimator
The most straightforward way of estimating R is by means of maximum likelihood estimation. The estimator obtained in this way is denoted as ${\stackrel{\u02c6}{R}}_{\mathrm{MLE}}$. From the invariance property of MLEs (Bain & Engelhardt, 1992, p. 296), we obtain ${\stackrel{\u02c6}{R}}_{\mathrm{MLE}}$ by substituting the MLEs of μ_{x}, μ_{y}, ${\sigma}_{x}^{2}$ and ${\sigma}_{y}^{2}$ in Eq. (1). These MLEs are given by
Parameter  Maximum likelihood estimator 

μ_{x}  $\stackrel{\u0304}{x}=\frac{1}{{n}_{x}}\sum _{i=1}^{{n}_{x}}{x}_{i}$ 
μ_{y}  $\stackrel{\u0304}{y}=\frac{1}{{n}_{y}}\sum _{i=1}^{{n}_{y}}{y}_{i}$ 
${\sigma}_{x}^{2}$  ${\stackrel{\u02c6}{\sigma}}_{x}^{2}=\frac{1}{{n}_{x}}\sum _{i=1}^{{n}_{x}}{\left({x}_{i}\stackrel{\u0304}{x}\right)}^{2}$ 
${\sigma}_{y}^{2}$  ${\stackrel{\u02c6}{\sigma}}_{y}^{2}=\frac{1}{{n}_{y}}\sum _{i=1}^{{n}_{y}}{\left({y}_{i}\stackrel{\u0304}{y}\right)}^{2}$ 
Equation (1) then becomes (2)${\stackrel{\u02c6}{R}}_{\mathrm{MLE}}=\Phi \left(\frac{\stackrel{\u0304}{x}\stackrel{\u0304}{y}}{\sqrt{{\stackrel{\u02c6}{\sigma}}_{x}^{2}+{\stackrel{\u02c6}{\sigma}}_{y}^{2}}}\right).$ Note that ${\stackrel{\u02c6}{\sigma}}_{x}^{2}$ and ${\stackrel{\u02c6}{\sigma}}_{y}^{2}$ are the MLEs of the variance, which are not unbiased.
Quasi maximum likelihood estimator
The QMLE is similar to the MLE, differing only in the use of unbiased estimators for ${\sigma}_{x}^{2}$ and ${\sigma}_{y}^{2}$ instead of the MLEs. We then obtain (3)${\stackrel{\u02c6}{R}}_{\mathrm{QMLE}}=\Phi \left(\frac{\stackrel{\u0304}{x}\stackrel{\u0304}{y}}{\sqrt{{s}_{x}^{2}+{s}_{y}^{2}}}\right)$ with ${s}_{x}^{2}=\frac{1}{{n}_{x}1}\sum _{i=1}^{{n}_{x}}{\left({x}_{i}\stackrel{\u0304}{x}\right)}^{2}$ and ${s}_{y}^{2}=\frac{1}{{n}_{y}1}\sum _{i=1}^{{n}_{y}}{\left({y}_{i}\stackrel{\u0304}{y}\right)}^{2}$.
Bayesian estimator
Our third way of estimating R is Bayesian. Whereas maximum likelihood estimation uses the data only, Bayesian estimation combines prior knowledge about the parameter(s) with the data. The prior knowledge is specified by a prior distribution and the information in the data by the likelihood. The prior distribution and the likelihood are then combined into what is called the posterior distribution of the parameter (Gelman et al., 2014). We will derive the joint posterior distribution of the parameters μ_{x}, μ_{y}, ${\sigma}_{x}^{2}$ and ${\sigma}_{y}^{2}$. This distribution together with Eq. (1) will provide us with the posterior distribution, f_{R}(r), of R.
Unfortunately we have often very little prior knowledge. Therefore, we derive ${\stackrel{\u02c6}{R}}_{\mathrm{Bayes}}$ assuming a noninformative prior distribution for the parameters, namely $p\left({\mu}_{x},{\sigma}_{x}^{2}\right)\propto \frac{1}{{\sigma}_{x}^{2}}$ for the joint prior distribution of ${\mu}_{x},{\sigma}_{x}^{2}$ and $p\left({\mu}_{y},{\sigma}_{y}^{2}\right)\propto \frac{1}{{\sigma}_{y}^{2}}$ for ${\mu}_{y},{\sigma}_{y}^{2}$ (Gelman et al., 2014, p. 64).
The posterior distributions are then given by (Gelman et al., 2014, p. 65)

${\mu}_{x}\mid {\sigma}_{x}^{2}\sim N\left(\stackrel{\u0304}{x},\frac{{\sigma}_{x}}{\sqrt{{n}_{x}}}\right)$

${\mu}_{y}\mid {\sigma}_{y}^{2}\sim N\left(\stackrel{\u0304}{y},\frac{{\sigma}_{y}}{\sqrt{{n}_{y}}}\right)$

${\sigma}_{x}^{2}\sim \text{Inversegamma}\left(\frac{{n}_{x}1}{2},\frac{\left({n}_{x}1\right){s}_{x}^{2}}{2}\right)$

${\sigma}_{y}^{2}\sim \text{Inversegamma}\left(\frac{{n}_{y}1}{2},\frac{\left({n}_{y}1\right){s}_{y}^{2}}{2}\right)$.
From this, we obtain the conditional posterior distribution of $\frac{{\mu}_{x}{\mu}_{y}}{\sqrt{{\sigma}_{x}^{2}+{\sigma}_{y}^{2}}}$ (Weerahandi & Johnson, 1992) ${\mu}_{x}{\mu}_{y}\mid {\sigma}_{x}^{2},{\sigma}_{y}^{2}\sim N\left(\stackrel{\u0304}{x}\stackrel{\u0304}{y},\sqrt{\frac{{\sigma}_{x}^{2}}{{n}_{x}}+\frac{{\sigma}_{y}^{2}}{{n}_{y}}}\right)$ $\therefore \frac{{\mu}_{x}{\mu}_{y}}{\sqrt{{\sigma}_{x}^{2}+{\sigma}_{y}^{2}}}\mid {\sigma}_{x}^{2},{\sigma}_{y}^{2}\sim N\left(\frac{\stackrel{\u0304}{x}\stackrel{\u0304}{y}}{\sqrt{{\sigma}_{x}^{2}+{\sigma}_{y}^{2}}},\sqrt{\frac{\frac{{\sigma}_{x}^{2}}{{n}_{x}}+\frac{{\sigma}_{y}^{2}}{{n}_{y}}}{{\sigma}_{x}^{2}+{\sigma}_{y}^{2}}}\right).$ Using the variable transformation method and integrating ${\sigma}_{x}^{2}$ and ${\sigma}_{y}^{2}$ out of the joint pdf, $f\left(r,{\sigma}_{x}^{2},{\sigma}_{y}^{2}\right)$, we obtain the marginal pdf, f_{R}(r) (see Appendix Result A.2 for details). This marginal posterior pdf of R (Eq. (A.4)) can be evaluated using numerical integration.
Alternatively, we can use Monte Carlo sampling to approximate the marginal posterior pdf of R. We used the Method of Composition (Lesaffre & Lawson, 2012, p. 93–94) in which we sample from the known posterior distributions of ${\sigma}_{x}^{2}$ and ${\sigma}_{y}^{2}$, then μ_{x} and μ_{y} from their known posterior conditional distributions and then apply Eq. (1) to obtain the corresponding R value. Figure 2 shows histograms of samples drawn from the marginal posterior distribution of R (sample size of 10,000) together with the marginal pdf computed by numerical integration using Eq. (A.4) for different sample sizes and R values. It can be seen that the Monte Carlo method gives a very good approximation to the theoretical posterior pdf evaluated by numerical integration.
In this paper, we obtained the posterior distribution of R by sampling, because it required less computing time than numerical integration in our implementation. The posterior mean is often taken as the Bayesian point estimator but we also investigated the posterior median and mode as point estimators of R.
Nonparametric estimator
As a benchmark comparison for the parametric estimators, we included a basic nonparametric estimator. This estimator, ${\stackrel{\u02c6}{R}}_{\mathrm{NP}}$, is calculated from the data without any distributional assumptions by (4)${\stackrel{\u02c6}{R}}_{{\mathrm{NP}}_{0}}=\frac{1}{{n}_{x}{n}_{y}}\sum _{i=1}^{{n}_{x}}\sum _{j=1}^{{n}_{y}}\left[I\left({x}_{i}>{y}_{j}\right)+\frac{1}{2}I\left({x}_{i}={y}_{j}\right)\right]$ where $I\left(\mathcal{S}\right)=1$ if $\mathcal{S}$ is true and 0 otherwise (Krzanowski & Hand, 2009, p. 65). Alternatively, Eq. (4) can be written as $\stackrel{\u02c6}{R}}_{{\mathrm{NP}}_{0}}=\frac{U}{{n}_{x}{n}_{y}$ where U is the Mann–Whitney statistic (Gibbons & Chakraborti, 2011). Equation (4) is also known as the area under the ROC curve and its equivalence to the Mann–Whitney statistic has been shown (Bamber, 1975).
The nonparametric estimator of R is related to estimating the success probability, p, in a binomial experiment. As noted in the Introduction, we encounter the zero problem. One possible solution is making use of Laplace’s Law of Succession (Zabell, 1989). This law states that given k successes in n trials of a binomial experiment, the probability of a success on the next trial is $\frac{k+1}{n+2}$. The validity of this expression has a Bayesian basis. On assuming a uniform prior for p, the posterior distribution of p is a Beta(k + 1, n − k + 1) distribution (Gelman et al., 2014, p. 30), so that the posterior mean is $\frac{k+1}{n+2}$. This expression, denoted ${\stackrel{\u02c6}{R}}_{{\mathrm{NP}}_{\mathrm{LLS}}}$, is then used instead of the estimator in Eq. (4). Note that the posterior mean is equal to the predictive probability of a success on the next trial.
An alternative solution is to replace the zero with some nonzero value. One option is to estimate the probability of an outcome outside the range of the data as $\frac{1}{2{n}_{x}{n}_{y}}$. This method is used by Matlab and Genstat to compute quantiles. Another alternative is to use $\frac{1}{{n}_{x}{n}_{y}+1}$ which is used by Minitab and SPSS.
Interval estimation
We propose interval estimators by calculating credible intervals for Bayesian methods and confidence intervals for others. For the Bayesian estimator, we calculated 90% twosided highest posterior density (HPD) credible intervals (Box & Tiao, 1973, p. 123). These intervals are obtained by finding the interval of the posterior distribution with the highest density, for which we used the ‘HPDinterval’ function in the ‘coda’ package in R (Plummer et al., 2006). HPD intervals produce the shortest intervals on a chosen scale, e.g., R or a transformation thereof, but are not transformation invariant. To estimate an upper credible bound of the risk, we also calculated the 95% percentile of the posterior. The upper bound of the 90% twosided HPD intervals is not necessarily equal to the 95% percentile as the probabilities to the left and right of the twosided HPD interval can be unequal.
For the nonBayesian estimators, we calculated 90% Bias corrected and accelerated (BCa) parametric bootstrap confidence intervals using the ‘boot’ package in R (Canty & Ripley, 2013; Davison & Hinkley, 1997) with 1,000 bootstrap samples. For the nonparametric estimator, the BCa interval algorithm did not converge for small R values and also had some difficulty with the small sample sizes. For these cases we calculated percentile confidence intervals. The percentile method obtains a symmetric 100(1 − α)% confidence interval by calculating the $\left(\frac{\alpha}{2}\right)\mathrm{th}$ and $\left(1\frac{\alpha}{2}\right)\mathrm{th}$ percentiles of the bootstrap sample. In the BCa method, these percentiles are adjusted to correct for bias and skewness. For symmetric distributions, the percentile and BCa intervals are equal. Both intervals are also transformation invariant (Efron & Tibshirani, 1993, p. 175, 187). When calculating the confidence intervals for small risks, all bootstrap values may be equal, resulting in a zero interval length. For the BCa and percentile interval, the upper bound of a 90% twosided interval is equal to the upper bound of a 95% onesided interval (Carpenter & Bithell, 2000). The upper 95% confidence bound is, therefore, trivially obtained from the 90% twosided interval.
For the MLElike estimators, we also calculated confidence intervals based on the noncentral t distribution (Reiser & Guttman, 1986). In this method, the sum of the two variances (${s}_{x}^{2}$ and ${s}_{y}^{2}$) are approximated with a chisquared distribution.
The upper confidence (credible) bounds are of special interest in the context of managing risks, as they indicate (with some certainty) that the risk will not be higher than the upper bound.
Simulation study
In this section, we discuss the design of the Monte Carlo (MC) simulation study following the guidelines provided in Burton et al. (2006).
Simulation setup
The simulation study is performed in R (R Core Team, 2013). We use the builtin rnorm function to sample from a normal distribution using the MersenneTwister pseudorandom number generator (Matsumoto & Nishimura, 1998). Starting seeds for the different scenarios were drawn from a discrete uniform distribution to produce independent samples for each sample size scenario. The four estimators are calculated on the same sample, thereby avoiding differences among the estimators due to sampling.
Substance  Sample size  Risks  

Aquatic  Soil  1  2  3  4  
Ag  12  1  0.007  0.397  0  0 
CNT  9  2  0  0  0  0 
TiO_{2}  18  2  7.2e−13  0.187  0  1.2e−7 
ZnO  17  2  0  0.011  0  0 
Fullerenes  4  0  0  0  0 
To make the simulation as realistic as possible, we chose scenarios that are in line with recent studies of environmental risk assessment. When exposures are measured, it is common to have small sample sizes (Johnson et al., 2011; Westerhoff et al., 2011), whereas any number of exposure values can be obtained when they are modeled (Gottschalk, Kost & Nowack, 2013). For the exposure sample size, therefore, we chose two scenarios: the case of a small number of exposures (n_{x} = 5) and the case of a (relatively) large number of exposures (n_{x} = 100). We chose sample size of effect concentrations and risks loosely suggested by data from Gottschalk, Kost & Nowack (2013). From this data (Table 1), we chose the following scenarios:

Sample sizes for effect concentrations (n_{y}): 2, 5, 12, 20, 100

risks: a grid of values from 1e−14 to 0.5.
The sample size of 100 for effect concentrations was added to study the influence of a large sample size. A risk of 0.5 is obtained when μ_{x} = μ_{y}. We, therefore, chose increasing values of μ_{x} − μ_{y} to obtain the required range of risks. Considering the standard deviations, we note that the standard deviation of the effect concentration data in the case study is 5.6 times larger than that of the exposure concentration data. Based on this, we chose three scenarios: σ_{y} = σ_{x}, ${\sigma}_{y}=\frac{1}{5}{\sigma}_{x}$ and σ_{y} = 5σ_{x}.
The number of simulations was determined by running a pilot simulation (1,000 simulations) for the MLE. From this pilot, we obtained the median empirical standard deviation (sd = 0.0719496) of ${\stackrel{\u02c6}{R}}_{\mathrm{MLE}}$ and the median absolute bias (δ = 0.002107476) in ${\stackrel{\u02c6}{R}}_{\mathrm{MLE}}$ over all scenarios. These were used to calculate the number of simulations, B, according to Burton et al. (2006) $B={\left(\frac{{z}_{0.95}sd}{\delta}\right)}^{2}={\left(\frac{1.96\cdot 0.0719496}{0.002107476}\right)}^{2}=4477.58\approx \text{4,500}.$ For each of the 4,500 MC simulations, we calculated ${\stackrel{\u02c6}{R}}_{\mathrm{MLE}}$, ${\stackrel{\u02c6}{R}}_{\mathrm{QMLE}}$, ${\stackrel{\u02c6}{R}}_{\mathrm{Bayes}}$ and ${\stackrel{\u02c6}{R}}_{\mathrm{NP}}$. Due to the skewness of their sampling distributions, especially for small R values, we decided to use a transformation. Due to the nature of the analytical expression for R, (see Eq. (1)), a probit (inverse standard normal cdf) transformation is a natural choice. Some comparisons between the original scale and the probit transformation are further discussed for the Bayesian case in Section ‘Comparison of Bayesian point estimators’.
Simulations were run on a HP desktop computer running Microsoft Windows 7 with processor specification Intel(R) Core(TM) i74770 CPU @ 3.40GHz, 3401 MHz, 4 Core(s), 8 Logical Processor(s). Three Rsessions were running the three cases σ_{y} = σ_{x}, ${\sigma}_{y}=\frac{1}{5}{\sigma}_{x}$ and σ_{y} = 5σ_{x} simultaneously. The σ_{y} = σ_{x} case took the longest with the following time (in hh: mm:sec) for each of the four estimators:

MLE: 03:43:06.18 (bootstrap); 00:10:21.82 (noncentral t)

QMLE: 03:35:36.44 (bootstrap); 00:10:13.30 (noncentral t)

Bayes: 03:12:56 (sample size 10,000); 00:41:02.9 (sample size 1,000)

Nonparametric: 19:13:17.12
The bootstrap of the MLE, QMLE and nonparametric estimator was the cause of the longer runtime. The runtime for the Bayesian estimator is directly related to the size of the posterior sample. All further results are given for the large sample case.
Performance measures
We calculated various performance measures to evaluate the performance of the four point estimators. We calculated the performance measures on the probit scale, so as to be able to highlight differences among methods for small values of R:

mean: $\mathrm{probit}\left(\stackrel{\u0304}{\stackrel{\u02c6}{R}}\right)=\sum _{i=1}^{\text{4,500}}\frac{\mathrm{probit}\left({\stackrel{\u02c6}{R}}_{i}\right)}{\text{4,500}}$

absolute bias: $\mathrm{bias}=\mathrm{probit}\left(\stackrel{\u0304}{\stackrel{\u02c6}{R}}\right)\mathrm{probit}\left(R\right)$

empirical (or MC) standard deviation: $SD=\sqrt{\sum _{i=1}^{\text{4,500}}\frac{{\left(\mathrm{probit}\left({\stackrel{\u02c6}{R}}_{i}\right)\mathrm{probit}\left(\stackrel{\u0304}{\stackrel{\u02c6}{R}}\right)\right)}^{2}}{\text{4,500}}}$

root mean squared error: $\mathrm{RMSE}=\sqrt{{\mathrm{bias}}^{2}+S{D}^{2}}$.
The quality of the interval estimators will be assessed by calculating the coverage probability for each scenario. Note that the confidence intervals we calculated are approximate and do not claim to deliver the correct coverage. In addition, Bayesian credible intervals also do not claim a coverage frequency. For each of the 4,500 simulations, we calculated the confidence (credible) intervals and calculated the proportion of intervals that contained the true R value. We also investigated lengths of confidence (credible) interval over the 4,500 simulations for each scenario and R value. Coverages and lengths of confidence intervals for the different nonparametric estimators were similar. We, therefore, only consider the estimator based on Laplace’s Law of Succession.
Simulation Results
All results given and discussed are for the scenario that most resembles the case study (σ_{y} = 5σ_{x} and sample sizes n_{x} = 100 and n_{y} = 12) unless explicitly stated otherwise. Graphs and tables for the other scenarios are given in the Supplemental Information. All sampling distribution graphs plot the estimated $\mathrm{probit}\left(\stackrel{\u02c6}{R}\right)$ (or $\stackrel{\u02c6}{R}$) value (yaxis) against the true probit(R) (or R) value (xaxis). A diagonal 1–1 line is drawn to indicate where $\mathrm{probit}\left(\stackrel{\u02c6}{R}\right)=\mathrm{probit}\left(R\right)$ (or $\stackrel{\u02c6}{R}=R$). A logarithmic scale is used when R is plotted.
Comparison of Bayesian point estimators
For Bayesian estimation, we considered three point estimators, namely, the posterior mean, median and mode. We summarize the sampling distribution of each by way of three quantiles (0.5 or median, 0.025 and 0.975). In Fig. 3 these quantiles are plotted as a function of the true value. Figures 3A and 3C show the median and Fig. 3B and 3D show the 0.025 and 0.975 quantiles.
In Figs. 3A and 3B, the quantiles are calculated from the sampling distribution of the estimators on the original scale ($\stackrel{\u02c6}{R}$) and plotted on log_{10}scale. The lines for the posterior mean are above the 1:1 line, so indicating large positive bias. The lines for the posterior mode go to log_{10}(0) = − ∞, due to the very skew posterior distributions for smaller risks (on original scale of R, as already illustrated in Fig. 2), so indicating large negative bias. The lines for the posterior median are in between and closer to the 1:1 line.
In Fig. 3C and 3D, the quantiles are calculated similarly but on the probittransformed $\stackrel{\u02c6}{R}$ and plotted on the probit scale as well. Here we see that quantiles of the posterior mean, median and mode almost coincide, indicating that the skewness problem is solved. Simulations for very small sample size of effect concentrations (n_{y} = 2) showed that the mean has a slight advantage because of narrower intervals between the 0.025 and 0.975 quantiles than that of the median and the mode. This difference, however, is very quickly lost for higher sample sizes (n_{y} ⩾ 5) as shown in Fig. S23.
From this study of Bayesian point estimators on different scales, we see the advantage of the use of the probit scale for the Bayesian case. Moreover, for ease and its transformation invariance, we chose the posterior median as the Bayesian point estimator of R. The probit scale stretches out small values of R, making possible differences between methods more clearly visible for small R. On this basis, we decided to perform, for all estimators, all further calculations in the simulation study on the probit scale.
Comparison of the four point and interval estimators
In this section, we show the simulation results for the four estimators.
We first compare the sampling distributions of the four estimators. Figure 4 illustrates the median (A) and 0.025 and 0.975 quantiles (B) of the sampling distributions of ${\stackrel{\u02c6}{R}}_{\mathrm{MLE}}$, ${\stackrel{\u02c6}{R}}_{\mathrm{QMLE}}$, ${\stackrel{\u02c6}{R}}_{\mathrm{Bayes}}$ and ${\stackrel{\u02c6}{R}}_{\mathrm{NP}}$. For ${\stackrel{\u02c6}{R}}_{\mathrm{NP}}$, we plotted both the standard estimator, ${\stackrel{\u02c6}{R}}_{{\mathrm{NP}}_{0}}$, (Eq. (4)) which goes to minus infinity on the probit scale and the Laplace version, ${\stackrel{\u02c6}{R}}_{{\mathrm{NP}}_{\mathrm{LLS}}}$. These provide the extreme endpoints of the different solutions in solving the zero problem in the nonparametric estimator.
The median of the Bayesian estimator lies closest to the true R (Fig. 4A). This is especially apparent in scenarios with n_{y} ≤ 12 (Fig. S24). The nonparametric estimators are clearly not able to estimate R for smaller values as they very quickly jump to their lower bound of either $\mathrm{probit}\left(\frac{1}{{n}_{x}x{n}_{y}+2}\right)$ or minus infinity, indicated by horizontal and vertical dash–dot lines respectively. As the sample sizes increase, the three parametric estimators converge (Figs. S24 and S25). The nonparametric estimators remain the worst estimators for all sample sizes when estimating small R values. In our further study, we consider the Laplace version only.
Next, we study the coverage and interval lengths of the twosided 90% confidence (credible) intervals of the estimators on the probit scale. For each of the 4,500 simulations, we calculated interval lengths and then obtained the median interval length for each combination of estimator, sample sizes and risk value combinations. In order to compare the median interval lengths across different risk values in a single graph, we standardized each one by dividing by the true $\left\mathrm{probit}\left(R\right)\right$ value to obtain the relative median interval length. Figure 5 plots the relative median interval length (yaxis) against the coverage probabilities (xaxis) for n_{x} = 5 and n_{x} = 100, all investigated sample sizes for effect concentrations, and all R values. Coverage probabilities of less than 0.5 were plotted at 0.5. The vertical line indicates a coverage probability of 90%. A good interval estimator gives points lying on this line with short interval length. This translates to good coverage and narrow intervals. Points corresponding to n_{y} = 12 are indicated by an open black circle.
We found that the MLE (not shown) and the QMLE had a similar pattern for both the bootstrap and noncentral t intervals, with the QMLE consistently having better coverage. Figure 5 shows that the bootstrap intervals have liberal coverage compared to the noncentral t intervals for small sample sizes of effect concentrations. As the sample sizes for effect concentrations increase (bigger dots), the estimators have better coverage. Very small sample size for effect concentrations (n_{y} = 2) gives the worst coverage. For the Bayesian estimator, the sample size has a lesser influence. For small exposure sample size (n_{x} = 5), the coverage of the Bayesian interval estimator tends to be too high. This problem largely disappears for n_{x} = 100, although there is some undercoverage for the n_{y} = 2 case. The parametric estimators have shorter interval lengths when n_{x} = 100 (right column) as compared to n_{x} = 5. For the nonparametric estimator, sample size has a slightly less systematic influence on the coverage.
Compared to the other estimators, the Bayesian interval estimator best maintains the nominal coverage without having larger median interval length and despite the fact of often having a higher than nominal coverage (Fig. 5). The nonBayesian estimators have smaller than nominal coverage for small sample size of effect concentrations with the nonparametric estimator being the worst. For better comparison of interval lengths among the estimators, the reader is referred to Fig. S26.
In risk assessment, one is often interested in an upper bound on the risk. We studied the coverages and interval lengths of the upper 95% confidence (credible) bounds of the estimators on the probit scale. The interval lengths were quantified as the difference between the upper bound and the true probit(R) value. The median of the 4,500 differences was obtained. In order to compare the median differences across different risk values in a single graph, each median difference was divided by the true $\left\mathrm{probit}\left(R\right)\right$ value being estimated to obtain the relative median difference. Figure 6 plots the relative median difference (yaxis) against the coverage probabilities (xaxis) for n_{x} = 5 and n_{x} = 100, all investigated sample sizes for effect concentrations, and all R values. Coverage probabilities of less than 0.5 were plotted at 0.5. The vertical line specifies a coverage probability of 95%. A good upper bound estimator gives points lying on this vertical line and being close to the horizontal 0 line. This translates to good coverage and an upper bound close to the true R. Points corresponding to n_{y} = 12 are indicated by an open black circle.
We see similar patterns as in the case of the twosided intervals, with the bootstrap intervals being too liberal. The Bayes estimator gives higher than nominal coverage for small exposure sample size (left column). For large exposure sample size, the Bayesian estimator clearly outperforms the other estimators with good coverage for all R values and sample sizes for effect concentrations without having larger median interval difference. The nonparametric estimator has a severe coverage problem.
The results for the performance measures of the different estimators are given in Tables S2–S5. Due to the lower bound of the nonparametric estimator, the SD, bias and RMSE are not reliable for small R values. Only for a few cases where R = 0.5, the nonparametric estimator has slightly lower SD and bias than the parametric estimators. The various graphs have also shown the inability of the nonparametric estimator to estimate small R values.
Among the parametric estimators, the Bayesian estimator as the smallest SD, bias and RMSE on probit scale for all sample sizes and R values. This confirms that the Bayesian estimator is better than the nonBayesian estimators as also seen for the interval estimator case. The QMLE has smaller SD, bias and RMSE than the MLE for all sample sizes and R values. This also confirms the results of the interval estimators where QMLE has better coverage than MLE.
The Bayesian estimator was in general the best estimator and specifically so for the scenario that is closest to the case study. The Bayesian point estimator was less biased than the MLE and the QMLE in all cases (σ_{y} = σ_{x} (Figs. S8 and S9), ${\sigma}_{y}=\frac{1}{5}{\sigma}_{x}$ (Figs. S16 and S17) and σ_{y} = 5σ_{x} (Figs. S24 and S25)), and this was especially apparent for small sample sizes for effect concentration (n_{y} ≤ 12) and small R values. The Bayesian interval estimator (90% twosided) had better coverage, with even higher than nominal coverage for exposure sample size n_{x} = 5. This is also seen for the case σ_{y} = σ_{x} (Fig. S10). For the case ${\sigma}_{y}=\frac{1}{5}{\sigma}_{x}$ (Fig. S18), the higher coverage is only seen for small sample size for effect concentrations as seen by the small dots. When considering the 95% upper bound of the Bayesian estimator compared to MLE and QMLE, we also see better coverage with similar higher than nominal coverage for exposure sample size n_{x} = 5. This is similar for the σ_{y} = σ_{x} case (Fig. S11) and slightly more pronounced for the ${\sigma}_{y}=\frac{1}{5}{\sigma}_{x}$ case (Fig. S19). Considering the performance measures, the Bayesian estimator performs better (lower values), also for the σ_{y} = σ_{x} case (Tables S2–S4) and the ${\sigma}_{y}=\frac{1}{5}{\sigma}_{x}$ case (Table S6–S8). For the corresponding case study scenario of σ_{y} = 5σ_{x} and n_{x} = 100, n_{y} = 12, the Bayesian estimator clearly outperformed the MLE and QMLE. It was less biased and maintained the nominal coverage in both the twosided and onesided cases. For better comparison of interval lengths among the estimators, the reader is referred to Fig. S27.
Case Study
In this section, we evaluate the case study results on the basis of the simulation study results. In the case study, we have a sample size of 1,000 of exposure concentrations and a sample size of 12 of effect concentrations. We note that the exposure concentrations come from a simulation model, so that it is possible to generate an arbitrary large sample exposure concentrations. We treat the size of 1,000 exposures as being effectively of size 100.
First, we verify that the normal–normal model is not in conflict with the data. Visually, the normal distribution fits the concentration data quite well (Fig. 1). The small sample size of the effect concentrations gives low power to any formal normality tests, where the large sample size of exposure concentrations gives high power, so that even small deviations from normality are detected. Even so, we cannot reject the null hypothesis of normality (see Table S1) for either the effect or the exposure samples at a 5% significance level. In the exposure concentration data, there is some indication for nonnormality evident from two of the normality tests which are only just not significant (pvalues of 0.0564 and 0.0538). Even so, we take the normal–normal model as a useful model.
Estimator  $\stackrel{\u02c6}{R}$  90% 2sided CI  95% upper CB 

MLE  0.0068  0.0003–0.0684  0.0684 
QMLE (noncentral t)  0.0090  0.0006–0.0784  0.0784 
QMLE (bootstrap)  0.0090  0.0002–0.0571  0.0571 
Bayes  0.0108  0.0006–0.0776  0.0806 
Empirical  0.0001  0.0001–0.0001  0.0001 
Next, we consider the estimates of the risk (Table 2). The estimates and intervals were calculated on the probitscale and then transformed back to the original scale so as to be able to evaluate the case study results in the light of the simulation study results. For the MLE we calculated the interval estimators based on the noncentral t distribution and for the QMLE, the noncentral t and parametric bootstrap .
The nonparametric estimator was calculated using Laplace’s Law of Succession. For the sample sizes of this case study, ${\stackrel{\u02c6}{R}}_{{\mathrm{NP}}_{\mathrm{LLS}}}$ then becomes $\frac{0+1}{\text{1,000}\cdot 12+2}=\frac{1}{\text{12,002}}=0.000083$. We note, however, that this value is very much dependent on the sample size. A larger exposure sample size will decrease the estimate. As seen in the simulation study results, it is impossible to draw any meaningful conclusions for small risks based on the nonparametric estimator.
The three parametric estimates are similar. The bootstrap 90% confidence interval of the QMLE is clearly narrower. From the simulation study, however, the bootstrap intervals showed liberal coverage and are, therefore, less trustworthy. Considering the 95% upper confidence bound, we note that the Bayesian bound is slightly higher than that of the MLE and QMLE and higher as well than the upper bound of the Bayesian 90% credible interval. Investigating these aspects in the simulation results, we found that these differences are to be expected (see Figs. S1–S3), although the difference between the Bayesian 95% upper bound and the upper bound of its 90% credible interval is not so typical. The distances between the 95% upper bound and $\stackrel{\u02c6}{R}$ as well as the ratio of the Bayesian upper bound to both the QMLE upper bound and the Bayesian upper bound of the twosided interval fall within the respective sampling distributions as obtained in the simulations.
Based on the simulation results we, therefore, conclude that the Bayesian estimate is the most appropriate. The upper bound (0.0806) is most reliable as it has the best coverage (compared to MLE and QMLE). This is clearly seen by the black circles in the Bayesian panel in the right column of Fig. 6. This case corresponds most closely to the case study data. Based on the model and the data used, we state with 95% confidence, that the risk will not be greater than 0.0806.
Discussion and Conclusion
In this paper we studied the problem of estimating the risk for the case of small sample size for effect concentrations and small R values. The case study data showed discrepancies between the parametric and nonparametric estimators which we investigated via a simulation study. We derived and compared three parametric estimators and one nonparametric estimator for the risk. This was done under the assumption of normality for both the exposure and effect concentration data. We found that, overall, the parametric estimators have better performance than the nonparametric estimator, and the Bayesian estimator outperformed the maximum likelihoodbased ones.
The Bayesian estimator in this paper was based on a noninformative prior on the underlying parameters. This resulted in a prior tendency of R toward 0.5. For small sample sizes, there was not enough data to counter this prior tendency and this resulted in an overestimation of R by the posterior mean estimator calculated on the original R scale (Figs. S4, S5, S12, S13, S20 and S21). To overcome this problem, it was needed to switch to the posterior median estimator or to switch to the probit(R) scale. We used both the probitscale and the posterior median resulting in an estimator that outperformed its parametric counterparts. More benefit can presumably be obtained from the Bayesian estimator if we can use an informative prior, at least when the prior is not in conflict with the data. In addition, the use of probability matching priors (Datta & Sweeting, 2005) may also improve on the coverage of the credible intervals. Ventura & Racugno (2011) used a strong matching prior for Bayesian estimation of P(X > Y) based on a profilelikelihood approach.
Using the probit(R) scale in the simulation study enabled us to more easily compare the estimators for small R values. Despite giving nice statistical properties, the probit scale may not directly address a specific risk assessment question.
Comparing the parametric bootstrap and noncentral t interval estimators for the MLE and the QMLE, we found the noncentral t intervals to have better coverage. The bootstrap intervals, although a good alternative, are liberal in coverage (i.e., resulted in smaller than nominal coverage) for small sample sizes of effect concentrations. This was also found by Tian (2008).
It was clearly seen that the nonparametric estimator was not able to estimate the risk for small sample sizes and small R values. For R values above the lower bound of $\mathrm{probit}\left(\frac{1}{{n}_{x}{n}_{y}+2}\right)$, the nonparametric estimator had performance measures similar to that of the MLE. As seen in our case study, however, the nonparametric estimator failed completely. The bootstrap cannot provide any variability of outcome with which to provide an interval for the estimate. In the simulation study, we also found that for small sample sizes, there was often too little variability in the data for the bootstrap to be able to quantify it (Fig. 4B). Although this translated to 0 coverage in Figs. 5 and 6, it really shows that the nonparametric estimator completely fails in these cases. For small sample sizes and small R values, therefore, we advise to use parametric estimators.
Considering the computation times of the simulation study, we note that, in addition to the Bayesian estimator being the best estimator, it can also require shorter computation time compared to the bootstrap alternatives depending on the posterior sample size. The larger posterior sample size (10,000) tends to result in slightly narrower estimates of the posterior distribution than those based on the smaller sample size (1,000). Nevertheless, the main results in Figs. 5 and 6 remain basically unchanged. In the case study, the credible interval becomes slightly wider for the larger sample size and the upper bound is slightly lower. Even so, not much is lost by taking the smaller sample size and this drastically reduces the computation time. The maximum likelihood based estimators have shorter computation time when calculating the interval estimators based on the noncentral t distribution than the bayesian estimator. The nonparametric estimator is by far the most computationally demanding due to the bootstrapping and the calculation of Eq. (4).
Assuming normality in the case and simulation study may seem as a strict assumption and going nonparametric is a way to avoid strict assumptions. For many situations in statistics the normal distribution is considered to have too light tails. In our case with very little data, going nonparametric leads to zero tails outside the range of the data. The usual areaunderthecurvebased nonparametric method can then severely underestimate the risk (often resulting in zero risk), whereas the estimate based on Laplace’s Law of Succession overestimates the risk for small true risks. To be able to draw any sensible conclusion, one has to use a parametric method. Our comparison of methods shows the advantage of using parametric methods in this case.
We conclude that making parametric assumptions, enabled us to estimate the risk for smaller sample sizes and small risks in the case the data is in fact normally distributed. Further research is needed to investigate the robustness of the parametric methods on nonnormal data. We need to investigate whether semiparametric methods and methods based on the extreme value distribution are able to estimate the tails of distributions sufficiently well from small data sets, so that they outperform the parametric methods used in this paper.