CoViD19: an automatic, semiparametric estimation method for the population infected in Italy
 Published
 Accepted
 Received
 Academic Editor
 Sharif Aly
 Subject Areas
 Epidemiology, Global Health, Health Policy, Infectious Diseases, Statistics
 Keywords
 Autoregressive metric, Covid19, Maximum entropy bootstrap, Model uncertainty, Number of Italian people infected
 Copyright
 © 2021 Fenga
 Licence
 This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ) and either DOI or URL of the article must be cited.
 Cite this article
 2021. CoViD19: an automatic, semiparametric estimation method for the population infected in Italy. PeerJ 9:e10819 https://doi.org/10.7717/peerj.10819
Abstract
To date, official data on the number of people infected with the SARSCoV2—responsible for the Covid19—have been released by the Italian Government just on the basis of a nonrepresentative sample of population which tested positive for the swab. However a reliable estimation of the number of infected, including asymptomatic people, turns out to be crucial in the preparation of operational schemes and to estimate the future number of people, who will require, to different extents, medical attentions. In order to overcome the current data shortcoming, this article proposes a bootstrapdriven, estimation procedure for the number of people infected with the SARSCoV2. This method is designed to be robust, automatic and suitable to generate estimations at regional level. Obtained results show that, while official data at March the 12th report 12.839 cases in Italy, people infected with the SARSCoV2 could be as high as 105.789.
Introduction
Covid19 epidemic has severely hit Italy, and its spread throughout Europe is expected soon. In such a scenario, the availability of reliable information related to its spread plays a significant role in many regards. In fact, many targeted measures, such as the coordination among emergency services or the implementation of operative actions (extensive or local lockdowns or even curfew) can only be efficiently taken when reliable estimates of the epidemic spread are available at the population level.
At the moment, official data on the infection in Italy are based on nonrandom, nonrepresentative samples of the population: people are tested for Covid19 on the condition that some symptoms related to the virus are present. These data can ensure a proper estimation of the number of both deaths and hospitalizations due to the virus and are crucial for the optimization of the available resources. Nonetheless, from a statistical point of view, the number of people tested positive for Covid19 represents a simple count which is not suitable to provide a reliable assessment of the “true”, unknown, number of infected people (thereafter “positive cases” ). In addition to the strong bias components induced by this testing strategy, there is at least another major obstacle to the construction of a valid estimator: the small sample size available. These issues are considered in the available literature: Feinstein & Esdaile (1987) point out how the statistical information in many cases can contain gross violations of epidemiological principles as well as of scientific standards for credible evidence. On the other hand, a substantial corpus of theory and methods are available to epidemiologists and/or the statisticians working on the field of epidemiology—see, for example, Kahn, Kahn & Sempos (1989) and, more recently, Clayton & Hills (2013) and Lawson (2013). Therefore, a “reasonable” trade–off between goodness of the outcomes of a statistical analysis and the available data, in some cases, is the best we can hope for. In the case of the present article, the shortness of the time series of interest is simply something that, at an early stage of an epidemic, cannot be avoided. It is well known that the shortness of the time series of interest might lead to a strong bias in the asymptotic results and therefore to the construction of biased confidence intervals. However, the results obtained in this article can be considered reliable as the approach used has been specifically designed to mitigate these negative effects. To confirm that, the estimates provided by this method have been proved to be in line with those published by official entities and have been reported on a number of nationally distributed daily newspaper published in Italy.
Based on the number of the deaths and of the observed positive cases and improving on an estimation equation proposed by Pueyo (2020), this article aims at estimating the “true” number of people infected by the Covid19 in each of the 20 Italian regions. Presently, to the best of the author’s knowledge, Puejo’s equation does not appear in the literature nevertheless its validity in the present context will be discussed later in “Data and Contagion Indicator”. In more details, the presented procedure is designed to reduce the impact of the biasing components on the parameter estimations, by employing a resampling scheme, called Maximum Entropy Bootstrap (MEBOOT) proposed by Vinod & Lópezde Lacalle (2009). This bootstrap method is particularly suitable in this context: as it will be outlined in the sequel it is designed to work with a broad class of time series (including non stationary ones) and—by virtue of its inherent simplicity—is able to generate bona fide replications in the case of short time series. In fact, unlike other schemes, long time series are not required. For example, in the case of the sieve bootstrap method Andre’es, Pena & Romo (2002), a lengthy series is needed in order to estimate an high order autoregressive model from which the bootstrap replications are generated. In conjunction with MEBOOT, a distance measure—based on the theory of stochastic processes and proposed by Piccolo (1990)—has been used to find pairs of similar regions. As it will be explained later, this has been done to maintain the same methodology in those cases where one of the variable employed in the model—that is, the number of deaths—was missing.
An Overview of the Proposed Method
In small data sets it is essential to save degrees of freedom (DOF) which are inevitably lost in an amount correlated with the complexity of the statistical model entertained (see, for example, Faes et al. (2009) and Barnard & Rubin (1999)). With this in mind, the proposed method is of the type semiparametric and consists of two parts: a purely nonparametric and a parametric one. The nonparametric part refers to the maximum entropy resampling method, which will be used to generate more robust estimations. On the other hand, a parametric approach has been chosen to select certain regions on the basis of a similarity function, as it will be explained at the end of the following “Data and Contagion Indicator”. While the former does not pose problems in terms of DOF, the latter clearly does. However, the sacrifice in terms of DOF is very limited as an autoregressive model of order 1 (employed in a suitable distance function, as below illustrated) has proved sufficient for the purpose. DOF—saving strategy is also the driving force behind the choice to not consider exogenous variables such as the regions geolocation or their population—for example, in a regressionlike scheme—but to implicitly assume these (and other) variables are embedded in the dynamic of the time series considered.
Data and Contagion Indicator
The paper makes use of official data, published by the Italian Authorities, related to the following two variables employed in the proposed method, that is, the number of

deaths from Covid19 (denoted by the symbol M_{t})

currently positive cases which have been recorded as a result of the administration of the test (denoted by the symbol C_{t}).
The data set includes 18 daily data points collected at the regional level during the period of February 24th to March 12th 2020. The total number of Italian regions considered is 20. However, one special administrative area (Trentino Alto Adige) is divided in two subregions, that is, Trento and Bolzano. Therefore, the set containing all the Italian regions—called Ω—has cardinality Ω = 21 (the cardinality function is denoted by the symbol ·). Two different subsets are built from Ω that is, Ω^{•}—containing the regions for which at least one death, out of the group of tested people, has been recorded and Ω^{○} (no recorded deaths). Those two sets are now specified:

$\left\{{\mathrm{\Omega}}^{\bullet}\right\}\equiv Piemonte,Lombardia,Veneto,Friuli,Liguria,Emilia,Toscana,Marche,Lazio,Abruzzo,ValleAosta,Bolzano,Campania,Puglia,Sicilia$

$\left\{{\mathrm{\Omega}}^{\u25cb}\right\}\equiv Trento,Umbria,Molise,Basilicata,Calabria,Sardegna$,
where Ω ≡ Ω^{•} ∪ Ω^{○}. In what follows, the two superscripts ^{•} and ^{○} will be always used respectively with reference to the regions $\{{r}_{1},{r}_{2},\dots {r}_{15}\}\in {\mathrm{\Omega}}^{\bullet}$ and $\{{s}_{1},{s}_{2},\dots {s}_{6}\}\in {\mathrm{\Omega}}^{\u25cb}$. The time span is denoted as $\{1,2,\dots ,T\}$. In the case of the regions included in the set Ω^{•}, following Pueyo (2020), the total number of positive is estimated as follows: (1) $${y}_{j,T}^{\bullet}={w}_{T}\ast {2}^{{\displaystyle \frac{\tau}{\delta}}}$$ (2) $${w}_{T}={\displaystyle \frac{{C}_{T}}{{M}_{T}}}$$Here, w_{T} (Eq. 2) is the ratio between the current positive cases (C) and the number of deaths (M) whereas, in Eq. (1), τ is the average doubling time for the Covid19 (i.e., the average span of time needed for the virus to double the cases) and δ the average time needed for an infected person to die. These two constant terms have been kept fixed as estimated according to the data so far available as reported and justified in the above mentioned Puejo’s web document. They are as follows: τ = 17.3 and δ = 6.2.
By construction, Eqs. (1) and (2) are able to properly describe the spread of the virus at the population level, as they are based on the key parameters average doubling τ and time to death (δ). To make this clear, suppose a situation where τ = δ (i.e., all the subjects, in average, die the following day after the disease has been contracted). In this case, Eq. (1) reduces to y^{•}_{j,T} = 2 * w_{T}, that is we will have the total positives equal to twice the mortality rate. As for the constants chosen, they appear to be in line with the data released by the Italian public authority.
The case of the regions belonging to Ω^{○} is more complicated. The related estimation procedure has been carried out as detailed below (the subscript t will be omitted for the sake of simplicity):

given the series s_{j} ∈ Ω^{○}, a series c^{π} ∈ Ω^{•} minimizer of a suitable distance function—denoted by the Greek letter π(·)—is found. In symbols: (3) $${c}^{\pi}=\underset{(c\in {\mathrm{\Omega}}^{\bullet})}{\mathrm{argmin}}\phantom{\rule{thickmathspace}{0ex}}\pi (s,c)$$

the estimated number of positives at the population level—already found for c^{π}, say ${I}_{{c}^{\pi}}$—becomes the weight for which the total cases recorded for s_{j}, are multiplied. Therefore, the estimate of the variable of interest for this case becomes (4) $${y}_{j,T}^{\u25cb}={\displaystyle \frac{{I}_{{c}^{\pi}}\ast {C}_{{s}_{j}}}{{C}_{{r}_{j}}}}$$
The distance function adopted π(·) (Eq. 3), called ARdistance, has been introduced by Piccolo (2007). Briefly, this metric can be applied if and only if the pair of series of interest are assumed to be realizations of two (possibly of different orders) Autoregressive Moving Average (ARMA) models (see, e.g., Makridakis & Hibon (1997)). Under this condition, each series can be expressed as an autoregressive model of infinite order, that is, AR(∞), whose (infinite) sequence of AR parameters is denoted by $\{\alpha {\}}_{j}^{\mathrm{\infty}}\equiv {\alpha}_{1},{\alpha}_{2},\dots $.
Without loss of generality, the distance between the series s and c, that is, π(s_{t}, c_{t}) (Eq. 4), under (s_{t}, c_{t}) ∼ ARMA(α, β), being α and β respectively the autoregressive and moving average parameters, is expressed as (5) $$\pi (s,c)={(\sum _{j=1}^{\mathrm{\infty}}{\alpha}_{j}(s){\alpha}_{j}(c))}^{1/2}$$
Equation 5 asymptotically converges under stationary condition of the autoregressive parameters, as proved in Piccolo (2010). In other words, considering for brevity only the autoregression in α_{j}, the roots of the polynomial $\mathrm{\Phi}(z):=1{\sum}_{j=1}^{S}{\alpha}_{j}{z}^{Sj}$ must lie outside the unit circle, that is, each root z_{j} must satisfy z_{1} > 1. For other asymptotic properties the reader is referred to Corduas & Piccolo (2008). It is well known that, with small sample sizes, the asymptotic properties of the ARMA parameters tend to deteriorate and therefore the statistical model might not perform optimally. However, in the present context their use is justified at least for two reasons: firstly the ARMA models have been here employed only for the construction of a simple distance measure used to build a similarity ranking of the Italian regions. As a simple way to pick a suitable “donor” (see the explanation below), that ARMA models tend to not perform optimally in such conditions can be considered a crucial issues. The second reason refers to the fact that, epidemics are an emergency situations and the the typical case where only a few (all the more so likely to be noisy) data points are available. Finally, in order to reach stationarity and thus correctly assess the distance functions, all the models have been estimated on properly differentiated time series.
The Resampling Method
The bootstrap scheme adopted proved to be adequate for the problem at hand. Given the pivotal role played in the proposed method, it will be briefly presented. In essence, the choice of the most appropriate resampling method is far from being an easy task, especially when the identical and independent distribution (iid) assumption (used in Efron’s initial bootstrap method) is violated. Under dependance structures embedded in the data, simple sampling with replacement has been proved—see, for example Carlstein (1986)—to yield suboptimal results. As a matter of fact, iid—based bootstrap schemes are not designed to capture, and therefore replicate, dependance structures. This is especially true under the actual conditions (small sample sizes) where the selection of the “right” resampling scheme becomes a particularly challenging task. Several ad hoc methods have been therefore proposed, many of which now freely and publicly available in the form of powerful routines working under software package such as Python^{®} or R^{®}. In more details, while in the classic bootstrap an ensemble $\mathrm{\Gamma}$ represents the population of reference the observed time series is drawn from, in MEB a large number of ensembles (subsets), say $\{{\gamma}_{1},\dots ,{\gamma}_{N}\}$ becomes the elements belonging to $\mathrm{\Gamma}$, each of them containing a large number of replicates $\{{x}_{1},\dots ,{x}_{J}\}$. Perhaps, the most important characteristic of the MEB algorithm is that its design guarantees the inference process to satisfy the ergodic theorem. Formally, recalling the symbol · to denote the cardinality function (counting function) of a given ensemble of time series $\{{x}_{t}\in {\gamma}_{i};\phantom{\rule{thickmathspace}{0ex}}i=1,\dots ,N\}$, the MEB procedure generates a set of disjoint subsets ${\mathrm{\Gamma}}_{\mathbf{N}}\equiv {\gamma}_{1}\cap {\gamma}_{1}\dots \cap {\gamma}_{N}$ s.t. $\mathbb{E}{\mathrm{\Gamma}}_{N}\approx \mu ({x}_{t})$, being μ(·) the sample mean. Furthermore, basic shape and probabilistic structure (dependency) is guaranteed to be retained $\mathrm{\forall}{x}_{t,j}^{\ast}\subset {\gamma}_{i}\subset \mathrm{\Gamma}$.
MEB resampling scheme has significant advantages over many of the available bootstrap methods: it does not require complicated tune up procedures (unavoidable, for example, in the case of resampling methods of the type Block Bootstrap) and it is effective under nonstationarity. MEB method relies on the entropy theory and the related concept of (un)informativeness of a system. In particular, the Maximum Entropy of a given density ρ(x), is chosen so that the expectation of the Shannon Information $\mathcal{H}=\mathbb{E}(\mathrm{log}\rho (x))$, is maximized, that is, $$\underset{(\rho )}{max}\phantom{\rule{thickmathspace}{0ex}}\mathcal{H}=\mathbb{E}(\mathrm{log}\rho (x))$$
Under mass and mean preserving constraints, this resampling scheme generates an ensemble of time series from a density function satisfying (4). Technically, MEB algorithm can be broken down, following Koutris, Heracleous & Spanos (2008), in 8 steps. They are:

a sorting matrix of dimension T × 2, say ${\mathcal{S}}_{1}$, accommodates in its first column the time series of interest x_{t} and an Index Set—that is, ${I}_{ind}=\{2,3,\dots ,T\}$—in the other one;

${\mathcal{S}}_{1}$ is sorted according to the numbers placed in the first column. As a result, the order statistics x_{(t)} and the vector I_{ord} of sorted I_{ind} are generated and respectively placed in the first and second column;

compute “intermediate points”, averaging over successive order statistics, that is, ${c}_{t}={\displaystyle \frac{{x}_{(t)}+{x}_{(t+1)}}{2},\phantom{\rule{thickmathspace}{0ex}}\phantom{\rule{thickmathspace}{0ex}}t=1,\dots T1}$ and define intervals I_{t} constructed on c_{t} and r_{t}, using ad hoc weights obtained by solving the following set of equations:
(i) $g(x)={\displaystyle \frac{1}{{r}_{1}}\mathrm{exp}\left({\displaystyle \frac{[x{c}_{1}]}{{r}_{1}}}\right);\phantom{\rule{1em}{0ex}}x\in {I}_{1};{r}_{1}={\displaystyle \frac{3{x}_{(1)}}{4}+{\displaystyle \frac{{x}_{(2)}}{4}}}}$
(ii) $\begin{array}{c}g(x)=\frac{1}{{c}_{k}{c}_{k1}};\text{\hspace{1em}}x\in ({c}_{k};{c}_{k+1}),\\ {r}_{k}=\frac{{x}_{(k1)}}{4}+\frac{{x}_{(k)}}{2}+\frac{{x}_{(k+1)}}{4};\text{\hspace{1em}}k=1,\dots ,T1;\end{array}$
(iii) $g(x)={\displaystyle \frac{1}{{r}_{T}}\mathrm{exp}{\displaystyle \frac{([{c}_{T1}x])}{{r}_{T}};\phantom{\rule{1em}{0ex}}x\in {I}_{T};\phantom{\rule{1em}{0ex}}{r}_{T}={\displaystyle \frac{{x}_{T1}}{4}+{\displaystyle \frac{3{x}_{T}}{4}}}}}$

from a uniform distribution in [0,1], generate T pseudorandom numbers and define the interval R_{t} = (t/T ; t + 1/T) for $t=0,1,\dots ,T1$, in which each p_{j} falls;

create a matching between R_{t} and I_{t} according to the following equations: $$\begin{array}{c}{x}_{j,t,me}={c}_{T1}\theta \mathrm{ln}(1{p}_{j})\phantom{\rule{1em}{0ex}}\mathrm{i}\mathrm{f}\phantom{\rule{1em}{0ex}}{p}_{j}\in {R}_{0}\\ {x}_{j,t,me}={c}_{1}\theta ln(1{p}_{j})\phantom{\rule{1em}{0ex}}\mathrm{i}\mathrm{f}\phantom{\rule{1em}{0ex}}{p}_{j}\in {R}_{T1}\end{array}$$
so that a set of T values {x_{j,t} }, as the jth resample is obtained. Here θ is the mean of the standard exponential distribution;

a new T × 2 sorting matrix ${\mathcal{S}}_{2}$ is defined and the T members of the set {x_{j,t}} for the jth resample obtained in Step 5 is reordered in an increasing order of magnitude and placed in column 1. The sorted I_{ord} values (Step 2) are placed in column 2 of ${\mathcal{S}}_{2}$;

matrix S_{2} is sorted according to the second column so that the order $\{1,2,\dots ,T\}$ is there restored. The jointly sorted elements of column 1 is denoted by $\{{x}_{\mathcal{S},j,t}\}$, where $\mathcal{S}$ recalls the sorting step;

Repeat Steps 1–7 a large number of times.
The Application of the Maximum Entropy Bootstrap
In what follows, the proposed procedure is presented in a stepbystep fashion.

For each time series ${y}_{t}^{\bullet}$ and ${y}_{t}^{\u25cb}$ the bootstrap procedure is applied so that B = 100 “bona fide” replications are available as a result, that is, ${\stackrel{~}{y}}_{t,b}^{\bullet};b=1,2,\dots B$ and ${\stackrel{~}{y}}_{t,b}^{\u25cb};b=1,2,\dots B$;

for both the series, the row vector related to the last observation T is extracted, that is, $\{{v}^{\u25cb}={\stackrel{~}{y}}_{T,1}^{\u25cb},{\stackrel{~}{y}}_{T,2}^{\u25cb}\dots {\stackrel{~}{y}}_{T,B}^{\u25cb}\}$ and $\{{v}^{\bullet}={\stackrel{~}{y}}_{T,1}^{\bullet},{\stackrel{~}{y}}_{T,2}^{\bullet}\dots {\stackrel{~}{y}}_{T,B}^{\bullet}\}$;

the expected values, that is, $\mathbb{E}({v}^{\bullet})$ and $\mathbb{E}({v}^{\u25cb})$, are then extracted along with the ≈ 95% confidence intervals (CI^{•} and CI^{○} ), which are computed according to the tpercentile method. In essence, through this method, suitable quantiles of an ordered bootstrap sample of tstatistics are selected and, as a result, the critical values for the construction of an appropriate confidence interval become available. A thorough explanation of the tpercentile method goes beyond the scope of this article, therefore the interested reader is referred to the excellent article by Berkowitz & Kilian (2000).
In particular, the lower (upper) CIs will be the lower (upper) bounds of our estimator while the quantities $\mathbb{E}({v}^{\bullet})$ $\mathbb{E}({v}^{\u25cb})$ are estimated through the mean operator, that is, (6) $${\mu}^{\u25cb}=\sum _{j=1}^{6}{v}_{j}^{\u25cb}$$and (7) $${\mu}^{\bullet}=\sum _{j=1}^{6}{v}_{j}^{\bullet}$$
At this point, it is worth emphasizing that the procedure not only, as just seen, requires very little in terms of input data (only the time series of the positives and the deaths are required) but also can be performed in an automatic fashion. In fact, once the data become available, one has just to properly assign the time series to the subsets Ω^{○} and Ω^{•} and the code will process the new data in an automatic way. The procedure is also very fast, as the computing time needed for the generation of the bootstrap samples requires—for the sample size in question—less than 2 min. Both code and dataset employed in this article have been uploaded as Supplemental Files. However, the data can also be downloaded free of charge at the following web address: https://github.com/pcmdpc/COVID19/tree/master/datiregioni (the file name is dpccovid19itaregioni20200323.csv).
Empirical Evidences
In order to give the reader the opportunity to gain a better insight on the different epidemic dynamical behaviors, in Figure 1–5 the time series of the variable C (as defined in Eq. 2) is reported for each region. Note that the sudden variations noticeable in Fig. 5 (Bolzano), Fig. 4 (Valle D’Aosta) and Fig. 3 (Molise and Campania) are due to the little number of tests administrated (i.e., the denominator of the variable w_{T} (2)) for these cases. In emergency situations the data are usually noisy, incomplete and might show large spikes, as in the case of Fig. 5.
That said, the main result of the article is summarized in Table 1, where three estimates of the number of positives are reported by region. The regions belonging to the set Ω^{○} (no deaths) are in Italics whereas all the others, belonging to the set Ω^{•}, are in a standard format. In the columns “Mean” and “Lower (Upper) Bounds”, the bootstrap estimates computed according to Eqs. (6) and (7) and the Lower (Upper) Bounds the lower (upper) bootstrap CIs are respectively reported. The column denominated “Official Cases” accounts for the number of positives cases released by the Italian Authorities, whereas the column “Morbidity” expresses the percentage ratio between μ^{•} (6) or μ^{○} (7) and the actual population of each region, as recorded by the Italian National Institute of Statistics. The latter source of data can be freely accessed at the web address http://dati.istat.it/Index.aspx? DataSetCode = DCIS_ POPRES1.
Lower bound  Mean  Upper bound  Official cases  Population  Morbidity (%)  

Abruzzo  526  600  807  78  1,311,580  0.06 
Basilicata  48  54  70  8  562,869  0.01 
Bolzano  697  730  795  103  531,178  0.15 
Calabria  182  238  493  32  1,947.131  0.03 
Campania  988  1,292  2,676  174  5,801.692  0.05 
Emilia Romagna  10,980  12,299  14,897  1,758  4,459,477  0.33 
Friuli Venezia Giulia  983  1,201  2,514  148  1,215,220  0.21 
Lazio  1,485  1,680  2,089  172  5,879,082  0.04 
Liguria  1,346  1,608  1,995  243  1,550,640  0.13 
Lombardia  37,744  45,020  49,723  6,896  10,060,574  0.49 
Marche  3,151  3,891  4,593  570  1,525,271  0.30 
Molise  119  134  167  16  305,617  0.05 
Piemonte  3,216  3,703  4,217  554  4,356,406  0.10 
Puglia  490  670  1292  98  4,029,053  0.03 
Sardegna  244  278  375  39  1,639,591  0.02 
Sicilia  776  865  1,098  111  4,999,891  0.02 
Toscana  2,352  2,755  3,965  352  3,729,641  0.11 
Trento  670  764  1,028  102  541,098  0.19 
Umbria  432  481  611  62  882,015  0.07 
Valle Aosta  139  183  356  26  125,666  0.28 
Veneto  8,382  9,343  12,028  1,297  4,905,854  0.25 
Totale Italia  74,950  87,789  105,789  12,839  60,359,546  0.18 
By examining the data for the whole Country, it is clear how the data collected by the Italian Authorities on the positive cases cannot be indicative of the situation at the population level, which appear to be greater by a factor of 8. Such a consideration, straightforward from a statistical point of view, might be worth outlining as many sources of information (e.g., newspaper, TV) mainly focus on the simple count of the positive cases so that the general public might miss the magnitude of this disease. As expected, the top three regions in terms of number of infected persons are Lombardia, Emilia Romagna and Veneto, where the estimated infected population is respectively (bootstrap mean) around 45,020, 12,299 and 9,343.
On the other hand, the risk of contagion is relatively low in some regions—mostly located in the Southern part of Italy—and in the island of Sardinia.
Regarding the regions included in the subset Ω^{○}, the application of the Piccolo distance (π) has generated the associations reported in Table 2.
Ω^{○}  Ω^{•}  π 

Basilicata  Veneto  0.0389 
Calabria  Campania  0.6211 
Molise  Lazio  0.4212 
Sardegna  Abruzzo  0.0157 
Trento  Abruzzo  0.00186 
Umbria  Sicilia  0.01398 
Model validation
The validation of the proposed approach is very simple and exploits the official Covid 19 mortality rate ( $K={\displaystyle \frac{\mathrm{D}\mathrm{E}\mathrm{A}\mathrm{T}\mathrm{H}}{\mathrm{I}\mathrm{N}\mathrm{F}\mathrm{E}\mathrm{C}\mathrm{T}\mathrm{E}\mathrm{D}}}$) issued by the WHO, which can be considered a well recognized and authoritative source. In essence, this constant—called Khas been used to make an estimate of the number of infected people (see Formula 8). Recalling that, in Italy, each and every person whose death was considered suspicious has been tested for Covid, it can be assumed the data related to these deaths to represent a population in itself (in other words, no inference procedures needed). The mortality rate, at the time of the writing of the paper, is K = 3.4%. By applying the simple formula (8) $$P={\displaystyle \frac{\mathrm{D}\mathrm{E}\mathrm{A}\mathrm{T}\mathrm{H}}{K}}$$where DEATH refers to the number of deceased people, it is possible to have a rough estimate of the total positives (P) at a population level. However, this is not the whole story. In fact, it is well known that the virus is not capable to kill a person instantly but it takes several days to do so. Therefore, Formula 8 is now rewritten to account for this temporal lag, that is, (9) $${P}_{t}={\displaystyle \frac{{\mathrm{D}\mathrm{E}\mathrm{A}\mathrm{T}\mathrm{H}}_{t+h}}{K}}$$where h is the delay time, which can be easily estimated by considering the empirical correlation function at different lags. In Fig. 6 such a structure is reported until lag h = 20. As it can be noticed, the highest correlation is at the lag h = 6.
Recalling that in Italy the number of Covid19 related deaths, at the date of March 12th 2020, reached the number of 2,978, by applying 9 and using h = 6, we have: $\frac{2,978}{0.35}=85,085.71$. This number is very consistent with the estimate given in the article, which is 87.789.
Even considering higher lags, that is, h = 7,8, Eq. (9) yields respectively the following number of deaths: 97,286 and 115,200. Both these results are still within the upper confidence interval given in the article (≈105,789). Shorter lags can always be considered, even though the scientific community seems to exclude them.
Additionally, to validate the number of deaths due to Covid19, the number of deaths occurred in the first quarter of 2020 with the average number of the deaths recorded in the first quarters of the years between 2015 and 2019 have been compared. It turns out that the total number of deaths ascribable to the Covid19 is roughly equal to the difference between these two quantities.
Conclusions
It is widespread opinion in the scientific community that current official data on the diffusion of SARSCoV2, responsible of the correlated disease, COIVD19,among population, are likely to suffer from a strong downward bias. In this scenario, the aim of this article is twofold: on one hand, it generates realistic figures on the effective number of people infected with SARSCoV2 at a national and regional level; on the other hand, it provides a methodology representing a viable alternative to those interested to apply inference procedures on the diffusion of epidemics.
This article proposes a methodology—illustrated in “Data and Contagion Indicator”—based on simple counts, that is, the number of deaths and the number of people tested positive to the virus for Italy, to:

provide an estimation at the national and regional level of the number of infected people and the related confidence intervals;

extend Eqs. (1) and (2) to those regions exhibiting no deaths as a consequence of the contraction of the Covid19.
The entire procedure has been written in the programing language R^{®} and uses official data as published by the Italian National Institute of Health. The whole code is available as a Supplemental Files in the Journal’s repository.
The results obtained show that, while official data at March 12th report, for Italy, a total of 12,839 cases, the people infected with the Covid19 could be as high as 105,789. This result, along with the estimated average doubling time for the Covid19 (≈6.2 days), confirms that this pandemic is to be regarded as much more dangerous than currently foreseen.
As it is well known, there are many critical questions about Covid19 that remain unanswered. One of them is related to the mortality rate Kused in Formula 8—whose accuracy should be carefully evaluated as more and more data become available. In fact, as observed by the WHO itself, K can be biased due to the denominator of the formula being not equal to the number of infected (clinical + asymptomatic).
Supplemental Information
R code to input data series.
The code import the data and build 21 matrices (positives and death).
Bootstrap generation and population estimation.
R code designed to bootstrapped the time series and apply the formula (1).