1. Introduction
Arellano-Valle et al. [1] introduced the skew-generalized-normal (SGN) distribution. We say that a random variable Z follows a SGN distribution, denoted as , if its probability density function (pdf) takes the following form:
(1)
where , , , and denote the pdf and the cumulative distribution function (cdf) of the distribution, respectively. For the case , the SGN distribution is reduced to standard normal distribution. On the other hand, non-zero values of parameter , directly influence the skewness of the model. In particular, when the SGN distribution is reduced to the well-known skew-normal (SN) distribution introduced by Azzalini [2]. Moreover, it converges to the half-normal distribution when .The SGN has been previously used in the literature. In this sense, Sever et al. [3] used this model in the context of discriminant analysis; Arnold et al. [4] considered the bivariate case; later Gómez et al. [5] examined the skew-curved-normal distribution, which is a subfamily of the SGN distribution. The singularity of the Fisher’s information matrix was examined by Arellano-Valle et al. [6]; in that paper it was concluded that for the SGN model with location and scale parameters, the Fisher information matrix is singular for the particular case when the normality is restored ().
Arrué et al. [7], considered bias reduction of the ML estimate of the shape parameter in the SGN distribution when . This submodel was named modified skew-normal (MSN). They also showed that Fisher’s information matrix of the MSN model is non-singular when the shape parameter is null. However, the issue of the divergence of the ML estimator for the shape parameter behaves in a similar way as in the SN model (see Sartori [8]). Hence, they applied the method introduced by Firth [9] to reduce the bias of this estimator for the shape parameter.
The MSN model is a convenient alternative to the SN model because in addition to regulating the asymmetry by means of a single parameter, it allows us to apply regular asymptotic theory to study the behavior of the MLE around normality. However, these two models do not have enough flexibility to adjust data from asymmetric distributions with heavy tails. With the purpose of producing robust inferences under these situations, we consider in this work a distribution that has more flexibility than the MSN distribution. Specifically, we consider the modified skew-t-normal (MStN) distribution, which extends the MSN distribution (this is a limiting case when the degrees of freedom tend to infinite) and consequently it is a more flexible model than the MSN distribution. In addition, it is a very useful model to describe abnormal data. On the other hand, the problem with the estimation of the shape parameter persists in this new distribution. For that reason, the main goal of this work is to examine bias reduction by using Firth’s method in the MStN distribution by assuming that the degrees of freedom are known. It will be seen that this leads us to similar results to those obtained in the MSN distribution.
The structure of the paper is as follows. In Section 2 the MStN is considered and some of its main properties are examined. In Section 3 the likelihood-based inference and the singularity of the Fisher’s information matrix of this distribution are studied. Next, in Section 4, the bias reduction methodology introduced by Firth [9] is described. This approach was used by Sartori [8] for the skew-normal and skew-t distributions and by Arrué et al. [7] for the MSN distribution. Later, this method is applied, in particular, to the shape parameter of the MStN distribution. It is shown that modified ML estimation is always finite. Furthermore, for the case of location, scale, and shape parameters, the methodology is applied by combining with the ML estimates of the location and shape parameters. Two applications of this methodology to real datasets are considered in Section 5. Finally, the conclusions are given in Section 6.
2. MStN Distribution
We say that a random variable Z follows a modified skew t-normal distribution, denoted by , if its pdf can be expressed as
(2)
where , and . Here, and denote the pdf of the Student’s-t distribution with degrees of freedom and the cdf of the distribution, respectively. For the particular case that , then the pdf of the MStN distribution given in (2) is equivalent to the pdf of the Student’s-t distribution. On the other hand, non-zero values of directly affect the symmetry and kurtosis of the model. In particular, when the model converges to the half-t distribution. Additionally, when , MStN approaches to the MSN distribution. Figure 1 shows the behavior of the pdf of the MStN distribution for different values of and .The thick solid line corresponds to the case , i.e., Student’s-t distribution ( left panel and right panel). The two panels illustrate that skewness and kurtosis are modified with values of and . Furthermore, for the limiting case it coincides with the distribution. Finally, the normal density is obtained when . We can also incorporate location and scale parameters, i.e., , with , and . Then, the new model is denoted by , with pdf given by
(3)
The MStN distribution given by (3) provides a suitable parametric model for fitting empirical data from asymmetric and heavy-tailed distributions. An important property of the MStN distribution given by (3) is the existence and non-singularity of the Fisher’s information matrix when . This means that we can use regular asymptotic theory to test if there is symmetry in the underlying distribution of the data.2.1. Properties
The MStN distribution has a series of interesting formal properties. The following proposition lists some of them in the standardized case with and .
The MStN distribution satisfies the following properties:
-
1.
If , then .
-
2.
If , then (with positive support); in particular, if , then .
-
3.
.
-
4.
(Modified Skew-Cauchy-Normal).
-
5.
If , then ; in particular, if , then .
-
6.
If , with conditional pdf
-
7.
If , with conditional pdf
-
8.
If , with conditional pdf
Property 8 shows the genesis of the MStN distribution, indicating that it belongs to the family of scale-shape mixtures of SN distributions defined in Arellano-Valle et al. [10], with a gamma and a normal mixing distributions for the scale V and shape S random variable, respectively. The proof of this property is obtained by integrating the conditional pdf in , and then using the following well-known facts: (i) if , then ; (ii) if , then . Properties 6 and 7 are direct consequences of Property 8 by the removal of one of the mixing variables. Property 6 represents the MStN distribution as a skew-scale mixture of the MSN distribution (Ferreira et al. [11]), while Property 7 represents it as a shape mixture of the MStN distribution. The remaining properties are proved straightforward.
2.2. Moments
For , the moment of order k of is given by
That is, we have where by Property 2, the random variable has a half- distribution with pdf , . This means that has the same distribution of , where and is independent of and so of , the half-normal distribution with pdf , . Thus, we find for and each that , with(4)
This is proved by the following proposition:Proposition 2.If , with , then the random variable has moment of order k, , given by
with and defined in (4) and .Note in Proposition 2 that , , and . In particular, when k is odd, the first of these relationships implies that . For k even, is constant on and hence it becomes a even function of .
2.3. Skewness and Kurtosis
Assuming , the skewness coefficient can be computed by using the results in Proposition 2 and the expression given by
Since when k is odd, then is an odd function of . This can also be observed in the left panel of Figure 2. The range of its values can be determined in terms of its minimum and maximum for each value of , as can be observed in Table 1. These values are obtained from the expression
Similarly, for the kurtosis coefficient can be obtained from Proposition 2 and the expression given by In this case, for k even does not depend on and hence is an even function of , as is displayed in the right hand side panel of Figure 2. Again, minimum and maximum values allow us to compute the range of this coefficient. These values can be obtained from the expressionsThe ranges of values of the skewness and kurtosis coefficients coincide with the range of the respective coefficients of the StN distribution, as displayed in Table 1 for .
In Figure 2 it is observed, that the skewness coefficient is an odd function and the width of the interval decreases when increases. When , we obtain the skewness range of the MSN and SN models, which is given by . On the other hand, the kurtosis coefficient is an even function. It is observed that the lower end and upper end of the interval, together with the width of the interval, decreases with the value of . Once again when , the kurtosis range of the models MSN and SN is obtained, whose interval corresponds to .
3. Inference
In this section we use the in order to obtain a robust inference on the parameters . In this sense, the degrees of freedom parameter , with , will be considered as initially known, and in a second stage from a grid of values for we will select that value with the best likelihood and AIC.
Given a random sample from the random variable , the log-likelihood function for parameter vector is
(5)
where and .The associated score functions are
where . From these equations, ML estimates of must be obtained numerically.The MStN Fisher’s information matrix computed for as has entries given by
where with for k odd, , and , which must be computed assuming that .In the symmetric case with , we have that , and so the Fisher information matrix reduces to
The expressions for and , with , were computed using the
On the order hand, as , the matrix , where is the Fisher’s information matrix of the model (see also Arrué et al. [7]).
As the MStN model is a regular parametric model, we have for each that the ML estimator of is consistent and such that as . This fact holds also when , and hence for the ML estimator of obtained from the model.
4. Bias Reduction for the ML Estimates
For each value of , the ML estimate of obtained from the MStN model overestimates the true value of this parameter. This fact can be seen in Table 2 below.
In addition, the ML estimate of could be infinite, with a certain probability, when all observations have the same sign, e.g., if all of them are positive so that . This non-zero probability of divergence when estimating , increases when the true values of and grow, however it quickly declines with the sample size. This can be observed in Figure 4.
For the case of the shape parameter with location and scale parameters, the overestimation of these parameters only occurs for . This can be noted in Table 3.
As the bias of the ML estimates of and is virtually zero, it is sensible to apply Firth’s method (see Ref. [9]), to reduce the bias of the ML estimate of to (see Cox and Snell [13]). Therefore, by doing this, we obtain a new ML bias-corrected estimate of , i.e., unbiased to . Since then it is verified that , and we only proceed for the case .
4.1. Preliminary Results
Consider a regular parametric model with log-likelihood function , and let us also suppose that the parameter is a scalar; for instance . In addition, let and , where and are the first and second derivatives of l, respectively, and consider the following expected quantities:
For a random sample of size n, we will consider of order and , and are of order (see Sartori [8]). In order to obtain the reduced ML estimator (least-bias estimator), denoted by , we must solve the following likelihood equation:(6)
where and (see Sartori [8]).The quasi-likelihood function associated with (6) is given by
where c is an arbitrary real number. This function allows us to numerically find a reduced bias estimate of , say, . Moreover, it can be used to compute confidence intervals of by means of the likelihood ratio statistics. In fact, since is a penalized likelihood with a bound penalty function of order , then the log-likelihood ratio statistics based on the expression(7)
has the usual asymptotic distribution . It can be used to calculate confidence intervals for , since it better captures the skewness of the log-likelihood compared to the normal asymptotic distribution.4.2. Shape Parameter Case
Consider now the baseline case with , and we take a sample of size n, say , from , with being the unknown parameter. In this case, the log-likelihood function is obtained from (5) by letting and . However, since is assumed to be known, it becomes proportional to the function
The ML estimate is infinite, with a certain probability, e.g., when , since the log-likelihood is an increasing function in and goes to zero when . The score function and the observed information are given by and , respectively.
Now, by using the notation the modified function takes the following expression,
(8)
Note that when k is odd, and when both k and h are even.
The left panel of Figure 5 below shows the graphs of the modified function of the SN, MStN for and and the MSN models. All of them are bound and odd functions for all and they tend to zero when (see Proposition 3). They take the maximum value at , with , with and respectively. Furthermore, it is also observed that the larger the value of , the more closely the modified function associated with the MStN distribution approaches the modified function of the MSN model. On the right panel of Figure 5, the different graphs of the integrated modified function M are displayed. It is noticeable that this is a decreasing and even function with respect to .
In order to guarantee the existence of estimator , the following proposition is needed.
Let be the modification of the score function for the distribution, then whatever the value of i.e., the rate of convergence of the tails of , is of the order .
See Appendix A. □
First Simulation Study
We have performed a simulation analysis with 5000 iterations for a random variable Z that follows a MStN distribution, by assuming that , and are known, for different sample sizes and different values of and .
It can be inferred from Table 4 that an overestimation of the parameter exists and there are also cases where the estimation is ∞. Obviously, it depends on the sample size, degrees of freedom (), and the true value of the shape parameter (). After applying Firth’s method to the shape parameter , we obtain a new estimate , which is finite and always exists. This is consistent with Proposition 3. The bias reduction of is quite good, taking into account that this method is applied whenever the estimate is finite and/or infinite. In addition, there exists an underestimation of when its value is large and the sample size is small. The empirical coverage is very close to the nominal value , although it is slightly lower when the sample size is small. This is due to the fact that the coverage is affected by the percentage of , which approaches infinity.
4.3. Location, Scale, and Shape Case
Similar to the previous case (shape parameter), the ML estimate of obtained from the log-likelihood function for given by (5), could be infinite, with a certain probability, when the random sample satisfies , where is the respective ML estimate of . Since the bias of the ML estimates of and is virtually zero, then it seems reasonable to apply the bias reduction method to only the shape parameter .
Let be the profile log-likelihood for , where and are the ML estimates for a known value of . We also define the profile of the modified log-likelihood equation as follows
(9)
where is the profile of the score function, and M is the modified function given in (8). We also use the profile of the quasi-log-likelihood associated with (9), given by where c is an arbitrary real number. As M is bound, then the likelihood ratio statistics has the usual asymptotic distribution . This is useful to calculate the confidence interval for .Second Simulation Study
Again, we have performed a simulation analysis with 5000 iterations for a random variable , by assuming that , are unknown and is known, for different sample sizes and different values of and .
In a similar way as in the case of the scalar parameter, it is observed in Table 5, that there exists an overestimation of the parameter . Moreover, there are cases where these estimates approach ∞ (with a higher percentage than in the previous case). Nevertheless, ML estimates of the location parameter and scale parameter are quite good since they always exist, they are finite, and bias is close to zero for both cases. For that reason, we only apply Firth’s method to the parameter . The new estimate, , always exists and it is finite. Besides, the bias is reduced for both cases, i.e., when is finite and/or infinite. Similarly to the previous case (without location and scale parameter), there also exists an underestimation of the parameter for large values of and small sample size, however the size of the underestimation is of lower magnitude. The empirical coverage is very close to the nominal value , although it is slightly lower when the sample size is small and the value of is large. This is due to the fact that the coverage is affected by the percentage of , which approaches infinity.
5. Applications
In this section we present two applications to real datasets; the first database is available as Supplementary Materials and second at
5.1. First Application
We consider a dataset that deals with the nickel concentration from 86 soil samples analyzed by the Mining Department of Universidad de Atacama, Chile. Table 6, shows the descriptive statistics of this dataset including the sample skewness coefficient () and the sample kurtosis coefficient (),
Now, an exploration of the ML estimates for the MStN distribution, assuming different known values for the parameter , is carried out to examine the behavior of the log-likelihood function. Table 7 illustrates the performance of the latter function. It is observed that the maximum value of this function () occurs when .
Table 8, shows the ML estimates of the parameters for the SN, MSN, and MStN models, respectively. Standard errors (shown in brackets) were obtained by inverting the Fisher’s information matrix for each model. In addition, two measures of model selection, the maximum of the log-likelihood function, and Akaike’s Information Criterion (AIC), are displayed in Table 8. It can be concluded that the model introduced in this paper provides a good fit to data as compared to other competing models.
Figure 6 exhibits the histogram of the nickel concentration dataset. Furthermore, we have superimposed the densities of the (dotted line) and (solid line) models.
QQ-plots of MStN and MSN models and cdf of the empirical distribution (ogive) and MSN and MStN distributions are shown in Figure 7 and Figure 8 respectively, obtained from the parameter estimates given above. Both tables confirm the good fit to data of the model presented in this paper.
Table 9 shows the ML estimates , , and the modified ML estimates .
The value of the modified ML estimates is lower than the ML estimates . In addition, by construction, it has a lower bias.
Table 10 displays the confidence interval of parameter for three different confidence levels. It is observed that the confidence intervals computed using the modified ML estimate of , , are more accurate than the interval based on the ML estimate of , , since they are narrower.
5.2. Second Application
We present a second application of the proposed model, to a dataset related with Betaplasma. The data were obtained from a study of patients () subjected to Betaplasma ingestion in their diet with the object of measuring plasma concentrations of betacarotene (ng/ml). The human body converts betacarotene into vitamin A, which then acts as an antioxidant, preventing oxidation damage to cells.
Table 11, presents a summary of the descriptive statistics, reflecting high dispersion, asymmetry, and a marked kurtosis value.
We studied the ML estimates for the MStN distribution, assuming different known values for parameter . This enabled us to analyze the behavior of the log-likelihood function. Table 12, illustrates the performance of the log-likelihood function. The maximum value of this function () occurs when .
Table 13, shows the ML estimates of the parameters for the SN, MSN, and MStN models, respectively, and the standard errors (in parentheses). The table also shows two model selection measures, the maximum log-likelihood function, and the Akaike information criterion (AIC). We may conclude that the model presented in this paper provides a better fit with the data than the SN and MSN models.
Figure 9, shows the histogram of the plasma concentration of betacarotene. It also presents the densities of the model (dotted line) and model (continuous line).
Bias reduction was applied to parameter of the MStN model, but the same estimated value was obtained as for the ML. This result is not surprising, since when the sample size () is relatively large the estimates tend to be more precise.
The left panel of Figure 9 visually shows the good fit of the MStN model with the plasma concentration of the betacarotene data. The right panel presents a close-up of the right tail of the distribution, showing how the proposed model is better at capturing the extreme values than the MSN model.
The QQ plots of the MStN and MSN models, and the CDF of the empirical distributions (ogive) of the MSN and MStN models are shown in Figure 10 and Figure 11, respectively, obtained from the parameter estimates given above. These graphs show the good fit of the proposed model for datasets with high extreme values.
6. Concluding Remarks
In this work, a modified maximum likelihood estimator is proposed to solve the issue of overestimation for the shape parameter in the MStN model. This problem is solved by finding a new maximum likelihood estimator () by using Firth’s method. Although the estimates of can be finite or infinite, the bias of this new estimator is lower. Moreover, the existence of is proved. A simulation study was carried out for the shape parameter case and for the case where location, scale, and shape parameters were considered. The conclusions for this simulation analysis are given below:
For both cases, the bias of the new ML estimator are reduced after applying Firth’s method. For both cases, this methodology satisfactorily reduces the bias of the new estimator even in situations where the estimates of are infinite;
For the second case, the method is not applied to parameters y , since the bias is very close to zero;
For the first case, the empirical coverage is very close to the nominal value , and is slightly lower when the sample size is small; this is due to the higher percentage of infinite values for . For the second case, the empirical coverage is lower for situations where is large and n is small;
For the first case, there exists an underestimation of the parameter , when the parameter takes large values and the sample size is small. For the second case, similar results are observed but the degree of the underestimation is of lower magnitude.
Conceptualization, J.A. and H.W.G.; methodology, R.B.A.-V. and H.W.G.; software, J.A. and E.C.-O.; validation, R.B.A.-V., O.V. and H.W.G.; formal analysis, R.B.A.-V., E.C.-O. and H.W.G.; investigation, J.A.; writing—original draft preparation, J.A. and R.B.A.-V.; writing—review and editing, R.B.A.-V., E.C.-O. and O.V.; funding acquisition, O.V. and H.W.G. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
The first data is available as
The authors declare no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 1. Different plots of the pdf of the MStN distribution for different values of [Forumla omitted. See PDF.] when [Forumla omitted. See PDF.] (left panel) and [Forumla omitted. See PDF.] (right panel).
Figure 4. Probability of divergence for the ML estimate of [Forumla omitted. See PDF.] for the MStN model when [Forumla omitted. See PDF.] (left panel) and [Forumla omitted. See PDF.] (right panel).
Figure 5. Modified function (left panel) and integrated modified function (right panel) for SN (dashed line), MStN with [Forumla omitted. See PDF.] (solid line) and [Forumla omitted. See PDF.] (thick solid line) and MSN (dotted line).
Figure 6. Histogram of the nickel concentration data together with the pdf of MSN model (dotted) and MStN (solid).
Figure 8. Empirical cdf (ogive) versus theoretical cdf for the MStN and MSN models.
Figure 9. Histogram of the plasma concentration of betacarotene data with the pdf of the MSN (dotted) and MStN (continuous) models.
Figure 11. Empirical cdf (ogive) versus theoretical cdf for the MStN and MSN models.
Skewness and kurtosis ranges for different values of
|
Skewness Range | Kurtosis Range |
---|---|---|
5 | (−2.550, 2.550) | (9.00, 23.109) |
7 | (−1.798, 1.798) | (5.000, 9.461) |
9 | (−1.539, 1.539) | (4.200, 7.054) |
11 | (−1.407, 1.407) | (3.857, 6.082) |
13 | (−1.326, 1.326) | (3.667, 5.561) |
15 | (−1.272, 1.272) | (3.545, 5.237) |
17 | (−1.233, 1.233) | (3.462, 5.017) |
19 | (−1.204, 1.204) | (3.400, 4.857) |
∞ | (−0.995, 0.995) | (3.000, 3.869) |
ML estimate of
|
|
|
|||||
---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
5 | 3 | 6.99 | 70.68 (71.04) | 7.00 | 95.10 (95.49) | 5.82 | 99.74 (99.80) |
5 | 7.17 | 71.48 (72.28) | 7.00 | 95.36 (95.95) | 5.86 | 99.90 (99.84) | |
10 | 7.06 | 74.52 (73.24) | 6.97 | 96.24 (96.29) | 5.83 | 99.84 (99.86) | |
10 | 3 | 11.52 | 45.62 (45.09) | 14.58 | 77.32 (77.66) | 13.73 | 95.18 (95.01) |
5 | 12.31 | 46.44 (46.15) | 15.00 | 79.08 (78.72) | 13.64 | 96.16 (95.47) | |
10 | 12.67 | 47.26 (47.00) | 14.38 | 80.40 (79.55) | 13.88 | 95.54 (95.82) |
a Calculated when
Bias of
n |
|
|
Bias ( |
Bias ( |
|
|
---|---|---|---|---|---|---|
50 | 5 | 3 | 0.003 | 0.006 | 6.860 | 83.78 |
100 | 5 | 3 | −0.005 | 0.010 | 6.797 | 96.98 |
200 | 5 | 3 | −0.001 | 0.003 | 5.665 | 99.88 |
50 | 10 | 3 | 0.016 | −0.010 | 11.052 | 61.80 |
100 | 10 | 3 | 0.004 | 0.001 | 13.499 | 85.24 |
200 | 10 | 3 | 0.000 | 0.002 | 13.211 | 97.60 |
50 | 5 | 5 | 0.003 | 0.004 | 6.840 | 85.54 |
100 | 5 | 5 | −0.001 | 0.000 | 6.637 | 97.52 |
200 | 5 | 5 | −0.001 | 0.003 | 5.666 | 99.96 |
50 | 10 | 5 | 0.016 | −0.011 | 11.254 | 61.96 |
100 | 10 | 5 | 0.004 | −0.003 | 13.645 | 86.08 |
200 | 10 | 5 | 0.001 | −0.001 | 12.845 | 98.00 |
50 | 5 | 10 | 0.008 | −0.006 | 6.798 | 87.10 |
100 | 5 | 10 | 0.000 | 0.001 | 6.525 | 98.22 |
200 | 5 | 10 | 0.000 | 0.002 | 5.641 | 99.94 |
50 | 10 | 10 | 0.017 | −0.015 | 11.404 | 65.28 |
100 | 10 | 10 | 0.002 | −0.001 | 13.621 | 87.80 |
200 | 10 | 10 | −0.001 | 0.000 | 13.200 | 98.20 |
a Calculated when
Bias of
n |
|
|
Bias( |
Bias( |
|
|
---|---|---|---|---|---|---|
20 | 5 | 3 | 1.867 | −1.583 | 0.94 | 71.64 (71.04) |
50 | 5 | 3 | 1.754 | −0.298 | 0.95 | 95.06 (95.49) |
100 | 5 | 3 | 0.788 | −0.030 | 0.95 | 99.82 (99.80) |
20 | 10 | 3 | 1.991 | −6.034 | 0.90 | 44.78 (45.09) |
50 | 10 | 3 | 4.299 | −2.866 | 0.94 | 76.80 (77.66) |
100 | 10 | 3 | 3.856 | −0.694 | 0.94 | 94.84 (95.01) |
20 | 5 | 5 | 2.148 | −1.513 | 0.94 | 72.62 (72.28) |
50 | 5 | 5 | 1.802 | −0.293 | 0.95 | 96.48 (95.95) |
100 | 5 | 5 | 0.815 | −0.004 | 0.95 | 99.80 (99.84) |
20 | 10 | 5 | 2.197 | −5.949 | 0.90 | 46.46 (46.15) |
50 | 10 | 5 | 4.116 | −2.751 | 0.94 | 79.38 (78.72) |
100 | 10 | 5 | 3.862 | −0.626 | 0.95 | 95.38 (95.47) |
20 | 5 | 10 | 2.177 | −1.479 | 0.94 | 72.82 (73.24) |
50 | 5 | 10 | 2.103 | −0.236 | 0.96 | 96.64 (96.29) |
100 | 5 | 10 | 0.776 | 0.018 | 0.95 | 99.90 (99.86) |
20 | 10 | 10 | 2.274 | −5.888 | 0.91 | 47.42 (47.00) |
50 | 10 | 10 | 4.169 | −2.626 | 0.94 | 79.18 (79.55) |
100 | 10 | 10 | 4.338 | −0.600 | 0.95 | 95.88 (95.82) |
a Calculated when
Bias of
n |
|
|
Bias ( |
Bias ( |
Bias ( |
Bias ( |
|
|
---|---|---|---|---|---|---|---|---|
50 | 5 | 3 | 0.004 | 0.005 | 1.889 | −0.898 | 0.93 | 84.24 |
100 | 5 | 3 | −0.002 | 0.005 | 1.617 | −0.306 | 0.94 | 97.42 |
200 | 5 | 3 | −0.001 | 0.004 | 0.712 | −0.106 | 0.95 | 99.84 |
50 | 10 | 3 | 0.017 | −0.008 | 1.097 | −3.971 | 0.87 | 61.86 |
100 | 10 | 3 | 0.004 | −0.003 | 3.301 | −1.699 | 0.91 | 85.40 |
200 | 10 | 3 | 0.001 | 0.000 | 3.188 | −0.534 | 0.94 | 97.96 |
50 | 5 | 5 | 0.004 | 0.002 | 1.828 | −0.855 | 0.94 | 86.70 |
100 | 5 | 5 | −0.002 | 0.005 | 1.755 | −0.255 | 0.94 | 97.58 |
200 | 5 | 5 | −0.001 | 0.002 | 0.628 | −0.098 | 0.95 | 99.78 |
50 | 10 | 5 | 0.016 | −0.013 | 1.185 | −3.842 | 0.87 | 63.74 |
100 | 10 | 5 | 0.005 | −0.002 | 3.736 | −1.508 | 0.92 | 86.72 |
200 | 10 | 5 | 0.000 | 0.000 | 3.053 | −0.401 | 0.94 | 97.78 |
50 | 5 | 10 | 0.006 | −0.003 | 1.832 | −0.813 | 0.92 | 86.84 |
100 | 5 | 10 | 0.000 | 0.002 | 1.523 | −0.245 | 0.95 | 98.00 |
200 | 5 | 10 | 0.000 | 0.000 | 0.578 | −0.108 | 0.94 | 99.94 |
50 | 10 | 10 | 0.014 | −0.013 | 1.530 | −3.689 | 0.88 | 64.18 |
100 | 10 | 10 | 0.004 | −0.002 | 3.622 | −1.311 | 0.92 | 86.40 |
200 | 10 | 10 | 0.002 | −0.002 | 2.700 | −0.450 | 0.93 | 98.06 |
Descriptive statistics of the nickel concentration dataset.
Data | n |
|
s |
|
|
---|---|---|---|---|---|
Nickel | 86 |
|
|
|
|
ML estimates of the parameters of the MStN model when
|
|
|
|
|
---|---|---|---|---|
13.092 | 6.284 | 0.954 | 1 | −343.057 |
8.676 | 11.182 | 2.235 | 2 | −338.775 |
7.083 | 13.767 | 2.994 | 3 | −338.260 |
6.335 | 15.345 | 3.492 | 4 | −338.483 |
5.858 | 16.458 | 3.875 | 5 | −338.864 |
21.439 | 11.410 | −0.518 | 6 | −349.399 |
Parameter estimates, standard errors,
MLE | SN | MSN | MStN |
---|---|---|---|
|
2.625 (1.136) | 2.571 (1.260) | 7.083 (1.402) |
|
24.968 (1.913) | 25.027 (2.153) | 13.767 (1.838) |
|
10.261 (4.751) | 10.619 (5.239) | 2.994 (0.789) |
|
- | - | 3 |
|
−344.762 | −344.769 | −338.260 |
AIC | 693.524 | 693.538 | 682.520 |
ML estimate and modified ML estimate of
|
|
|
|
|
|
---|---|---|---|---|---|
7.083 (1.402) | 13.767 (1.838) | 2.994 (0.789) | - | −338.260 | - |
7.083 (1.447) | 13.767 (1.843) | - | 2.838 (0.731) | - | −338.305 |
Confidence interval for
MLE |
|
|
|
---|---|---|---|
IC | (1.696, 4.292) | (1.373, 4.615) | (1.158, 4.830) |
IC* | (1.635, 4.040) | (1.336, 4.339) | (1.137, 4.539) |
Descriptive statistics of the plasma concentration of betacarotene.
Data | n |
|
s |
|
|
---|---|---|---|---|---|
Betaplasma | 315 |
|
|
|
|
ML estimates of the parameters of the MStN model when
|
|
|
|
|
---|---|---|---|---|
64,737 | 74,446 | 2514 | 1 | −1.931,939 |
52,120 | 107,392 | 3946 | 2 | −1.917,631 |
45,800 | 127,352 | 5081 | 3 | −1.918,405 |
42,250 | 140,472 | 5930 | 4 | −1.921,499 |
39,815 | 150,177 | 6608 | 5 | −1.924,907 |
Parameter estimates, standard errors,
MLE | SN | MSN | MStN |
---|---|---|---|
|
21.332 (5.180) | 21.332 (5.170) | 52.120(5.686) |
|
248.613 (10.510) | 248.594 (10.505) | 107.391(8.519) |
|
18.834(6.968) | 18.954(6.971) | 3.945(0.728) |
|
- | - | 2 |
|
−1976.317 | −1976.319 | −1917.631 |
AIC | 3956.634 | 3956.638 | 3839.262 |
BIC | 3969.896 | 3969.896 | 3852.520 |
Supplementary Materials
The following supporting information can be downloaded at:
References
1. Arellano-Valle, R.B.; Gómez, H.W.; Quintana, F.A. A New Class of Skew-Normal Distributions. Commun. Stat. Theory Methods; 2004; 33, pp. 1465-1480. [DOI: https://dx.doi.org/10.1081/STA-120037254]
2. Azzalini, A. A class of distributions which includes the normal ones. Scand. J. Stat.; 1985; 12, pp. 171-178.
3. Sever, M.; Lajovic, J.; Rajer, B. Robustness of the Fishers discriminant function to skew-curved normal distribution. Metodoloskizvezki; 2005; 2, pp. 231-242.
4. Arnold, B.C.; Castillo, E.; Sarabia, J.M. Distributions with Generalized Skewed Conditionals and Mixtures of such Distributions. Commun. Stat. Theory Methods; 2007; 36, pp. 1493-1504. [DOI: https://dx.doi.org/10.1080/03610920601125862]
5. Gómez, H.W.; Castro, L.M.; Salinas, H.S.; Bolfarine, H. Properties and Inference on the Skew-curved-symmetric Familiy of Distributions. Commun. Stat. Theory Methods; 2010; 39, pp. 884-898. [DOI: https://dx.doi.org/10.1080/03610920902807887]
6. Arellano-Valle, R.B.; Gómez, H.W.; Salinas, H.S. A note on the Fisher information matrix for the skew-generalized-normal model. SORT; 2013; 37, pp. 19-28.
7. Arrué, J.; Arellano-Valle, R.B.; Gómez, H.W. Bias reduction of maximum likelihood estimates for a modified skew normal distribution. J. Stat. Comput. Simul.; 2016; 86, pp. 2967-2984. [DOI: https://dx.doi.org/10.1080/00949655.2016.1143471]
8. Sartori, N. Bias prevention of maximum likelihood estimates for scalar skew normal and skew t distributions. J. Stat. Plan. Inference; 2006; 136, pp. 4259-4275. [DOI: https://dx.doi.org/10.1016/j.jspi.2005.08.043]
9. Firth, D. Bias reduction of maximum likelihood estimates. Biometrika; 1993; 80, pp. 27-38. Erratum in Biometrika 1993, 82, 667 [DOI: https://dx.doi.org/10.1093/biomet/80.1.27]
10. Arellano-Valle, R.B.; Ferreira, C.S.; Genton, M.G. Scale and shape mixtures of multivariate skew-normal distributions. J. Multivar. Anal.; 2018; 166, pp. 98-110. [DOI: https://dx.doi.org/10.1016/j.jmva.2018.02.007]
11. Ferreira, C.S.; Bolfarine, H.; Lachos, V.H. Skew scale mixtures of normal distributions: Properties and estimation. Stat. Methodol.; 2011; 8, pp. 154-171. [DOI: https://dx.doi.org/10.1016/j.stamet.2010.09.001]
12. Wolfram Research, Inc. Mathematica, Version 10.0; Wolfram Research, Inc.: Champaign, IL, USA, 2014.
13. Cox, D.R.; Snell, E.J. A general definition of residuals. J. R. Stat. Soc. B Stat. Methodol.; 1968; 30, pp. 248-275. [DOI: https://dx.doi.org/10.1111/j.2517-6161.1968.tb00724.x]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
In this paper, likelihood-based inference and bias correction based on Firth’s approach are developed in the modified skew-t-normal (MStN) distribution. The latter model exhibits a greater flexibility than the modified skew-normal (MSN) distribution since it is able to model heavily skewed data and thick tails. In addition, the tails are controlled by the shape parameter and the degrees of freedom. We provide the density of this new distribution and present some of its more important properties including a general expression for the moments. The Fisher’s information matrix together with the observed matrix associated with the log-likelihood are also given. Furthermore, the non-singularity of the Fisher’s information matrix for the MStN model is demonstrated when the shape parameter is zero. As the MStN model presents an inferential problem in the shape parameter, Firth’s method for bias reduction was applied for the scalar case and for the location and scale case.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details




1 Departamento de Estadística y Ciencias de Datos, Facultad de Ciencias Básicas, Universidad de Antofagasta, Antofagasta 1240000, Chile
2 Departamento de Estadística, Facultad de Matemáticas, Pontificia Universidad Católica de Chile, Santiago 7820436, Chile
3 Centre for Actuarial Studies, Department of Economics, The University of Melbourne, Melbourne, VIC 3010, Australia
4 Departamento de Ciencias Matemáticas y Físicas, Facultad de Ingeniería, Universidad de Católica de Temuco, Temuco 4780000, Chile