Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: We develop new models for imperfect repair and the corresponding generalized renewal processes for stochastic description of repairable items that fail when their degradation reaches the specified deterministic or random threshold. The discussion is based on the recently suggested notion of a random virtual age and is applied to monotone processes of degradation with independent increments. Imperfect repair reduces degradation of an item on failure to some intermediate level. However, for the nonhomogeneous processes, the corresponding age reduction, which sets back the ‘clock’ of the process, is also performed. Some relevant stochastic comparisons are obtained. It is shown that the cycles of the corresponding generalized imperfect renewal process are stochastically decreasing/increasing depending on the monotonicity properties of the failure rate that describes the random failure threshold of an item. PubDate: 2022-05-05
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Modeling bivariate (or multivariate) count data has received increased interest in recent years. The aim is to model the number of different but correlated counts taking into account covariate information. Bivariate Poisson regression models based on the shock model approach are widely used because of their simple form and interpretation. However, these models do not allow for overdispersion or negative correlation, and thus, other models have been proposed in the literature to avoid these limitations. The present paper proposes copula-based bivariate finite mixture of regression models. These models offer some advantages since they have all the benefits of a finite mixture, allowing for unobserved heterogeneity and clustering effects, while the copula-based derivation can produce more flexible structures, including negative correlations and regressors. In this paper, the new approach is defined, estimation through an EM algorithm is presented, and then different models are applied to a Spanish insurance claim count database. PubDate: 2022-05-04
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract We analyze reliability systems with components whose lifetimes are identically distributed, and whose joint distribution admits a Samaniego signature representation of the system lifetime distribution. Our main result is the following. We assume that two systems have the same structure and that the lifetimes of the components of the systems share the same dependence copula. If the first system lifetime precedes (succeeds) its single component lifetime in the convex transform order, and if also the component lifetime of the second system precedes the (succeeds) component lifetime of the first system in the convex transform order then the system-component ordering property is preserved by the second system lifetime, i.e., the system lifetime precedes (succeeds) the component lifetime in the second system also. This allows us to conclude various sufficient and necessary conditions on the system signatures under which the monotone failure rate and density properties of the component lifetimes are inherited by the system lifetime under the condition that the component lifetimes are independent. PubDate: 2022-04-27
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract In this study we propose an adaptive bi-level variable selection method to analyze multivariate failure time data. In the regression setting, we treat the coefficients corresponding to the same predictor variable as a natural group and then consider variable selection at the group level and individual level simultaneously. By imitating the group variable selection procedure with adaptive bi-level penalty, the proposed variable selection method can select a predictor variable at two different levels allowing different covariate effects for different event types: the group level where the predictor is important to all failure types, and the individual level where the predictor is only important to some failure types. An algorithm based on cycle coordinate descent is developed to carry out the proposed method. Based on the simulation results, our method outperforms the classical penalty methods, especially in removing unimportant variables for different failure types. We obtain the asymptotic oracle properties of the proposed variable selection method in the case of a diverging number of covariates. We construct a generalized cross-validation method for the tuning parameter selection and assess model performance using model errors. We also illustrate the proposed method using a real-life data set. PubDate: 2022-04-16
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract In this paper, we study the commonness of nonparametric component functions at different quantile levels in spatial additive autoregressive models. We propose two fused adaptive group LASSO penalties to shrink the difference of functions between neighbouring quantile levels. Using these methods, we can estimate the nonparametric functions and identify the quantile regions where functions are unvarying simultaneously. Therefore, when there exists a quantity-level region where the functions are unvarying, its performance is expected to be better than the conventional spatial quantile additive autoregressive model. Under some regularity conditions, the proposed penalized estimators can reach the optimal rate of convergence in theory and also recognize the true varying/unvarying regions accurately. At the same time, our proposed method shows good numerical results in simulated and real datasets. PubDate: 2022-04-10
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract Estimating a precision matrix is an important problem in several research fields when dealing with large-scale data. Under high-dimensional settings, one of the most popular approaches is optimizing a Lasso or \(\ell _1\) norm penalized objective loss function. This penalization endorses sparsity in the estimated matrix and improves the accuracy under a proper calibration of the penalty parameter. In this paper, we demonstrate that the problem of minimizing Lasso penalized D-trace loss can be seen as solving a penalized Sylvester matrix equation. Motivated by this method, we propose estimating the precision matrix using penalized generalized Sylvester matrix equations. In our method, we develop a particular estimating equation and a new convex loss function constructed through this equation, which we call the generalized D-trace loss. We assess the performance of the proposed method using detailed numerical analysis, including simulated and real data. Extensive results show the advantage of the proposed method compared to other estimation approaches in the literature. PubDate: 2022-04-09
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract The \(\delta \) -shock model is one of the basic shock models which has a wide range of applications in reliability, finance and related fields. In existing literature, it is assumed that the recovery time of a system from the damage induced by a shock is constant as well as the shocks magnitude. However, as technical systems gradually deteriorate with time, it takes more time to recover from this damage, whereas the larger magnitude of a shock also results in the same effect. Therefore, in this paper, we introduce a general \(\delta \) -shock model when the recovery time depends on both the arrival times and the magnitudes of shocks. Moreover, we also consider a more general and flexible shock process, namely, the Poisson generalized gamma process. It includes the homogeneous Poisson process, the non-homogeneous Poisson process, the Pólya process and the generalized Pólya process as the particular cases. For the defined survival model, we derive the relationships for the survival function and the mean lifetime and study some relevant stochastic properties. As an application, an example of the corresponding optimal replacement policy is discussed. PubDate: 2022-04-01
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract Joint distribution between two or more variables could be influenced by the outcome of a conditioning variable. In this paper, we propose a flexible Wald-type statistic to test for such influence. The test is based on a conditioned multivariate Kendall’s tau nonparametric estimator. The asymptotic properties of the test statistic are established under different null hypotheses to be tested for, such as conditional independence or testing for constant conditional dependence. Two simulation studies are presented: The first shows that the estimator proposed and the bandwidth selection procedure perform well. The second presents different bivariate and multivariate models to check the size and power of the test and runs comparisons with previous proposals when appropriate. The results support the contention that the test is accurate even in complex situations and that its computational cost is low. As an empirical application, we study the dependence between some pillars of European Regional Competitiveness when conditioned on the quality of regional institutions. We find interesting results, such as weaker links between innovation and higher education in regions with lower institutional quality. PubDate: 2022-04-01
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract For a given pair of random lifetimes whose dependence is described by a time-transformed exponential model, we provide analytical expressions for the distribution of their sum. These expressions are obtained by using a representation of the joint distribution in terms of bivariate distortions, which is an alternative approach to the classical copula representation. Since this approach allows one to obtain conditional distributions and their inverses in simple form, then it is also shown how it can be used to predict the value of the sum from the value of one of the variables (or vice versa) by using quantile regression techniques. PubDate: 2022-03-17
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract Statistical inference with nonresponse is quite challenging, especially when the response mechanism is nonignorable. In this case, the validity of statistical inference depends on untestable correct specification of the response model. To avoid the misspecification, we propose semiparametric Bayesian estimation in which an outcome model is parametric, but the response model is semiparametric in that we do not assume any parametric form for the nonresponse variable. We adopt penalized spline methods to estimate the unknown function. We also consider a fully nonparametric approach to modeling the response mechanism by using radial basis function methods. Using Pólya–gamma data augmentation, we developed an efficient posterior computation algorithm via Gibbs sampling in which most full conditional distributions can be obtained in familiar forms. The performance of the proposed method is demonstrated in simulation studies and an application to longitudinal data. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00774-y
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract This paper considers a high-dimensional linear regression problem where there are complex correlation structures among predictors. We propose a graph-constrained regularization procedure, named Sparse Laplacian Shrinkage with the Graphical Lasso Estimator (SLS-GLE). The procedure uses the estimated precision matrix to describe the specific information on the conditional dependence pattern among predictors, and encourages both sparsity on the regression model and the graphical model. We introduce the Laplacian quadratic penalty adopting the graph information, and give detailed discussions on the advantages of using the precision matrix to construct the Laplacian matrix. Theoretical properties and numerical comparisons are presented to show that the proposed method improves both model interpretability and accuracy of estimation. We also apply this method to a financial problem and prove that the proposed procedure is successful in assets selection. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00779-7
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract We develop a method originally proposed by R. A. Fisher into a general procedure, called tailoring, for deriving goodness-of-fit tests that are guaranteed to have a \(\chi ^{2}\) asymptotic null distribution. The method has a robustness feature that it works correctly in testing a certain aspect of the model while some other aspect of the model may be misspecified. We apply the method to small area estimation. A connection, and difference, to the existing specification test is discussed. We evaluate performance of the tests both theoretically and empirically, and compare the performance with several existing methods. Our empirical results suggest that the proposed test is more accurate in size, and has either higher or similar power compared to the existing tests. The proposed test is also computationally less demanding than the specification test and other comparing methods. A real-data application is discussed. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00772-0
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract This paper discusses disadvantages and limitations of the available inferential approaches in sequential clinical trials for treatment comparisons managed via response-adaptive randomization. Then, we propose an inferential methodology for response-adaptive designs which, by exploiting a variance stabilizing transformation into a bootstrap framework, is able to overcome the above-mentioned drawbacks, regardless of the chosen allocation procedure as well as the desired target. We derive the theoretical properties of the suggested proposal, showing its superiority with respect to likelihood, randomization and design-based inferential approaches. Several illustrative examples and simulation studies are provided in order to confirm the relevance of our results. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00777-9
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract Penalized estimation can conduct variable selection and parameter estimation simultaneously. The general framework is to minimize a loss function subject to a penalty designed to generate sparse variable selection. The majorization–minimization (MM) algorithm is a computational scheme for stability and simplicity, and the MM algorithm has been widely applied in penalized estimation. Much of the previous work has focused on convex loss functions such as generalized linear models. When data are contaminated with outliers, robust loss functions can generate more reliable estimates. Recent literature has witnessed a growing impact of nonconvex loss-based methods, which can generate robust estimation for data contaminated with outliers. This article investigates MM algorithm for penalized estimation, provides innovative optimality conditions and establishes convergence theory with both convex and nonconvex loss functions. With respect to applications, we focus on several nonconvex loss functions, which were formerly studied in machine learning for regression and classification problems. Performance of the proposed algorithms is evaluated on simulated and real data including cancer clinical status. Efficient implementations of the algorithms are available in the R package mpath in CRAN. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00770-2
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract In many existing methods of multiple comparison, one starts with either Fisher’s p value or the local fdr. One commonly used p value, defined as the tail probability exceeding the observed test statistic under the null distribution, fails to use information from the distribution under the alternative hypothesis. The targeted region of signals could be wrong when the likelihood ratio is not monotone. The oracle local fdr based approaches could be optimal because they use the probability density functions of the test statistic under both the null and alternative hypotheses. However, the data-driven version could be problematic because of the difficulty and challenge of probability density function estimation. In this paper, we propose a new method, Cdf and Local fdr Assisted multiple Testing method (CLAT), which is optimal for cases when the p value based methods are optimal and for some other cases when p value based methods are not. Additionally, CLAT only relies on the empirical distribution function which quickly converges to the oracle one. Both the simulations and real data analysis demonstrate the superior performance of the CLAT method. Furthermore, the computation is instantaneous based on a novel algorithm and is scalable to large data sets. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00775-x
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract Mixtures of t factor analyzers (MtFA) have been well recognized as a prominent tool in modeling and clustering multivariate data contaminated with heterogeneity and outliers. In certain practical situations, however, data are likely to be censored such that the standard methodology becomes computationally complicated or even infeasible. This paper presents an extended framework of MtFA that can accommodate censored data, referred to as MtFAC in short. For maximum likelihood estimation, we construct an alternating expectation conditional maximization algorithm in which the E-step relies on the first-two moments of truncated multivariate-t distributions and CM-steps offer tractable solutions of updated estimators. Asymptotic standard errors of mixing proportions and component mean vectors are derived by means of missing information principle, or the so-called Louis’ method. Several numerical experiments are conducted to examine the finite-sample properties of estimators and the ability of the proposed model to downweight the impact of censoring and outlying effects. Further, the efficacy and usefulness of the proposed method are also demonstrated by analyzing a real dataset with genuine censored observations. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00766-y
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract We introduce a new class of spatial Cox processes driven by a Hilbert-valued random log-intensity. We adopt a parametric framework in the spectral domain, to estimate its spatial functional correlation structure. Specifically, we consider a spectral functional, approach based on the periodogram operator, inspired on Whittle estimation methodology. Strong consistency of the parametric estimator is proved in the linear case. We illustrate this property in a simulation study under a Gaussian first-order Spatial Autoregressive Hilbertian scenario for the log-intensity model. Our method is applied to the spatial functional prediction of respiratory disease mortality in the Spanish Iberian Peninsula, in the period 1980–2015. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00773-z
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract In one-sided testing, Bayesians and frequentists differ on whether or not there is discrepancy between the inference based on the posterior model probability and that based on the p value. We add some arguments to this debate analyzing the discrepancy for moderate and large sample sizes. For small and moderate samples sizes, the discrepancy is measured by the probability of disagreement. Examples of the discrepancy on some basic sampling models indicate the somewhat unexpected result that the probability of disagreement is larger when sampling from models in the alternative hypothesis that are not located at the boundary of the hypotheses. For large sample sizes, we prove that the Bayesian one-sided testing is, under mild conditions, consistent, a property that is not shared by the frequentist procedure. Further, the rate of convergence is \(O(e^{nA})\) , where A is a constant that depends on the model from which we are sampling. Consistency is also proved for an extension to multiple hypotheses. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00778-8
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract We consider the problem of statistical inference in a parametric finite Markov chain model and develop a robust estimator of the parameters defining the transition probabilities via minimization of a suitable (empirical) version of the popular density power divergence. Based on a long sequence of observations from a first-order stationary Markov chain, we have defined the minimum density power divergence estimator (MDPDE) of the underlying parameter and rigorously derived its asymptotic and robustness properties under appropriate conditions. Performance of the MDPDEs is illustrated theoretically as well as empirically for some common examples of finite Markov chain models. Its applications in robust testing of statistical hypotheses are also discussed along with (parametric) comparison of two Markov chain sequences. Several directions for extending the MDPDE and related inference are also briefly discussed for multiple sequences of Markov chains, higher order Markov chains and non-stationary Markov chains with time-dependent transition probabilities. Finally, our proposal is applied to analyze corporate credit rating migration data of three international markets. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00771-1
Please help us test our new pre-print finding feature by giving the pre-print link a rating. A 5 star rating indicates the linked pre-print has the exact same content as the published article.
Abstract: Abstract The assessment of prevalence on regional levels is an important element of public health reporting. Since regional prevalence is rarely collected in registers, corresponding figures are often estimated via small area estimation using suitable health data. However, such data are frequently subject to uncertainty as values have been estimated from surveys. In that case, the method for prevalence estimation must explicitly account for data uncertainty to allow for reliable results. This can be achieved via measurement error models that introduce distribution assumptions on the noisy data. However, these methods usually require target and explanatory variable errors to be independent. This does not hold when data for both have been estimated from the same survey, which is sometimes the case in official statistics. If not accounted for, prevalence estimates can be severely biased. We propose a new measurement error model for regional prevalence estimation that is suitable for settings where target and explanatory variable errors are dependent. We derive empirical best predictors and demonstrate mean-squared error estimation. A maximum likelihood approach for model parameter estimation is presented. Simulation experiments are conducted to prove the effectiveness of the method. An application to regional hypertension prevalence estimation in Germany is provided. PubDate: 2022-03-01 DOI: 10.1007/s11749-021-00776-w