Weighted Target Estimation
- Weighted target estimation is a statistical method that reweights each observation's score based on its fit, ensuring robust and efficient parameter estimation.
- It modifies the maximum likelihood equation by assigning adaptive weights derived from residuals comparing the empirical and model distributions.
- Empirical studies demonstrate that this approach maintains full efficiency under ideal models while effectively mitigating the influence of outliers and model violations.
Weighted target estimation refers to a set of statistical methodologies in which individual data contributions are reweighted to produce more robust, efficient, or targeted parameter estimates. These reweighting techniques are often motivated by the need for robustness to outliers, handling contamination, or ensuring that the estimator remains fully efficient under the assumed parametric model. The work by Markatou, Basu, and Lindsay (Majumder et al., 2016) develops a new weighted likelihood approach aimed at achieving robustness without sacrificing efficiency—by constructing carefully designed weights for each observation’s contribution in the likelihood score equation.
1. Fundamentals of the Weighted Likelihood Approach
The central innovation is a modification of the maximum likelihood estimating equation such that each observation’s score contribution is multiplied by an adaptive weight. These weights depend on the compatibility of the observation with the fitted parametric model. If an observation fits well under the model, its score receives full weight (weight ≈ 1). If it appears incompatible (e.g., is an outlier), the corresponding weight is diminished. By attaching the weight to the score function of each observation—not to the likelihood itself—this approach avoids routinely downweighting observations and only penalizes those not aligning with the bulk of the data.
The general weighted score equation for parameter θ is:
where , is the model density, and is a weight function dependent on a residual that measures discrepancy between the empirical and model distributions at .
2. Construction of Residuals and Weight Functions
2.1 Residual Function
For univariate i.i.d. data, the residual quantifies how much the empirical distribution deviates from the model distribution at . Specifically:
with and . The parameter controls the fraction of observations, in each tail, that may be downweighted.
2.2 Weight Function Specification
The weight function must satisfy:
- (full weight for central, well-modeled points),
- is at least twice differentiable at $0$,
- and ,
- decays to zero as and is small at .
A general device is to start with a density , defined on with and mode at (after reparametrization), then set:
Several concrete are used:
- Gamma density: , with , so mode is at $1$.
- Weibull, generalized extreme value (GEV), and -related densities.
For the gamma weight,
As , for all , and weighted likelihood estimation reduces to MLE.
3. Weighted Score Equation and Estimator
The weighted likelihood estimator (WLE) is defined as the solution to:
Each data point thus has its score function adaptively weighted by , responsive to its degree of model fit.
4. Theoretical Properties
4.1 Fisher Consistency
If is the true data-generating model, then the estimator is Fisher consistent: .
4.2 Influence Function
At the true model, the WLE’s influence function is the same as the MLE:
where is the Fisher information. Higher-order influence function analysis is required to reveal robustness away from the model.
4.3 Asymptotic Efficiency and Consistency
Under standard regularity conditions, is consistent, asymptotically normal, and efficient at the model:
4.4 Equivariance
The estimator is equivariant for location-scale families: if , then the transformation of the estimate is .
5. Comparison to Prior Weighted Likelihood Methods
The classical weighted likelihood approach of Markatou et al. constructed weights via smoothed residuals from kernel density estimates, requiring bandwidth selection and being unsuitable for bounded support models. In contrast, the present method compares empirical and model cdfs, avoiding kernel estimation subtleties, and providing a broader class of possible weight functions (gamma, Weibull, GEV, F-based, etc.). Moreover, by ensuring and , asymptotic full efficiency is retained.
6. Empirical Studies and Real Data Evidence
Simulation studies with contaminated normal and exponential models demonstrate that, under no contamination, the WLE achieves MSE virtually identical to the MLE. Under moderate contamination (e.g., with additional high-variance or shifted mean components), WLE’s MSE remains stable while the MLE’s MSE increases significantly, showing the robustness of the new approach.
Applications to real data include:
- Poisson regression on Drosophila data: WLE matches the outlier-deleted MLE.
- Normal model fit to Newcomb’s speed of light data: WLE with small closely tracks the cleaned MLE, downweighting outliers.
- Exponential fit to Melbourne rainfall data: WLE downweights extreme right-tail observations.
- Multivariate and regression cases: WLE delivers robust fits (e.g., robust concentration ellipses and regression lines) resistant to influential outliers.
7. Robustness, Efficiency, and Tuning
Comprehensive theoretical analysis confirms Fisher consistency, equivariance, and asymptotic normality. Robustness arises from the second-order influence function, while first-order local efficiency is preserved. Practical performance is controlled via the tuning parameters in the weight function (typical values include , , , etc.), balancing robustness against efficiency.
With H(0) = 1 and a suitable tuning, the WLE is fully efficient when the model is correct and robust under mild to moderate model violations. Compared with kernel-based weighted likelihood approaches, this method is both algorithmically simpler and theoretically more general, being deployable in a broad array of models—including location-scale, Poisson, exponential, regression, and multivariate contexts.
In sum, the weighted target estimation method introduced in this work systematically attaches a data-driven, model-adaptive weight to each observation’s score contribution. This results in estimators with strong robustness properties against outliers and contamination, while retaining the efficiency and invariance expected of the maximum likelihood estimator under correct model specification. The design of the residual-based weight function, the clean theoretical guarantees, and the empirical performance across diverse settings position this approach as a foundational tool for robust inference in statistical modeling.