Skip to Main content Skip to Navigation
Journal articles

Reducing errors-in-variables bias in linear regression using compact genetic algorithms

Abstract : A new technique is devised to mitigate the errors-in-variables bias in linear regression. The procedure mimics a 2-stage least squares procedure where an auxiliary regression which generates a better behaved predictor variable is derived. The generated variable is then used as a substitute for the error-prone variable in the first-stage model. The performance of the algorithm is tested by simulation and regression analyses. Simulations suggest the algorithm efficiently captures the additive error term used to contaminate the artificial variables. Regressions provide further credit to the simulations as they clearly show that the compact genetic algorithm-based estimate of the true but unobserved regressor yields considerably better results. These conclusions are robust across different sample sizes and different variance structures imposed on both the measurement error and regression disturbances.
Complete list of metadata
Contributor : Diyarbakirlioglu Erkin Connect in order to contact the contributor
Submitted on : Wednesday, April 27, 2022 - 10:55:49 AM
Last modification on : Wednesday, April 27, 2022 - 10:55:52 AM


  • HAL Id : hal-01134592, version 1



Mehmet Hakan Satman, Erkin Diyarbakirlioglu. Reducing errors-in-variables bias in linear regression using compact genetic algorithms. Journal of Statistical Computation and Simulation, Taylor & Francis, 2014, pp.20. ⟨hal-01134592⟩



Record views