Article ID: | iaor2017512 |
Volume: | 58 |
Issue: | 4 |
Start Page Number: | 515 |
End Page Number: | 532 |
Publication Date: | Dec 2016 |
Journal: | Australian & New Zealand Journal of Statistics |
Authors: | Prendergast L A, Garnham A L |
Keywords: | forecasting: applications |
In the regression setting, dimension reduction allows for complicated regression structures to be detected via visualisation in a low‐dimensional framework. However, some popular dimension reduction methodologies fail to achieve this aim when faced with a problem often referred to as symmetric dependency. In this paper we show how vastly superior results can be achieved when carrying out response and predictor transformations for methods such as least squares and sliced inverse regression. These transformations are simple to implement and utilise estimates from other dimension reduction methods that are not faced with the symmetric dependency problem. We highlight the effectiveness of our approach via simulation and an example. Furthermore, we show that ordinary least squares can effectively detect multiple dimension reduction directions. Methods robust to extreme response values are also considered.