This package delivers a scikit-learn
compatible Python 3 package for sundry state-of-the art multivariate statistical methods, with
a focus on dimension reduction.
The categories of methods delivered in this package, are:
- Projection pursuit dimension reduction (
ppdire
folder; cf. ppdire Documentation file and ppdire Examples Notebook) - Sufficient dimension reduction (
sudire
folder; cf. sudire Documentation file and sudire Examples Notebook)) - Robust M-estimators for dimension reduction (
sprm
folder; cf. SPRM Documentation file and SPRM Examples Notebook)
Further documentation can be found in the direpack
vignette.
We hope that this package leads to scientific success. If it does so, we kindly ask to cite the direpack
publication [0], as well as the original publication of the corresponding method.
The package also contains a set of tools for pre- and postprocessing:
- The
preprocessing
folder provides classical and robust centring and scaling, as well as spatial sign transforms [4] - The
dicomo
folder contains a versatile class to access a wide variety of moment and co-moment statistics, and statistics derived from those. Check out the dicomo Documentation file and the dicomo Examples Notebook. - Plotting utilities in the
plot
folder - Cross-validation utilities in the
cross-validation
folder
- The estimator (
sprm.py
) [1] - The Sparse NIPALS (SNIPLS) estimator [3](
snipls.py
) - Robust M regression estimator (
rm.py
) - Ancillary functions for M-estimation (
_m_support_functions.py
)
The ppdire
class will give access to a wide range of projection pursuit dimension reduction techniques.
These include slower approximate estimates for well-established methods such as PCA, PLS and continuum regression.
However, the class provides unique access to a set of robust options, such as robust continuum regression (RCR) [5], through its native grid
optimization algorithm, first
published for RCR as well [6]. Moreover, ppdire
is also a great gateway to calculate generalized betas, using the CAPI projection index [7].
The code is orghanized in
ppdire.py
- the main PP dimension reduction classcapi.py
- the co-moment analysis projection index.
The sudire
folder gives access to an extensive set of methods that resort under the umbrella of sufficient dimension reduction.
These range from meanwhile long-standing, well-accepted approaches, such as sliced inverse regression (SIR) and the closely related SAVE [8,9],
through methods such as directional regression [10] and principal Hessian directions [11], and more. However, the package also contains some
of the most recently developed, state-of-the-art sufficient dimension reduction techniques, that require no distributional assumptions.
The options provided in this category are based on energy statistics (distance covariance [12] or martingale difference divergence [13]) and
ball statistics (ball covariance) [14]. All of these options can be called by setting the corresponding parameters in the sudire
class, cf. the docs.
Note: the ball covariance option will require some lines to be uncommented as indicated. We decided not to make that option generally available,
since it depends on the Ball
package that seems to be difficult to install on certain architectures.
The package is distributed through PyPI, so install through:
pip install direpack
Detailed documentation on how to use the classes is provided in the docs
folder per class.
Jupyter Notebooks with Examples are provided for each of the classes in the examples
folder.
direpack
: A Python 3 package for state-of-the-art statistical dimension reduction methods- Sparse partial robust M regression, Irene Hoffmann, Sven Serneels, Peter Filzmoser, Christophe Croux, Chemometrics and Intelligent Laboratory Systems, 149 (2015), 50-59.
- Partial robust M regression, Sven Serneels, Christophe Croux, Peter Filzmoser, Pierre J. Van Espen, Chemometrics and Intelligent Laboratory Systems, 79 (2005), 55-64.
- Sparse and robust PLS for binary classification, I. Hoffmann, P. Filzmoser, S. Serneels, K. Varmuza, Journal of Chemometrics, 30 (2016), 153-162.
- Spatial Sign Preprocessing: A Simple Way To Impart Moderate Robustness to Multivariate Estimators, Sven Serneels, Evert De Nolf, Pierre J. Van Espen, Journal of Chemical Information and Modeling, 46 (2006), 1402-1409.
- Robust Continuum Regression, Sven Serneels, Peter Filzmoser, Christophe Croux, Pierre J. Van Espen, Chemometrics and Intelligent Laboratory Systems, 76 (2005), 197-204.
- Robust Multivariate Methods: The Projection Pursuit Approach, Peter Filzmoser, Sven Serneels, Christophe Croux and Pierre J. Van Espen, in: From Data and Information Analysis to Knowledge Engineering, Spiliopoulou, M., Kruse, R., Borgelt, C., Nuernberger, A. and Gaul, W., eds., Springer Verlag, Berlin, Germany, 2006, pages 270--277.
- Projection pursuit based generalized betas accounting for higher order co-moment effects in financial market analysis, Sven Serneels, in: JSM Proceedings, Business and Economic Statistics Section. Alexandria, VA: American Statistical Association, 2019, 3009-3035.
- Sliced Inverse Regression for Dimension Reduction Li K-C, Journal of the American Statistical Association (1991), 86, 316-327.
- Sliced Inverse Regression for Dimension Reduction: Comment, R.D. Cook, and Sanford Weisberg, Journal of the American Statistical Association (1991), 86, 328-332.
- On directional regression for dimension reduction , B. Li and S.Wang, Journal of the American Statistical Association (2007), 102:997–1008.
- On principal hessian directions for data visualization and dimension reduction:Another application of stein’s lemma, K.-C. Li. , Journal of the American Statistical Association(1992)., 87,1025–1039.
- Sufficient Dimension Reduction via Distance Covariance, Wenhui Sheng and Xiangrong Yin in: Journal of Computational and Graphical Statistics (2016), 25, issue 1, pages 91-104.
- A martingale-difference-divergence-based estimation of central mean subspace, Yu Zhang, Jicai Liu, Yuesong Wu and Xiangzhong Fang, in: Statistics and Its Interface (2019), 12, number 3, pages 489-501.
- Robust Sufficient Dimension Reduction Via Ball Covariance Jia Zhang and Xin Chen, Computational Statistics and Data Analysis 140 (2019) 144–154
Release Notes can be checked out in the repository.
A list of possible topics for further development is provided as well. Additions and comments are welcome!