YeoJohnsonTransformationen
The Yeo–Johnson transformation is a statistical method used to stabilize variance and improve normality of real-valued data. It extends the Box–Cox transformation to accommodate both positive and negative observations without the need to shift data. Introduced by Ishk I. K. Yeo and Raymond A. Johnson in 2000, it provides a continuous, monotone family of power transformations controlled by a single parameter lambda.
For a real-valued observation x, the transformed value T(x; lambda) is defined piecewise as follows:
- If x >= 0: T = ((x + 1)^lambda - 1) / lambda, for lambda not equal to 0; T = log(x
- If x < 0: T = -((-x + 1)^(2 - lambda) - 1) / (2 - lambda), for lambda not equal to 2;
The parameter lambda is typically chosen to maximize normality or to stabilize variance, often via maximum
Key advantages include handling of zero and negative values without pre-shifting, and the ability to approximate