Some algorithms pre-assume that our
data is centered at “0”. For example, if we initialize the weights of a small
multi-layer perceptron with tanh activation units to 0 or small random values centered
around zero, we want to update the model weights “equally.” As a rule of thumb
I would say: When in doubt, just standardize the data.
Normalization
Also known as min-max scaling or min-max normalization, is
the simplest method and consists in re-scaling the range of features to scale
the range in [0, 1] or [−1, 1]. Selecting the target range depends on the
nature of the data.
It will be useful when we are sure enough that there are no anomalies (i.e. outliers) with extremely large or small values. For example, in a recommender system, the ratings made by users are limited to a small finite set like
.
In some situations, we may prefer to map data to a range like
with zero-mean.2 Then we should choose mean normalization
x′=
x−mean(x)
max(x)−min(x)
In this way, it will be more convenient for us to use other techniques like matrix factorization
Standardization
In machine learning, we can handle various types of data,
e.g. audio signals and pixel values for image data, and this data can include
multiple dimensions.
Feature standardization makes the values of each feature in the data have
zero-mean (when subtracting the mean in the numerator) and unit-variance. This
method is widely used for normalization in many machine learning algorithms
(e.g., support vector machines, logistic regression, and artificial neural networks)
No comments:
Post a Comment