Feature scaling is relatively easy with Python. Note that it is recommended to split data into test and training data sets BEFORE scaling. If scaling is done before partitioning the data, the data may be scaled around the mean of the entire sample, which may be different than the mean of the test and mean of the train data. Standardization:
Any scripts or data that you put into this service are public. DMwR documentation built on May 1, 2019, 9:17 p.m. This package includes functions and data accompanying the book "Data Mining with R, learning with case studies" by Luis Torgo, CRC Press 2010. I'm going to assume that you mean , when you say "using a Gaussian Mixture Model", you mean fitting a mixture of (possibly multivariate) Gaussians to some data, for the purposes of clustering. In this case, provided you use maximum-likelihood as your condition for fitting the model, you don't need to scale your data.
What I've done first, is rescaled the data using min-max normalization: # Normalize data between 0 and 1 from sklearn.preprocessing import MinMaxScaler min_max = MinMaxScaler() dataframe2 = pd.DataFrame(min_max.fit_transform(dataframe), columns = dataframe.columns)
Select the desired columns from each downloaded dataset. Concatenate the DataFrames. Drop all NaNs from the new, merged DataFrame. Normalize each column (independently) to 0.0-1.0 in the new DataFrame using the code. df = (df - df.min ()) / (df.max () - df.min ()) Feed the normalized data into my neural network. Hence, having all variables on the same scale will facilitate easy comparison of the “importance” of each variable, as now all variables are on the same scale. The most common way to standardize the variable X X is to use the z z transformation: zi = xi −μ sdX z i = x i − μ s d X. . 60 293 310 285 134 398 31 190 155

how to unscale data in r