Search icon CANCEL
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Big Data Analysis with Python

You're reading from   Big Data Analysis with Python Combine Spark and Python to unlock the powers of parallel computing and machine learning

Arrow left icon
Product type Paperback
Published in Apr 2019
Publisher Packt
ISBN-13 9781789955286
Length 276 pages
Edition 1st Edition
Languages
Tools
Concepts
Arrow right icon
Authors (3):
Arrow left icon
Ivan Marin Ivan Marin
Author Profile Icon Ivan Marin
Ivan Marin
Sarang VK Sarang VK
Author Profile Icon Sarang VK
Sarang VK
Ankit Shukla Ankit Shukla
Author Profile Icon Ankit Shukla
Ankit Shukla
Arrow right icon
View More author details
Toc

Table of Contents (11) Chapters Close

Big Data Analysis with Python
Preface
1. The Python Data Science Stack 2. Statistical Visualizations FREE CHAPTER 3. Working with Big Data Frameworks 4. Diving Deeper with Spark 5. Handling Missing Values and Correlation Analysis 6. Exploratory Data Analysis 7. Reproducibility in Big Data Analysis 8. Creating a Full Analysis Report Appendix

Chapter 07: Reproducibility in Big Data Analysis


Activity 14: Test normality of data attributes (columns) and carry out Gaussian normalization of non-normally distributed attributes

  1. Import the required libraries and packages in the Jupyter notebook:

    import numpy as np
    import pandas as pd
    import seaborn as sns
    import time
    import re
    import os
    import matplotlib.pyplot as plt
    sns.set(style="ticks")
  2. Now, import the libraries required for preprocessing:

    import sklearn as sk
    from scipy import stats
    from sklearn import preprocessing
  3. Set the working directory using the following command:

    os.chdir("/Users/svk/Desktop/packt_exercises")
  4. Now, import the dataset into the Spark object:

    df = pd.read_csv('bank.csv', sep=';')
  5. Identify the target variable in the data:

    DV = 'y'
    df[DV]= df[DV].astype('category')
    df[DV] = df[DV].cat.codes
  6. Generate training and testing data using the following command:

    msk = np.random.rand(len(df)) < 0.8
    train = df[msk]
    test = df[~msk]
  7. Create the Y and X data, as illustrated here:

    # selecting the target variable (dependent variable) as y
    y_train = train[DV]
  8. Drop the DV or y using the drop command:

    train = train.drop(columns=[DV])
    train.head()

    The output is as follows:

    Figure 7.22: Bank dataset

  9. Segment the data numerically and categorically and perform distribution transformation on the numeric data:

    numeric_df = train._get_numeric_data()

    Perform data preprocessing on the data.

  10. Now, create a loop to identify the columns with a non-normal distribution using the following command (converting to NumPy arrays for more efficient computation):

    numeric_df_array = np.array(numeric_df)
    loop_c = -1
    col_for_normalization = list()
    
    for column in numeric_df_array.T:
        loop_c+=1
        x = column
        k2, p = stats.normaltest(x) 
        alpha = 0.001
        print("p = {:g}".format(p))
            
        # rules for printing the normality output
        if p < alpha:
            test_result = "non_normal_distr"
            col_for_normalization.append((loop_c)) # applicable if yeo-johnson is used
            
            #if min(x) > 0: # applicable if box-cox is used
                #col_for_normalization.append((loop_c)) # applicable if box-cox is used
            print("The null hypothesis can be rejected: non-normal distribution")
            
        else:
            test_result = "normal_distr"
            print("The null hypothesis cannot be rejected: normal distribution")

    The output is as follows:

    Figure 7.23: Identifying the columns with a non-linear distribution

  11. Create a PowerTransformer based transformation (box-cox):

    pt = preprocessing.PowerTransformer(method='yeo-johnson', standardize=True, copy=True)

    Note

    box-cox can handle only positive values.

  12. Apply the power transformation model on the data. Select the columns to normalize:

    columns_to_normalize = numeric_df[numeric_df.columns[col_for_normalization]]
    names_col = list(columns_to_normalize)
  13. Create a density plot to check the normality:

    columns_to_normalize.plot.kde(bw_method=3)

    The output is as follows:

    Figure 7.24: Density plot to check the normality

  14. Now, transform the columns to a normal distribution using the following command:

    normalized_columns = pt.fit_transform(columns_to_normalize)
    normalized_columns = pd.DataFrame(normalized_columns, columns=names_col)
  15. Again, create a density plot to check the normality:

    normalized_columns.plot.kde(bw_method=3)

    The output is as follows:

    Figure 7.25: Another density plot to check the normality

  16. Use a loop to identify the columns with non-normal distribution on the transformed data:

    numeric_df_array = np.array(normalized_columns) 
    loop_c = -1
    
    for column in numeric_df_array.T:
        loop_c+=1
        x = column
        k2, p = stats.normaltest(x) 
        alpha = 0.001
        print("p = {:g}".format(p))
            
        # rules for printing the normality output
        if p < alpha:
            test_result = "non_normal_distr"
            print("The null hypothesis can be rejected: non-normal distribution")
            
        else:
            test_result = "normal_distr"
            print("The null hypothesis cannot be rejected: normal distribution")

    The output is as follows:

    Figure 7.26: Power transformation model to data

  17. Bind the normalized and non-normalized columns. Select the columns not to normalize:

    columns_to_notnormalize = numeric_df
    columns_to_notnormalize.drop(columns_to_notnormalize.columns[col_for_normalization], axis=1, inplace=True)
    1. Use the following command to bind both the non-normalized and normalized columns:

      numeric_df_normalized = pd.concat([columns_to_notnormalize.reset_index(drop=True), normalized_columns], axis=1)
      numeric_df_normalized

    Figure 7.27: Non-normalized and normalized columns

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $19.99/month. Cancel anytime