alternative
  • Home (current)
  • About
  • Tutorial
    Technologies
    C#
    Deep Learning
    Statistics for AIML
    Natural Language Processing
    Machine Learning
    SQL -Structured Query Language
    Python
    Ethical Hacking
    Placement Preparation
    Quantitative Aptitude
    View All Tutorial
  • Quiz
    C#
    SQL -Structured Query Language
    Quantitative Aptitude
    Java
    View All Quiz Course
  • Q & A
    C#
    Quantitative Aptitude
    Java
    View All Q & A course
  • Programs
  • Articles
    Identity And Access Management
    Artificial Intelligence & Machine Learning Project
    How to publish your local website on github pages with a custom domain name?
    How to download and install Xampp on Window Operating System ?
    How To Download And Install MySql Workbench
    How to install Pycharm ?
    How to install Python ?
    How to download and install Visual Studio IDE taking an example of C# (C Sharp)
    View All Post
  • Tools
    Program Compiler
    Sql Compiler
    Replace Multiple Text
    Meta Data From Multiple Url
  • Contact
  • User
    Login
    Register

Machine Learning - Machine Learning Development Life Cycle - Mathematical Transformation Tutorial

Mathematical Transformation will turns one function or graph into another  related function or graph by applying mathematical formula on column. In machine learning, dataset should be normal distributed or close to normal distributed (mean=median=mode). There are different ways to transform a continuous (numeric) variable so that the resulting variable looks more normally distributed. Some of them are-

  1. Function Transformer
  1. Log Trans
  2. Reciprocal
  3. Power (Sq | Sqrt)
  4. Custom
  1. Power Transformer
  1. Box-Cox
  2. Yeo-Johnson
  1. Quantile Transformer

How to find if data is normal?

  1. Using sns.distplot()
  2. Using pd.skew()
  3. QQ Plot

Advanced Statistical Concepts in Data Science


 

  1. Function Transformer
  1. Log Transformation

Log transformation is a data transformation method in which it replaces each variable x with a log(x). In other words, the log transformation reduces or removes the skewness of our original data.

Usually used for right skewed data.

Func = np.log1p i.e it will add 1 to value to prevent it making 0 after applying log

While np.log is simple log. It can become zero after applying log on 1

  1. Reciprocal

The reciprocal transformation is defined as the transformation of x to 1/x.

  1. Power (Sq | Sqrt)

Squared is usually used for left skewed data

  1. Custom
  1. Power Transformer
  1. Box-Cox

The exponent here is a variable called lambda that varies over the range of -5 to 5, and in the process of searching, we examine all values of lambda. Finally, we  choose the optimal value (resulting in the best approximation to a normal distribution) for your variable.

 

  1. Yeo-Johnson

This transformation is somewhat of an adjustment to the Box-Cox transformation, by which we can apply it to negative number.

  1. Quantile Transformer
Machine Learning

Machine Learning

  • Introduction
  • Overview
    • Type Of Machine Learning
    • Batch Vs Online Machine Learning
    • Instance Vs Model Based Learning
    • Challenges in Machine Learning
    • Machine Learning Development Life Cycle
  • Machine Learning Development Life Cycle
    • Framing the Problem
    • Data Gathering
    • Understanding your Data
    • Exploratory Data Analysis (EDA)
    • Feature Engineering
    • Principal Component Analysis
    • Column Transformer
    • Machine Learning Pipelines
    • Mathematical Transformation
    • Binning and Binarization | Discretization | Quantile Binning | KMeans Binning
  • Supervised Learning
    • Overview
    • Linear Regression [Regression]
    • Multiple Linear Regression
    • Polynomial Linear Regression [Regression]
    • Bias Variance Trade Off
    • Regularization
    • LOGISTIC REGRESSION [Regression & Classification]
    • Polynomial Logistic Regression
    • Support Vector Machines / Support Vector Regressor
    • Naïve Bayes Classifier [classification]
    • Decision Tree
    • Entropy
    • Information Gain
    • K Nearest Neighbor (KNN)
    • Neural Network (MultiLayer Perceptron)
  • Ensemble Learning
    • Introduction to Ensemble Learning
    • Basic Ensemble Techniques
    • Advanced Ensemble Techniques
    • Random Forest Classifier
    • Boosting
  • UnSupervised Learning
    • Overview
    • K Mean Clustering

About Fresherbell

Best learning portal that provides you great learning experience of various technologies with modern compilation tools and technique

Important Links

Don't hesitate to give us a call or send us a contact form message

Terms & Conditions
Privacy Policy
Contact Us

Social Media

© Untitled. All rights reserved. Demo Images: Unsplash. Design: HTML5 UP.

Toggle