## Simple linear regression with Python

Simple linear regression is the most basic form of regression. It is the foundation of statistical or machine learning modelling technique. All advance techniques you may use in future will be based on the idea and concepts of linear regression. It is the most primary skill to explore your data and have the first look into it.

Simple linear regression is a statistical model which studies the relationship between two variables. These two variables will be such that one of them is dependent on the other. A simple example of such two variables can be the height and weight of the human body. From our experience, we know that the bodyweight of any person is correlated with his height.

The body weight changes as the height changes. So here body weight and height are dependent and independent variable respectively. The task of simple linear regression is to quantify the change happens in the dependent variables for a unit change in the independent variable.

## Mathematical expression

We can express this relationship using a mathematical equation. If we express a person’s height and weight with X and Y respectively, then a simple linear regression equation will be:

Y=a.X+b

With this equation, we can estimate the dependent variable corresponding to any known independent variable. Simple linear regression helps us to estimate the coefficients of this equation.  As a is known now, we can say for one unit change in X, there will be exactly a unit change in Y.

See the figure below, the a in the equation is actually the slope of the line and b is the intercept from X-axis.

As the primary focus of this post is to implement simple linear regression through Python, so I would not go deeper into the theoretical part of it. Rather we will jump straight into the application of it.

Before we start coding with Python, we should know about the essential libraries we will need to implement this. The three basic libraries are NumPy,  pandas and matplotlib. I will discuss about these libraries briefly in a bit.

## Application of Python for simple linear regression

I know you were waiting for this part only. So, here is the main part of this post i.e. how we can implement simple linear regression using Python. For demonstration purpose I have selected an imaginary database which contains data on tree total biomass above the ground and several other tree physical parameters like tree commercial bole height,  diameter, height, first forking height, diameter at breast height, basal area. Tree biomass is the dependent variable here which depends on all other independent variables.

Here is a glimpse of the database:

From this complete dataset, we will use only Tree_height_m and Tree_biomass (kg) for this present demonstration. So, here the dataset name is tree_height and has the look as below:

## Python code for simple linear regression

### Importing required libraries

Before you start the coding, the first task is to import the required libraries. Give them a short name to refer them easily in the later part of coding.

``````import pandas as pd
import numpy as np
import matplotlib.pyplot as plt``````

These are the topmost important libraries for data science applications. These libraries contain several classes and functions which make performing data analysis tasks in Python super easy.

For example, numPy and Pandas are the two libraries which encapsulate all the matrix and vector operation functions. They allow users to perform complex matrix operations required for machine learning and artificial intelligence research with a very intuitive manner. Actually the name numPy comes from “Numeric Python”.

Whereas Matplotlib is a full-fledged plotting library and works as an extension of numPy. The main function of this library to provide an object-oriented API for useful graphs and plots embedded in the applications itself.

These libraries get automatically installed if you are installing Python from Anaconda, which is a free and opensource resource for R and Python for data science computation. So as the libraries are already installed you have to just import them.

### Importing dataset

``````dataset=pd.read_csv('tree_height.csv')
x=dataset.iloc[:,:-1].values
y=dataset.iloc[:, 1].values``````

Before you use this piece of code, make sure the .csv file you are about to import is located in the same working directory where the Python file is located. Otherwise, the compiler will not be able to find the file.

Then we have to create two variables to store the independent and dependent data. Here the use of matrix needs special mention. Please keep in mind that the dataset I have used has the dependent (Y) variable in the last column. So, while storing the independent variable in x, the last column is excluded and for dependent variable y, the location of the last column is considered.

### Splitting the dataset in training and testing data

``````from sklearn.model_selection import train_test_split
x_train, x_test, y_train, y_test=train_test_split(x,y,test_size=1/4, random_state=0)``````

This is of utmost importance when we are performing statistical modelling. Any model developed should be tested with an independent dataset which has net been used for model building. As we have only one dataset in our hand so, I have created two independent datasets with 80:20 ratio.

The train data consists of 80% of the data and used for training the model. Whereas rest of the 20% data was kept aside for testing the model. Luckily the famous sklearn library for Python already has a module called model_selection which contains a function called train_test_split.  We can easily get this data split task done using this library.

### Application of linear regression

``````from sklearn.linear_model import LinearRegression
regressor=LinearRegression()
regressor.fit(x_train,y_train)``````

This is the main part where the regression takes place using Linear Regression function of sklearn library.

### Printing coefficients

``````#To retrieve the intercept:
print(regressor.intercept_)
#For retrieving the slope:
print(regressor.coef_)``````

Here we can get the expression of the linear regression equation with the slope  and intercept constant.

### Validation plot to check homoscedasticity assumption

``````#***** Plotting residual errors in training data
plt.scatter(regressor.predict(x_train), (regressor.predict(x_train)-y_train),
color='blue', s=10, label = 'Train data')
# ******Plotting residual errors in testing data
plt.scatter(regressor.predict(x_test),regressor.predict(x_test)-y_test,
color='red',s=10,label = 'Test data')
#******Plotting reference line for zero residual error
plt.hlines(y=0,xmin=0,xmax=60)
plt.title('Residual Vs Predicted plot for train and test data set')
plt.xlabel('Residuals')
plt.ylabel('Predicted values')``````

For the data used here this part will create a plot like this:

This part is for checking an important assumption of a linear regression which is the residuals are homoscedastic. That means the residuals have equal variance. If this assumption fails then the whole regression process does not stand.

### Predicting the test results

``y_predict=regressor.predict(x_test)``

The independent test dataset is now in use to predict the result using the newly developed model.

### Printing actual and predicted values

``````new_dataset=pd.DataFrame({'Actual':y_test.flatten(), 'Predicted':y_predict.flatten()})
new_dataset``````

### Creating scatterplot using the training set

``````plt.scatter(x_train, y_train, color='red')
plt.plot(x_train, regressor.predict(x_train), color='blue')
plt.title('Tree heihgt vs tree weight')
plt.xlabel('Tree height (m)')
plt.ylabel('Tree wieght (kg)')``````

### Visualization of model’s performance using test set data

``````plt.scatter(x_test, y_test, color='red')
plt.plot(x_test, regressor.predict(x_test), color='blue')
plt.title('Tree heihgt vs tree weight')
plt.xlabel(‘Tree height (m)')
plt.ylabel('Tree wieght (kg)')``````

### Calculating fit statistics for the model

``````r_square=regressor.score(x_train, y_train)
print('Coefficient of determination(R square):',r_square)
from sklearn import metrics
print('Mean Absolute Error:', metrics.mean_absolute_error(y_test, y_predict))
print('Mean Squared Error:', metrics.mean_squared_error(y_test,y_predict))
print('Root Mean Squared Error:', np.sqrt(metrics.mean_squared_error(y_test, y_predict)))``````

This is th final step of finding the goodness of fit of the model. This piece of code generates some statistics which will quantitatively tell the performance of your model. Here the most important and popular four fit statistics are calculated. Except for the coefficient of determination, the lower the value of all other statistics better is the model.

## Getting started with Python for Machine Learning: beginners guide

If you are reading this article, then you are a Machine Learning enthusiast without any doubt. You must have already gone through the theoretical basics of it and getting impatient to try hand in your first Machine Learning application. Python is the most popular programming language for machine learning. I would suggest that if you want a carrier in data science it is Python which you should bet on for.

```Learn about two main types of Machine Learning
>Supervised machine learning
>Unsupervised machine learning```

But first of all the question is….

## Why Python for machine learning?

Why I have chosen Python to carry on Machine Learning? There are lots of tools available and some of them are very popular too. For example, R is a very reputed language and also present there for a long time.

Especially people with traditional statistical or mathematical background have a strong inclination towards R too. One of the reasons behind this popularity is R actually came into existence replacing S which was a pure statistical programming language developed on C platform and hence was hugely popular amongst statisticians.

### Python Vs R

R was developed in 1992 and has a specific edge for data analysis tasks. And being a procedural language it breaks down the total tasks into a series of steps and procedures. Both of R and Python being open source are freely available to use and online resources are huge.

R is mainly helpful for core statistical and data analytics purpose. The language was developed by statisticians keeping the need for statisticians in mind mainly. It has very powerful graphical functions like ggplot, ggvis, shiny etc. If you want to create eye-catching plots from your data, R should be your best friend.

On the other hand, Python came a little early in 1989 developed by Guido Van Rossum, a Dutch scientist. It has a slow steady growth till 2010 but after that with the start of data explosion era, its popularity also shoots quickly.

The main reason behind is so quick popularity is its simplicity and versatility. Machine Learning and Artificial Intelligence have many complex algorithms to perform several complex tasks. But the beauty of Python is that it makes tasks easy for both machine learning and AI with its vast collection of simple to use functions.

Use of Python in data science is just one of its capability. Being a general-purpose language, Python can be used for developing web applications, software, mobile applications development and even read-modifying files connecting to the database. This versatility of this language has won the heart of millions of people irrespective of whether they are data scientists or computer science enthusiasts.

If you are a beginner in data science you can jump-start the learning and application of Python even with little or no background in programming languages. It is also a far better performer compare to R when it comes to analyzing large size database.

The following chart from Economist.com will help you to realize how popular Python has become recently surpassing all other big names like Java, R, C++ etc.

Source: steelkiwi.com, economist.com

In the data science world these two programming languages are close competitors. Both of them are very popular and have their own plus and minuses. And ultimately which platform you should use is purely your choice.

Having said this, I think the popularity and simplicity of Python in its application in machine learning will keep it slightly ahead of R. And if you are looking ahead to build a career as a data scientist, in my opinion, the future is brighter with Python skill.

## Setting up Python in your computer

To start with python application, the first step is to install Python in your computer. If your desktop/laptop is a new one, then there is a chance that it might have Python preinstalled in it. You can check your start menu for it. If you get it there then skip this step.

After you downloaded the file, click it to start the installation. Just go with the recommended installation process. It is a quick process and within minutes python is installed in your system.

The following window will appear as the Python installation is finished.

Now you can check your computer start menu and the python folder with associated applications will be there.

### As I have installed it just now so it is having a “New” tag with all its application. Now as the python is installed you can directly launch its application and start your code.

Here in the above screenshot, you can see that the console is showing all the details of the Python version installed. I have also done some basic command like print and simple calculation.

But to start with your Python coding we will need a good IDE which will help us with Python syntax writing in an intuitive way.

### Selecting a Python IDE

Although while installing Python a simple IDE called IDLE gets installed automatically. We prefer to use a more popular and advanced IDE called PyCharm. The reason is to get familiar with one IDE of any programming language takes significant time. So, we should choose a good IDE to start with so that we can continue our task in it.

PyCharm is currently the most popular IDE for python. See the following table which compares some popular Python IDEs. PyCharm also comes with a paid version. But you will get full-featured integrated environment in both of them.

Source: www.softwaretestinghelp.com

Except for thesse IDEs some simple text editor like Notepad++ is also very popular amongst data scientists. The only issue with text editors is that you have to use some additional plugins to compile the code written in them. In this context IDEs come handy and you can do the complete task starting from writing code to its compilation in there itself.

Having said these, the final selection is completely on your choice. Python being the most popular programming language, users have the luxury of choosing an IDE from a vast collection of it. And honestly speaking, it is difficult to judge any single IDE to be the best one. Every one of them has its own strengths and weakness.

So though I have selected PyCharm here, you can select any other too. It will not take you much time to switch between IDEs once you make your basics strong.

### So, let’s start installing PyCharm

PyCharm is a product of Jetbrains. Open the concern page following this link

#### Step 1:

Click the download button under Community, it is the free open source version of PyCharm

#### Step: 3

In the next step, if you want to change the program location then provide the path or you can go with the default path assigned. I am here going with the default folder. Then click next.

#### Step: 4

Next window will allow you to create a desktop icon of PyCharm and you can also update the path variable. To proceed click next.

#### Step: 5

Here you can change the start menu folder, then click “Install”.

#### Step: 6

The installation will start. It takes a few minutes. As the installation is done click next.

#### Step: 7

In the next window click “Finish” to complete the installation process.

## Starting PyCharm for the first time

Now PyCharm is installed on your computer. Go to your computer start menu and launch the programme. The first window appears is of the privacy policy. Click on to agree with the terms & conditions and click continue.

Next is the data sharing window. It’s completely your choice. Choose any of the options and proceed.

In the next window, you will get to choose the appearance of your IDE. Choose any of them you feel comfortable with and click skip remaining and set defaults. You can change all these options anytime you want later.

The next window is important which allows you to choose the location where you want to create your Python project. For me, I like to save all my important files at the cloud, so I have provided that particular path there. You can change it here or later.

So now you are all set to start your journey with Python programming with PyCharm IDE.

## References:

• https://www.python.org
• https://towardsdatascience.com
• https://www.geeksforgeeks.org
• https://steelkiwi.com/blog

## Unsupervised Machine Learning: a detailed discussion

Unsupervised Machine Learning is a kind of Machine Learning where the algorithm identifies some hidden pattern in the data on its own. This type of Machine Learning is used when there is no labeled data available to train the algorithm.

Unlike Supervised Machine Learning here the input dataset is not tagged with some known answers. This is because in many cases we need to predict such situations which are completely new. The experimenter has no experience about the data in hand, its distribution and parameters are also unknown.

So, in this case, the application of Supervised Learning is not feasible. So we have to go for Unsupervised Machine Learning. The main problem with this approach is that we have no test dataset labeled with the correct answer to check the accuracy of such an unsupervised learning process. That’s why it has lesser accuracy than supervised learning.

## Learning process of a baby

Application of unsupervised learning resembles the learning process of babies. They start learning process themselves at the first. No one teaches them. They start identifying objects from their experience.

For example, since birth they see human and no one teach them about characteristics of it. But whenever the baby sees a new human around he matches the characteristics and recognizes the new object as a human being. This is a very basic example of unsupervised learning.

## Application of Unsupervised Machine Learning

Although this approach has a problem of lesser accuracy, it is useful to find out hidden pattern in the data.

### Speech recognition

You might have used google’s speech recognition tool. It is such a handy tool to convert your speech into text. When you have to write a lot of text, you can certainly use it to your advantage. I also use it frequently during writing my articles in Google doc.

So the point is the technology used for this handy tool for speech recognition is nothing but unsupervised machine learning. The annotation process from voice to text is very costly so, labeled data is not available to train the algorithm.

### Detection of anomaly

Unsupervised classification can also come handy to detect extreme values in the dataset. Such data generally comprises outliers which are erroneous observation due to mechanical error or error during data collection, fraudulent transaction data in bank transaction statement likewise.

### Clustering of data

Clustering is a grouping of data on the basis of some uniformity. It reveals the data structure and helps to design the classifier.

### Finds hidden patterns and feature of the data

Unsupervised learning finds out all kinds of hidden pattern and features of the which consequently helps in categorization.

## Issues with unsupervised machine learning

• The process has some inherent issues which you must consider before its application.
• Unsupervised learning results are less accurate compare to that of supervised learning and it is very obvious too.
• Performing unsupervised learning is much more complicated than a supervised one.
• Validation of the model is not possible due to lack of labeled data.

## Types of unsupervised machine learning

Unsupervised machine learning can be further grouped into two broad categories which are clustering and association problems.

### Clustering

It is of great importance when we discuss unsupervised learning. This technique finds out some similarity in the uncategorized data and groups them to create different clusters. This clustering process is hugely beneficial to gather some basic information about the data in hand. For finding patterns and features of the dataset which is otherwise completely unknown to the researchers

We can decide how many clusters we should create. The clusters are so formed so that the within-cluster variance is lower compare to between cluster variance. In similarity measure it can be phrased as the members of a cluster are similar whereas members of different clusters are dissimilar.

We perform this clustering through several approaches.

### Hierarchical clustering

Here every data point is considered an individual cluster to start with. Then in similarity basis, the most similar data points are clubbed to form a single cluster. This process continues until the decided number of clusters is achieved.

#### Probabilistic clustering

Here as the name suggests, we do the clustering on the basis of a probability distribution. For example, if there are keywords like

“Boys’ school”

“Girls’ school”

“Girls’ college”

“Boys’ college”

Then the clusters can form two categories either “boy” and “girl” or “school” and “college”

#### Exclusive clustering

If data points are such that they are very exclusive to a particular category. Then in a straight manner, we form the clusters according to data points exclusivity. Here no single data point can belong to more than one clusters.

#### Overlapping clustering

In contrast to exclusive clusters in overlapping clustering, one particular data point can belong to more than one clusters. To achieve such clustering, we use fuzzy sets.

### Clustering algorithms

There are some popular algorithms to perform clustering. In this article, I will briefly discuss them. Each of them will have an elaborate discussion in separate articles.

#### K-means

K-means clustering is a type of clustering where data points are grouped into k clusters. If the value of k is large them the cluster size is small and if k has small value then cluster size is bigger.

Every cluster has a value called the centroid. This is kind of the heart of the cluster. The distance of other data points from this centroid determines if they qualify for the cluster or not.

#### K- Nearest Neighbors

It is a simple algorithm and performs well when there is a significant distance between the sample data points. It is the most simple classification method under unsupervised machine learning but takes considerable time when the dataset is large.

#### Principal Component analysis

It is a variable reduction technique. The basic objective of PCA is to calculate fewer number of new variables maintaining the variance of the data as explained by the original variables.

#### Hierarchical clustering

This is a hierarchical clustering technique. Hierarchical in the sense that it starts with considering each data points as a cluster and then goes on forming clusters by including close clusters. This process continues until only one cluster remains.

#### Fuzzy K-means

This is a more generalized form of K-means clustering. Here also clusters are formed using a centroid value. But the difference is that in simple K-means clustering, the data points are either same as the centroid or it is different, there is no in-between position; whereas in fuzzy k-means clustering algorithm assigns a probability to each data points depending on its distance from the centroid. K-means clustering simply a special case of fuzzy K-means clustering where the probability is either 1 or 0.

### Association

This also about pattern or feature identification from large database. Unsupervised machine learning uses this association rules to find out the interesting relationship between variables. For example, students in a class can be a subject of this association rule based on their choice of subject.

## Summary

So, we  can summarize some important points about unsupervised machine learning which are as follows:

Unsupervised machine learning is the type of machine learning where we don’t use any lebeled data.

No labeled data, so no supervision of the result and no validation

It has less accuracy compare to that of supervised machine learning

Unsupervised learning is more complicated than supervised learning

Unsupervised learning proves helpful when we have no idea about the data, its distribution and parameters are also unknown.

Two main methods of conducting unsupervised machine learning are clustering and association.

## References:

• https://towardsdatascience.com
• https://www.guru99.com
• https://www.geeksforgeeks.org

## Supervised Machine Learning: a beginner’s guide

The most common type of Machine Learning is Supervised Machine Learning. The nomenclature is due to the fact that the learning process being supervised by the result which is already known. The learning process goes through several iterations. The process continues until the difference between the actual and estimated result comes under an acceptable level.

“Computers are able to see, hear and learn. Welcome to the future.”

~Dave Waters. Department of Earth Sciences, University of Oxford Associate Professor of Metamorphic Petrology (retired)

The data used in supervised machine learning are called “labelled data” because these data are already tagged with the right answer. Once the training part is complete and a robust model is achieved, some new inputs are provided. The task of the model now is to predict the label of this unforeseen inputs based on the labelled data used before.

In mathematical notation, it can be represented as the output variable Y which is a function of input variable X

Y=f(X)

During the training phase of supervised machine learning both X and Y remains unknown. The algorithm tries to find out the mapping function which can predict the Y most precisely.

## Example of Supervised machine learning

You must have come across the term pattern recognition from any online or offline source. This is a kind off buzz word today and is in use to make our life more sophisticated and comfortable. Starting from a very simple application like your smartphone’s face recognition or handwriting recognition to advance use of cancer cell detection, this supervised learning is the essence of pattern recognition.

Its simple applications are already making our lives easier be it your smartphone’s face lock feature, handwriting recognition or your voice recognition. The auto-driving car concept also heavily depends on supervised learning concept. In every sector of the industry, you can find presence of this theory nowadays.

### An application in agriculture

Now to understand how this system works we will take an example of its application in the agriculture field.

Prediction for the crop yield well before its harvesting is very essential for proper policy planning. It helps the government to fix its price, to provide better storage of the produce and farmers also able to plan its marketing channels if there is a precise prediction about how much production is expected.

Now crop yield is determined by several factors, some of them are physical parameters of the crop itself like crop height, number of tillers etc. weather parameters like rainfall, humidity, sunshine hours etc. other than these soil health factors like carbon balance, organic matters and several others play an important role and contribute to the ultimate yield.

Now if we have a sufficient amount of labelled data that is a set of data which has all these independent variables affecting the yield along with the corresponding yield, we can train the algorithm with this training dataset. So, it will be supervised learning. As if the learning process has been supervised by any teacher.

The learning process stops only when a robust model is achieved and the prediction is of an acceptable level.

## A real-world problem solved by Supervised Machine learning

Here I am going to cite an example of supervised learning in modern research and how it is being used to address complex problems of the real world.

A Project work was taken up by a group of scientists to identify the endangered species of Mojave desert of California. The main objective of the study was to locate the two threatened species Mohave Ground Squirrel and desert tortoise of the area by analyzing images captured by smartphones.

The challenge faced by the biologists was to track and rescue these two endangered species as they were very tough to spot. Nature has given them such a capability to camouflage with the desert background and vegetation that it becomes almost impossible for the human eye to see them.

So here the scientists used computer vision and develop a machine learning algorithm to identify the pattern, distinguish it from the desert backdrop and classify them according to the characteristics.

## Types of supervised machine learning

There are two main categories of supervised machine learning.

• Classification
• Regression

### Classification:

It is applicable when the variable in hand is a categorical variable and the objective is to classify it. If the algorithm classifies into two classes, it is called binary classification and if the number of classes is more than two, then it is called multiclass classification.

In the given figure, a binary classification has been demonstrated. Here a group of people has been classified according to their genders depending on a dataset consisting their height and weight.

The task is done in the same way as discussed before. First of all, the algorithm is trained with a dataset with an assigned category. Then based on this training the algorithm has categorized the values when provided with an input data.

#### Example of classification

A most common example of classification problem is identifying if a new mail is a spam or not spam, identifying loan defaulters also a problem of classification.

The algorithm is provided with a dataset of mails and a corresponding column indicating if it is a spam or not spam. Similarly, a list is first provided with the customers labelled with if they are a loan defaulter or not to train the algorithm. Then the supervised learning model is used to identify the type of customer from an independent input dataset.

There are a number of algorithms for classification. The most popular ones are

• Naive Baye’s theorem
• Linear classifier
• Support vector machine
• Random forest
• Decision tree
• K-Nearest neighbour

### Regression

Regression is a statistical process which tries to find out the relationship between the dependent and independent variables. The major difference with classification is that in regression we deal with continuous variables.

If a regression equation is a linear one between the independent and dependent variables then it is a simple linear regression equation. If the regression equation of Y on X is linear, then it does not necessarily suggest that the regression equation of X on Y is also linear and vice-versa. The dependent variable a function of independent variables with respective constant parameters and an error term which is again a random variable. A regression model has the expression:

Y=f 0,1,2,…, n+ϵ

Where Y is the dependent variable, X1, X2+…Xn are independent variables, 0,1,2,…, n are the regression coefficients and is the error term and normally distributed with mean 0 and variance 2.  This type of regression model is also known as a deterministic model.

#### Example of regression

An example of simple linear regression can be regressing the weight of a group of people on the basis of their height. Here Height and weight are the independent and dependent variable respectively. As a person height determines his weight, not the vice versa.

The blue line in the above figure is the regression line fitted with a supervised machine learning technique. This represents the best-fitted line obtained through a rigorous training process until a robust model with acceptable accuracy is achieved.

To perform regression a number of algorithms are used by researchers. The most frequently used ones are:

• Simple linear regression
• Multiple linear regression
• Logistic regression
• Polynomial regression etc.

## Machine Learning: Some lesser known facts

Machine Learning (ML) has become a buzz word in today’s world. Although we can have its references since the middle of the twentieth century it has gained its popularity during the last few years. Mainly because of its immense capability to explore a large amount of data without the need for any programming and hence the simplicity to use.

Since Machine learning is still a new concept and there are several doubts and misconception about it. In this article, I will try to explore some of these facts that are less known about Machine Learning along with very basic ideas like what is Machine Learning and how it is making our lives better.

Let’s start with a famous conversation of an interview to hire a Machine Learning expert. You must have read this before but I like this so much and it can give a good start to this article.

So as the interview starts, the interviewer starts asking questions to the candidate:

Candidate: Machine Learning

Interviewer: What is 23+34?

Candidate: It’s 10

Interviewer: No, wrong answer, its 57

Candidate: It’s 35

Interviewer: No, wrong answer again, it’s 57

Candidate: It’s 50

Interviewer: No, the answer is still 57

Candidate: It’s 57

Interviewer: You are hired !!!

Although it is a joke, to some extent it reflects how the Machine Learning works. Machine Learning is all about learning from the data it is fed with. Here is a famous quote from Thomas H. Davenport, Analytics thought-leader from the Wall Street Journal which reflects the power of Machine Learning;

“Human can create one or two good models a week; Machine Learning can create thousands of good models a week”

Thomas H. Davenport, Analytics thought-leader from the Wall Street Journal

## Importance of Machine Learning in the present context

Today we have a huge amount of data popularly known as big data. This can be a gold mine of knowledge if used and explored properly. Data mining, Baysian analysis all these are getting popular only because they also cater to extract information from a big pile of data.

As the volume of data increased, so its complexity. The data comes from varieties of sources, consists of numerous fields. We need modelling techniques which can analyze such kind of data quickly with improved accuracy. So here is Machine Learning for you.

## So, what is Machine Learning ?

Machine learning in simple term is converting knowledge from information. We have a huge amount of data in our custody, generated throughout a period over more than 50 years. If it is not used to generate knowledge out of it then this huge volume of data is of no use and we are just scrapping a very valuable resource that can help solve many challenges of humanity.

It is as such a very vast field of data science and assimilates many concepts of other associated fields like Artificial Intelligence.

The beauty of Machine Learning is that it does not need programming by human rather as the name suggests it learns from the data it was fed. In this sense, it is similar to a human who also learns from their past experiences.

This learning comes through a rigorous process of observing the data, finding out the pattern in order to minimize the difference between actual and estimation.

Machine Learning has three main categories, which are

## Application of Machine Learning?

Recent advances made in Machine Learning enables computer some of the tasks which can only be handled by human until very recent time. In our daily life, we take help or use applications which use this technique and most of the time we don’t even know that it is Machine Learning which is making our lives easier.

### In daily life

We can take a simple example of getting personalised Google news. This application which type of news you are interested in by keeping an account of your likes and dislikes as you time to time input in Google’s database. The same technique is used by Facebook to suggest you groups or pages that you may like. Ever wonder how your email service identifies spam emails for you and discriminates from important mails, thanks to ML.

Online video streaming services like Netflix, Amazon Prime, Hotstar etc. or music streaming applications like Spotify all of them have a nice feature which automatically populates your account with contents you prefer. Here the essence is Machine Learning; it analyzes your popular choices and suggests content according to your choice.

### Image/speech recognition & medical research

Image recognition uses this technology to answer whether an animal is a cat or dog, identifying persons crossing the road, identifying your handwriting and converting into texts and many more.

In a similar way converting voice into text which is predominantly in use in several platforms like speech to text tool in Google doc and here also ML plays an important role.

In medical research, ML is a fast-growing technology. It helps in analyzing voluminous data and to identify trends and patterns.  Especially with the advent of wearable devices and sensors which keep track of vital parameters of patient’s health. The data generated by these devices are analyzed through ML often in real-time to enable medical practitioners to detect any trend and red flag any symptom for better diagnosis.

### Oil and gas sector

In this sector, ML finds its use to identify natural resources like minerals under the ground, pointing out any risk involved in the performance of the refinery sensors and chance of failure, also preparing an optimized oil distribution plan to make it more cost-effective and efficient.

Thus almost in every sector of our society, the use of Machine Learning is rapidly expanding. In absence of Machine Learning, performing such a resource-intensive and time-consuming process would not be at all feasible in traditional ways.

### Futuristic applications

Few applications of Machine Learning which are still in the testing phase, are always been the popular topics of science fiction stories. We are now frequently hearing and reading about self-driving cars of Google or Tesla. This is already a reality now, but go back 10 years, such a concept used to be a subject of science fiction only. The basic concept behind this revolutionary invention is Machine Learning.

Almost every industry who deal with a large amount of data has realized the importance of Machine Learning. Be it banking and finance sector, automobile, research or health care sector ML enables them to work more efficiently and have an edge over their competitors with the help of data insights often in real-time.

## So, what is Artificial Intelligence (AI) then?

If you have read up to this, then this question is most probably rising in your mind and it is bound to. Although most of the times we use the terms AI and Machine Learning interchangeably they are not the same. AI makes machines to emulate human intelligence whereas ML helps machines to learn from data.

Artificial Neural Network (ANN) as its name suggests it mimics the neural network of our brain hence it is artificial. The human brain has a highly complicated network of nerve cells to carry the sensation to its designated section of the brain. The nerve cell or neurons form a network and transfer the sensation one to another. Similarly in ANN also a number of inputs pass through several layers similar to neurons and ultimately produce an estimation.

Machine Learning is a way to implement Artificial Intelligence. Machine Learning has been in application since decades but in recent days as Artificial Intelligence came into action Machine Learning, to be more specific Deep Learning has become more popular.

## ANN: a deep learning process

ANN is a deep learning process, the burning topic of data science. Deep learning is basically a subfield of Machine Learning. You may be familiar to the machine learning process and if not you can refer to this article for a quick working knowledge on it. Talking about deep learning, it is in recent times find its application in almost all ambitious projects. Starting from basic pattern recognition, voice recognition to face recognition, self-driving car, high-end projects in robotics and artificial intelligence deep learning is revolutionizing the modern applied science.

ANN is a very efficient and popular process of pattern recognition. But the process involves complex computations and several iterations. The advent of high-end computing devices and machine learning technologies have made our task much easier than ever. Users and researchers can now focus only on their research problem without taking the pain of implementing a complex ANN algorithm.

The concept of Artificial Intelligence although not very new, it was first used in 1950 and was supposed to use a computer to perform such activities which can only be done by human beings only.

So in that sense, AI is a much broader concept and ML can be considered as a subset of it. AI is as a whole mimics the concept of human intelligence and to achieve it ML plays a very important role by extracting information from data without the need for programming.

### Machine Learning Vs Deep Learning Vs Data Mining

Often these three concepts are little confusing and the main reason is all these techniques have the same goal, which is to get an insight, relationship or trend of the data in hand. But they differ in their execution and abilities.

#### Machine Learning

As we discussed, Machine Learning functions more like statistical models, where there is a mathematically proven strong theory about the distribution of the data and it is assumed that the data fulfil some assumptions too. The advantage of Machine Learning is that even if we do not have any theoretical idea about the distribution of the data it can learn from the data through several iterations until the best pattern is found. Hence, the process of ML can be easily automated too.

#### Data Mining

It is a much broader concept with the same objective as ML and encompassed a variety of concepts to achieve that. Like deep learning uses traditional statistical theories, text analytics, time series algorithm, data manipulation techniques and even Machine Learning too in order to identify an underlying pattern in the data.

#### Deep Learning

It is a more advance concept compare to the above two. Deep learning involves the state of the art technologies combining modern high-end computing and neural networks to identify complex patterns in a large amount of data. Advance technologies like image recognition, recognizing words from the sound which are still in the testing stage are all subject of deep learning.

## Some facts on Machine Learning

At the very beginning I have mentioned that being a new concept, some ideas about Machine Learning are also popular but not completely true. Here I will try to discuss all those lesser-known facts about ML.

### Fact 1: It is not complete automated process and human intervention is required

There is a misconception that ML is a 100% automated process, which is not completely true and human intervention is necessary to create and improve algorithms. The system needs context and parameters to operate which again provided by human operators.

### Fact 2: Having advance knowledge in Mathematics is not a prerequisite for simple application of Machine Learning

You can start the application of ML to analyze your data with some practice and guidance. There are lots of content available on the internet some of them are free whereas few are premium courses.

To start practising with ML you can choose any of the free courses. The main factor is you have to practice a lot. I can suggest you a free crash course on ML by Google Developers, developed by Google, so no question about the quality.

The MOOC’s course on ML in Coursera is also very good to start your learning session.

### Fact 3: Machine Learning and Artificial Intelligence are not the same

Some people have this notion that these two are same, even I used to have the same idea until I came across this article published in Forbes. It was a very good comprehensive discussion about the differences between these two, read it you will get your many doubts about ML and AI cleared.

### Fact 4: Even without a very sound knowledge of programming language you can learn the application of ML

Oh… it certainly helps, having good knowledge in a few programming languages can help you jump start your carrier in ML, but it is not at all an essential one. Its just you have to give some little more time when you are first time writing your code for Machine Learning. Be it R or Python or any other language, you learn it by making errors, this is the most effective way of learning any language.

So, in nutshell, if you are interested in learning ML, just start it now, take a small dataset, write a small piece of code. There will be errors in the beginning, don’t let it hold you back. Soon you will start enjoying its beauty and it will get more and more interesting.