# R AND PYTHON: GRADIENT DESCENT

My previous blog, I presented syntactical differences and similar functions between R & Python. Now, I want to take it to next level and write some machine learning algorithms using both R and Python. Here, one may use direct functions from the packages available. However, here I am presenting the way to write your own functions for algorithms. In this series, I am starting with Gradient descent algorithm. I briefly explain, what is gradient descent. After that, I apply gradient descent algorithm for a linear regression to identify parameters. For illustration, I simulate data for simple linear regression.

I show how you can write your own functions for simple linear regression using gradient descent in both R and Python. I present the results of both R and Python codes.

# Gradient descent algorithm

Gradient descent is one of the popular optimization algorithms. We use it to minimize the first order derivative of the cost function. You can define a cost function using an equation (example: f(x) = 3x2 + 2x+3) , then do the first order derivative for it ( in this case, f `(x)= 6x+2). Then you can randomly initialize value of x and iterate the value with some learning rate (slope), until we reach convergence (we reach local minimum point). For more details, you can refer simple computational example, with python code, given in Wikipedia .

I am taking the same concept of Gradient Descent and applying it for simple linear regression. In the above example, we have the equation to represent the line and we are finding the minimum value. However, in the case of linear regression, we have data points, and we come up with the linear equation, that fits these data points. Here, it works by efficiently searching the parameter space to identify optimal parameters i.e,or . intercept(theta0) and slope(theta1). We initialize the algorithm with some theta values and iterate the algorithm to update theta values simultaneously until we reach convergence. These final theta values (parameters or coefficients), when we put it in equation format, represent best regression line.

In machine learning, the regression function y = theta0 + theta1x is referred to as a 'hypothesis' function. The idea, just like in Least squares method, here is to minimize the sum of squared errors, represented as a 'cost function'. We minimize cost function to achieve the best regression line.

Let us start with simulating data for simple linear regression.

### R:

Importing the dummy data we created.

setwd("S:\\ANALYTICS\\R_Programming\\Gradient_Descent")

N x beta errors # errors errors y

data

colnames(data)

write.csv(data, file = "Data.csv", row.names = FALSE)

Below we are considering the above simulated data for the python coding purpose.

#Data simulation for python

colnames(data)

write.csv(data, file = "Data_Python.csv", row.names = FALSE)

Looking at the dimension of data this is because we are working on matrix method.

Now we change the dimension of x(predictor) and y(response) accordingly to make matrix multiplication possible.

create the x- matrix of explanatory variables. Here we are producing a matrix with 1's in 100 rows to make the matrix dimension is such a way that the multiplication is possible.

Data

head(Data)

## x y

## 1 55.01790 26.34163

## 2 63.62490 36.61669

## 3 56.07122 26.69499

## 4 61.00054 39.82381

## 5 63.45294 36.48111

## 6 62.82037 32.28978

dim(Data)

## [1] 100 2

head(Data)

## x y

## 1 55.01790 26.34163

## 2 63.62490 36.61669

## 3 56.07122 26.69499

## 4 61.00054 39.82381

## 5 63.45294 36.48111

## 6 62.82037 32.28978

attach(Data)

#n

a0 = rep(1,100)

a1 = Data[,1]

b = Data[,2]

x y

head(x)

## a0 a1

## [1,] 1 55.01790

## [2,] 1 63.62490

## [3,] 1 56.07122

## [4,] 1 61.00054

## [5,] 1 63.45294

## [6,] 1 62.82037

head(y)

## [,1]

## [1,] 26.34163

## [2,] 36.61669

## [3,] 26.69499

## [4,] 39.82381

## [5,] 36.48111

## [6,] 32.28978

# Derivative Gradient:

Derivative gradient is the next step after simulating the data in which we obtain the initial parameters. For initializing the parameter of hypothesis function, we considered the first order derivative of the cost function in the below function.

where m is the total number of training examples and h(x(i)) is the hypothesis function defined like this:

Hypothesis Equation

For determining the initial gradient values, we have to consider the first derivative of the cost function which shown below. Cost function J(theta0, theta1)

Cost Function

Initial_Gradient m hypothesis = x %*% t(theta)

loss = hypothesis - y

x_transpose = t(x)

gradient return(t(gradient))

}

Here 'i' represents each observation from the data.

For the derivative gradient, we are determining the first derivative of the cost function which gives initial coefficients to update through the gradient runner below. This derivative gradient is called by the gradient runner inside the for loop.

From the above function results we will get the matrix of thetas (a0 and a1). Now using the gradient runner, we have to iterate through the theta0 and theta1 until convergence.

# Gradient runner:

Gradient runner is the iterative process to update the theta values through the below function until it converges to its local minimum of the cost function.

Here in this function additional element we are using is learning rate(alpha) which help to converge to local minimum. For our requirement, we considered alpha as 0.0001.

Gradient descent is used to minimize the cost function to get the converged theta value matrix which fits the regression line in data.

gradient_descent # Initialize the coefficients as matrix

theta #theta #learning rate

alpha = 0.0001

for (i in 1:Numiterrations) {

theta }

return(theta)

}

print(gradient_descent(x, y, 1000))

## a0 a1

## [1,] 0.03548202 0.5777683

To speed up the convergence change learning rate or number of iterations. If you consider learning rate as too small it will take more time to converge and if it is high also it will never converge.

## Python:

Import all the required packages and the dummy data to run the algorithm. For python, also we are doing it in matrix method.

Same steps to be followed in python as we did using R.

from numpy import *

import numpy as np

import random

from sklearn.datasets.samples_generator import make_regression

import pylab

from scipy import stats

import pandas as pd

import os

path = "S:\\ANALYTICS\\R_Programming\\Gradient_Descent"

os.chdir(path)

points = genfromtxt("Data_Python.csv", delimiter=",", )

a= points[:,0]

b= points[:,1]

x= np.reshape(a,(len(a),1))

x = np.c_[ np.ones(len(a)), x] # insert column

y = b

#For determining the initial gradient values, we have to consider the first derivative of the cost function which shown below.

Cost function

def Initial_Gradient(x, y, theta):

m = x.shape[0]

hypothesis = np.dot(x, theta)

loss = hypothesis - y

x_transpose = x.transpose()

gradient = (1/m) * np.dot(x_transpose, loss)

return(gradient)

#Below gradient descent function is to update the theta0 and theta1 values(coefficients of hypothesis equation) with the initial values untill they converge. We are calling the initial theta value matrix inside the for loop.

def gradient_descent(alpha, x, y, numIterations):

m = x.shape[0] # number of samples

theta = np.zeros(2)

x_transpose = x.transpose()

for iter in range(0, numIterations):

hypothesis = np.dot(x, theta)

loss = hypothesis - y

J = np.sum(loss ** 2) / (2 * m) # cost

theta = theta - alpha * Initial_Gradient(x, y, theta) # update

return theta

#Now to verify the algorithm in R and Python we are providing learning rate and number of iteration to gradient descent to perform.

print(gradient_descent(0.0001,x,y,1000))

## [ 0.03548202 0.57776832]

If you observe from the both algorithms we can see they given the same theta values(coefficients) for hypothesis.

While understanding, and applying gradient descent algorithm. Hope you can see the syntactical differences to arrive at same results from R and Python. In this blog, we saw writing functions, handling matrices, loops, and several other things in R and Python. See you next time!**About Rang Technologies:**

Headquartered in New Jersey, Rang Technologies has dedicated over a decade delivering innovative solutions and best talent to help businesses get the most out of the latest technologies in their digital transformation journey. **Read More...**