View on GitHub
# https://en.wikipedia.org/wiki/Gradient_descent
import numpy as np
import random

# 1. calculate hypothesis h = X * theta
# 2. calculate loss = h - y, squared cost = (loss ** 2) / (2 * m)
# 3. calculate gradient = X' * loss / m
# 4. update parameters thete = theta - alpha * gradient
#
# m: number of examples

def gradient_descent(x, y, theta, alpha, m, num_iterations):
    x_transpose = x.transpose()
    
    for i in range(0, num_iterations):
        hypothesis = np.dot(x, theta)
        loss = hypothesis - y
        cost = np.sum(loss ** 2) / (2 * m)

        # print
        print("Iteration: %d, Cost: %f" % (i, cost))
        
        # average gradient per example
        gradient = np.dot(x_transpose, loss) / m
        
        # update
        theta = theta - alpha * gradient
    
    return theta

def generate_data(num_points, bias, variance):
    x = np.zeros(shape=(num_points, 2))
    y = np.zeros(shape=num_points)
    
    # straight line
    for i in range(0, num_points):
        # bias feature
        x[i][0] = 1
        x[i][1] = 1
        
        # target variable
        y[i] = (i + bias) + random.uniform(0, 1) * variance
        
    return x, y

# generate points with bias, variance (noise)
x, y = generate_data(100, 25, 10)
m, n = np.shape(x)
num_iterations = 10000
alpha = 0.0005
theta = np.ones(n)
theta = gradient_descent(x, y, theta, alpha, m, num_iterations)

# print
print(theta)

# Iteration: 0, Cost: 3440.881818
# Iteration: 1, Cost: 3434.861831
# .
# .
# .
# Iteration: 9998, Cost: 429.382775
# Iteration: 9999, Cost: 429.382775
# [39.80223561 39.80223561]
      

Updated on May 27, 2023 Changelog