Difference between revisions of "Machine Learning: Linear Regression"
Aplstudent (talk | contribs) |
Aplstudent (talk | contribs) |
||
(2 intermediate revisions by the same user not shown) | |||
Line 22: | Line 22: | ||
[[File:Scatterlinearpost.png]] | [[File:Scatterlinearpost.png]] | ||
− | The slope and intercept must be initialized by the user. In addition the learning rate is how quickly to change the parameters by each iteration. the data set is | + | The slope and intercept must be initialized by the user. In addition the learning rate is how quickly to change the parameters by each iteration,this is called alpha in the code. the data set is fit multiple times with updated starting weights, these are epochs. |
Line 42: | Line 42: | ||
</code> | </code> | ||
− | each pair is looped through and a | + | each pair is looped through and a output is predicted according to <math>Ypred=X*m+b</math> The error is then found according to <math>Ypred-Y</math> The weights are then updated according to the formula <math>m:=m-learningrate*error*x</math> and <math>b:=b-learningrate*error</math> The new value of the paramaters are used with the next set of points. |
<code> | <code> | ||
Line 61: | Line 61: | ||
the slope is found to be 1.04 and the y-intercept as -0.08. Imperfect data was used so the exact value was not found. The choice of alpha is important because a value that is too large will give an erroneous result. The algorithm will overshoot the exact value and then update it by overshooting it again resulting in the answer revolving around the exact value. | the slope is found to be 1.04 and the y-intercept as -0.08. Imperfect data was used so the exact value was not found. The choice of alpha is important because a value that is too large will give an erroneous result. The algorithm will overshoot the exact value and then update it by overshooting it again resulting in the answer revolving around the exact value. | ||
− | This is a graph of the final parameters: | + | This is a graph of the final parameters,an appropriate answer was found: |
[[File:Linearregression2tut.png]] | [[File:Linearregression2tut.png]] | ||
+ | |||
+ | |||
+ | |||
+ | |||
+ | sources: | ||
+ | Artificial Intelligence : A modern Aproach by Stuart Russel and Peter Norvig, chapter 19. | ||
+ | http://machinelearningmastery.com/linear-regression-tutorial-using-gradient-descent-for-machine-learning/ |
Latest revision as of 10:09, 12 November 2016
Machine learning is the subfield of Artificial intelligence interested in making agents that review their past actions and experiences to formulate a new action. The task is to learn the formula for a function. A performance standard is used to adjust the function that is predicted. Linear regression is a common pattern that can be approximated by learning the parameters for slope and intercept.
Here are the packages that are used and three points to approximate a function from.
import numpy as np
import matplotlib.pyplot as plt
pair=np.array([(1,1),(4,4.2),(2,2)])
plt.scatter([pair[0][0],pair[1][0],pair[2][0]],[pair[0][1],pair[1][1],pair[2][1]])
The slope and intercept must be initialized by the user. In addition the learning rate is how quickly to change the parameters by each iteration,this is called alpha in the code. the data set is fit multiple times with updated starting weights, these are epochs.
nepoch = 4
m=0.0
b=0.0
alpha = 0.2
each pair is looped through and a output is predicted according to The error is then found according to The weights are then updated according to the formula and The new value of the paramaters are used with the next set of points.
for i in range(0,nepoch,1):
print("i=" , i)
for i in [0,1,2]:
print(m,b)
x=pair[i][0]
y=pair[i][1]
y_pred=m*x+b
error=y_pred-y
m=m-alpha*error*x
b=b-alpha*error
print(m,b)
the slope is found to be 1.04 and the y-intercept as -0.08. Imperfect data was used so the exact value was not found. The choice of alpha is important because a value that is too large will give an erroneous result. The algorithm will overshoot the exact value and then update it by overshooting it again resulting in the answer revolving around the exact value.
This is a graph of the final parameters,an appropriate answer was found:
sources:
Artificial Intelligence : A modern Aproach by Stuart Russel and Peter Norvig, chapter 19.
http://machinelearningmastery.com/linear-regression-tutorial-using-gradient-descent-for-machine-learning/