# Lesson1.rtf

View again

All materials on our website are shared by users. If you have any questions about copyright issues, please report us to resolve them. We are always happy to assist you.
Information Report
Category:

## Documents

Published:

Views: 9 | Pages: 2

Share
Related documents
Description
Machine Learning: Lesson 1 Supervised Learning - A set of “training” data (data that is given to be “correct”) is given and a program makes a prediction (extrapolation/interpolation). The prediction is a regression. Unsupervised Learning - A given data set consists of undetermined information. The goal is to create a program that will find a structure amongst the given data set. Structure can be a cluster. Clusters can be studied and programs can gather information about them that will allow
Transcript
Machine Learning: Lesson 1Supervised Learning -> A set of “training” data (data that is given to be “correct”) is given and a program makes a prediction (extrapolation/interpolation). The prediction is aregression. Unsupervised Learning -> A given data set consists of undetermined information. The goal is to create a program that will ﬁnd a structure amongst the given data set. Structure can be a cluster. Clusters can be studied and programs can gather informationabout them that will allow data scientists to explore further.Hypothesis Function -> $h_{\theta}(x)= \theta_0 + \theta_1\cdot x$ where h is the hypothesis function, $x$ is the input value for a given data set, and $\theta_0$ and $\theta_1$ are parameters that help compute the cost function. The hypothesis function describes a set of data using a regression so that supervised learning can be applied to such a set. Depending on the values substituted for the parameters, a minimum cost function can be calculated to ﬁnd the best regression to ﬁt any data set.Cost Function -> is a univariate (or multivariate) function that takes a hypothesis functionand a data set, returning a value that describes how well such a hypothesis function ﬁt agiven data set. Minimizing the cost function is a common task endowed upon data scientists so that supervised learning can be applied to a given data set.Cost Function Formula -> $J_(\theta_1,\theta_2) = \frac{1}{2m}\sum_{i=0}^{N}(h_{\theta}(x^(i)) - y^(i))^2$ Where $m$ is the number of “training” data in a given data set and $i$ is the i’th datum in a set of “training” data. Gradient Descent Algorithm -> An algorithm that runs through a contour plot of a cost function with two or more parameters ($\left\{\theta_0,\theta_1,\theta_2,\dotsc,\theta_n\right\}$ and traverses the plot to ﬁnd the local minimum. Depending on where the algorithm sets its starting point, a diﬀerent minimum might be found. Gradient Descent Algorithm Implementation ->  /**  Convergence is of type value returned by Cost Function, which ought to be the lowest point on a contour plot. If the current point is the smallest nearby, i.e. Cost Function is minimized, then end loop return the point of convergence.**/while(!convergence){for(int j = 1, j < n; j++){$\theta_j = \theta_j + \alpha \frac{\partial}{\partial \theta_j} J(\theta_0,\theta_1)$ /**where $\alpha$ is the learning rate and $\frac{\partial}{\partial \theta_j}$ is the change in the position along the contour plot**/}}
Recommended

7 pages

11 pages

3 pages

4 pages

6 pages

12 pages

5 pages

12 pages

12 pages

12 pages

2 pages

16 pages

3 pages

3 pages

2 pages

View more...
Related Search