Machine Learning Introduction
-
Upload
nattaq12345 -
Category
Documents
-
view
2 -
download
0
description
Transcript of Machine Learning Introduction
Machine Learning: Lesson 1
Supervised Learning -> A set of training data (data that is given to be correct) is given and a program makes a prediction (extrapolation/interpolation). The prediction is a regression.
Unsupervised Learning -> A given data set consists of undetermined information. The goal is to create a program that will find a structure amongst the given data set. Structure can be a cluster. Clusters can be studied and programs can gather information about them that will allow data scientists to explore further.
Hypothesis Function -> $h_{\theta}(x)= \theta_0 + \theta_1\cdot x$ where h is the hypothesis function, $x$ is the input value for a given data set, and $\theta_0$ and $\theta_1$ are parameters that help compute the cost function. The hypothesis function describes a set of data using a regression so that supervised learning can be applied to such a set. Depending on the values substituted for the parameters, a minimum cost function can be calculated to find the best regression to fit any data set.
Cost Function -> is a univariate (or multivariate) function that takes a hypothesis function and a data set, returning a value that describes how well such a hypothesis function fit a given data set. Minimizing the cost function is a common task endowed upon data scientists so that supervised learning can be applied to a given data set.
Cost Function Formula -> $J_(\theta_1,\theta_2) = \frac{1}{2m}\sum_{i=0}^{N}(h_{\theta}(x^(i)) - y^(i))^2$
Where $m$ is the number of training data in a given data set and $i$ is the ith datum in a set of training data.
Gradient Descent Algorithm -> An algorithm that runs through a contour plot of a cost function with two or more parameters ($\left\{\theta_0,\theta_1,\theta_2,\dotsc,\theta_n\right\}$ and traverses the plot to find the local minimum. Depending on where the algorithm sets its starting point, a different minimum might be found.
Gradient Descent Algorithm Implementation -> /**Convergence is of type value returned by Cost Function, which ought to be the lowest point on a contour plot. If the current point is the smallest nearby, i.e. Cost Function is minimized, then end loop return the point of convergence.**/while(!convergence){for(int j = 1, j < n; j++){$\theta_j = \theta_j + \alpha \frac{\partial}{\partial \theta_j} J(\theta_0,\theta_1)$/**where $\alpha$ is the learning rate and $\frac{\partial}{\partial \theta_j}$ is the change in the position along the contour plot**/}}