Lecture Notes of 23/09/20


Supervised Learning

Supervised Machine Learning is where we give input(X) and output(Y) variables to an algorithm which approximates a mapping function between these from the training data so that when we give a new input data, it can predict the output variables for that data.

The Data in supervised machine learning consists of :

  1. Input : It could inlcude numbers, pictures, emails, measurements etc.
  2. Output : These could be numbers, labels (like spam/not-spam), vectors(like coordinates of a bounding box), sequences etc.

Now let us look at an example where supervised machine learning is used. One of the popular areas it is used is for spam detection in emails. It is used to categorize emails as either spam or not spam.


Note :

The above optimization problem works only if the data is linearly separable. If not, there would be no suitable set of parameters w and b which can simultaneously satisfy all the constraints. In such cases, people will either work with a slightly different model known as soft-margin SVM or move on to non-linear hyperplanes depending on which is suitable for the problem.

Let's summarise the steps involved in Supervised Machine Learning :

  1. Getting the Data.
  2. Converting into feature vectors.
  3. Generating the Model.
  4. Predicting the output for a new input data.

Fig. Supervised ML model (Image Credits: The NVIDIA Blog)


The type of decision boundary we want will depend on the following factors: For solving our problem using ML, the algorithms are choosen based on: A suitable algorithm is choosen by taking care of all the above parameters and keeping a balance between all of them.


References

  1. Support-Vector Machine.
  2. A Course in Machine Learning by Hal Daume .