[coursera machine learning] Week 1

1. machine learning 问题的分类:

Supervised Learning: right answers given in samples

           Regression: continuous result

           Classification: discrete valued output

Unsupervised Learning: learning about a dataset without correct answers

            Clustering: divide dataset into groups

            Non-clustering: separate different voices from a voice sample (cocktail party)

 

2. Model Representation:

training set -> learning algorithms -> hypothesis 

x -> hypothesis -> y

3. Cost Function:

m is the number of samples

4. Gradient Descent (not only for linear regression)

n is the number of features

minimization a function (ect. cost function)

the alpha is learning rate

all theta should be updated simultaneously.

5. Normal Equation Formula

comparison of gradient descent and normal equation formula.

normal equation is faster with less features.

gradient descent is faster with more features.

 

posted @ 2017-04-10 10:58  Gryffin  阅读(127)  评论(0编辑  收藏  举报