请注意,本文编写于 90 天前,最后修改于 86 天前,其中某些信息可能已经过时。
目录
在线资料
Linear Regression Model
Part One
Part Two
Cost Function
Cost Function inition
在线资料
中文资料:https://github.com/fengdu78/Coursera-ML-AndrewNg-Notes
在线文章:http://www.ai-start.com/ml2014/html/week1.html#header-n257
Linear Regression Model
mind
-
what, fitting a straight line to your data
-
how, build a linear regression model from this dataset
- use a dataset on house sizes and prices
- plot a graph where the horizontal axis is the size of the house in square feet, and the vertical axis is price of a house
- plot the data points
- fit a straight line to the data
-
estimate the price(how, predict the price of a house based on the size of the house)
- input the size of the house
- output price
key word
- first model
- That just means fitting a straight line(直线) to your data.
- let's start with a problem that you can address using linear regression.
- Say you want to predict the price of a house based on the size of the house.
- We're going to use a dataset on house sizes and prices from Portland
- Here we have a graph where the horizontal axis is the size of the house in square feet, and the vertical axis is the price of a house in thousands of dollars.
- Let's go ahead and plot the data points for various houses in the dataset.
- Supervised learning model
- Data has "right answers"
- Regression model
- Predicts numbers
- Regression problem
Part One
Part Two
key word
best-fit line
the linear function
why we are choosing a linear function
linear regression
linear regression with one variable = univariate linear rgression
construct a cost function
is the most universal and important ideas in machine learning and is used in both linear regression and in training many of the most advanced AI models in the world
从示例分析流程

抽象出公式

Cost Function
mind
- what
- why need
- how to formula
key process
**key word **
- parameters、coefficients、weights
- hellp you build intuition on how w and b the parameters determine f
- prediction,predictive value
- slop of the line
- line intersects
- superscript ? 上标
- y hat i
- close to the true targe
- Now the question is how do you find values for w and b so that the predict y hat i is close to the ture target y^i for many or maybe all training examples x^i y^i
- how to measure how well a line fits the training data
- to do that , we're going to construct a cost function
- The cost function takes the prediction y hat and compares it to the target y by taking y hat minus(减) y.
- This different called thd error .
- let's computes the square 平方 of this error.
- we're going to want to compute this term 项 for different training examples i in the training set .
- we want to measure the error across the entire 全部 training set.
- m is the number of training set examples
- notice that if we have more training examples m is larger and your cost function will calculate a bigger number
- To build a cost function that doesn't automatically get bigger as the training set size gets larger by convention,we will compute average squared error instead of the total squared error and we do that by dividing by m like this.
- By convention the cost function that machine learning people use actually divides by 2 times m.
- The extra division by 2 is just meant to make some of oue later calculations look neater, but the cost function still works whether you include this division by 2 or not.


Cost Function inition
key process
-
define goal
- model: you want to fit a straight line to the training data, so you have this model , fw
- parameters : the model's parameters are w, and b,
- depending on the values chosen for these parameters
- ,you get different straight lines like this.
- you want to find values for w, and b, so that the straight line fits the training data well
- cost function
- To measure how well a choice of w,and b fits the training data ,you have cost functiuon J.
- What does cost function J dose is, it measure different between the model's predictions and the actual true values for y.
- wht you see later, is that linear regression would try to find values for w, and b , then make a J of w be as small as possible.
- goal: in math, we want to minimize J
1.
-
simplified
- in order for us to better visualize the cost function J.
- getting rid of the parameter b
- It just goes away from the equation,
- so f is now just w times x
- the goal become J of w
-
f and j are related
-
f function , w is constant value , f is only a function of x.
-
j is only a function of w.
-
first pick value
- the plots of three points representing,1,1 2,2 3,3
- pick a value for w, Say w is 1. 似乎和上一步顺序颠倒,应该先有w,然后在选择f的点位
- compute f and j of three point, cost function J is equal to 0 .
-
let's look at how F and J change for differrent values of w.
-
second pick value
- w is equal to 0.5
- comput to y hat of three point
- comput to J,we work out the math
- The cost J is about 0.58
- plot J graph
-
pick too much values
- comput to J,we work out the math
- plot J graph
-
how to choose w ?
- Well, as you can imagine, choosing a value of w that casues J of w to be as small as possible seems like good bet.
-
J is cost function that measures hot big the squared errors are, so choosing w that minimizes these squared errors, makes them as small as possible, will give us a good model.
key word
- how the cost function can be used to find the best parameters for your model
- you want to fit a straight line to the training data
- you have , model , parameters, cost function
- goal
coursera ppt



本文作者:bob
本文链接:
版权声明:本博客所有文章除特别声明外,均采用 BY-NC-SA
许可协议。转载请注明出处!