热搜关键词: 电路基础ADC数字信号处理封装库PLC

pdf

统计机器学习(斯坦福大学讲义)1-12(全)

  • 1星
  • 2021-05-04
  • 5.76MB
  • 需要1积分
  • 4次下载
标签: 机器学习

机器学习

AI

AI

统计机器学习(斯坦福大学讲义)1-12(全)

CS229 Lecture notes
Andrew Ng
Supervised learning
Let’s start by talking about a few examples of supervised learning problems.
Suppose we have a dataset giving the living areas and prices of 47 houses
from Portland, Oregon:
Living area (feet
2
)
2104
1600
2400
1416
3000
.
.
.
We can plot this data:
housing prices
1000
900
800
700
price (in $1000)
600
500
400
300
200
100
0
500
1000
1500
2000
2500
3000
square feet
3500
4000
4500
5000
Price (1000$s)
400
330
369
232
540
.
.
.
Given data like this, how can we learn to predict the prices of other houses
in Portland, as a function of the size of their living areas?
1
CS229 Fall 2012
2
To establish notation for future use, we’ll use
x
(i)
to denote the “input”
variables (living area in this example), also called input
features,
and
y
(i)
to denote the “output” or
target
variable that we are trying to predict
(price). A pair (x
(i)
, y
(i)
) is called a
training example,
and the dataset
that we’ll be using to learn—a list of
m
training examples
{(x
(i)
, y
(i)
);
i
=
1,
. . . , m}—is
called a
training set.
Note that the superscript “(i)” in the
notation is simply an index into the training set, and has nothing to do with
exponentiation. We will also use
X
denote the space of input values, and
Y
the space of output values. In this example,
X
=
Y
=
R.
To describe the supervised learning problem slightly more formally, our
goal is, given a training set, to learn a function
h
:
X → Y
so that
h(x)
is a
“good” predictor for the corresponding value of
y.
For historical reasons, this
function
h
is called a
hypothesis.
Seen pictorially, the process is therefore
like this:
Training
set
Learning
algorithm
x
(living area of
house.)
h
predicted y
(predicted price)
of house)
When the target variable that we’re trying to predict is continuous, such
as in our housing example, we call the learning problem a
regression
prob-
lem. When
y
can take on only a small number of discrete values (such as
if, given the living area, we wanted to predict if a dwelling is a house or an
apartment, say), we call it a
classification
problem.
3
Part I
Linear Regression
To make our housing example more interesting, let’s consider a slightly richer
dataset in which we also know the number of bedrooms in each house:
Living area (feet
2
)
2104
1600
2400
1416
3000
.
.
.
#bedrooms Price (1000$s)
3
400
3
330
3
369
2
232
4
540
.
.
.
.
.
.
(i)
Here, the
x’s
are two-dimensional vectors in
R
2
. For instance,
x
1
is the
(i)
living area of the
i-th
house in the training set, and
x
2
is its number of
bedrooms. (In general, when designing a learning problem, it will be up to
you to decide what features to choose, so if you are out in Portland gathering
housing data, you might also decide to include other features such as whether
each house has a fireplace, the number of bathrooms, and so on. We’ll say
more about feature selection later, but for now let’s take the features as
given.)
To perform supervised learning, we must decide how we’re going to rep-
resent functions/hypotheses
h
in a computer. As an initial choice, let’s say
we decide to approximate
y
as a linear function of
x:
h
θ
(x) =
θ
0
+
θ
1
x
1
+
θ
2
x
2
Here, the
θ
i
’s are the
parameters
(also called
weights)
parameterizing the
space of linear functions mapping from
X
to
Y.
When there is no risk of
confusion, we will drop the
θ
subscript in
h
θ
(x), and write it more simply as
h(x).
To simplify our notation, we also introduce the convention of letting
x
0
= 1 (this is the
intercept term),
so that
n
h(x)
=
i=0
θ
i
x
i
=
θ
T
x,
where on the right-hand side above we are viewing
θ
and
x
both as vectors,
and here
n
is the number of input variables (not counting
x
0
).
4
Now, given a training set, how do we pick, or learn, the parameters
θ?
One reasonable method seems to be to make
h(x)
close to
y,
at least for
the training examples we have. To formalize this, we will define a function
that measures, for each value of the
θ’s,
how close the
h(x
(i)
)’s are to the
corresponding
y
(i)
’s. We define the
cost function:
1
J(θ)
=
2
m
i=1
(h
θ
(x
(i)
)
y
(i)
)
2
.
If you’ve seen linear regression before, you may recognize this as the familiar
least-squares cost function that gives rise to the
ordinary least squares
regression model. Whether or not you have seen it previously, let’s keep
going, and we’ll eventually show this to be a special case of a much broader
family of algorithms.
1
LMS algorithm
We want to choose
θ
so as to minimize
J(θ).
To do so, let’s use a search
algorithm that starts with some “initial guess” for
θ,
and that repeatedly
changes
θ
to make
J(θ)
smaller, until hopefully we converge to a value of
θ
that minimizes
J(θ).
Specifically, let’s consider the
gradient descent
algorithm, which starts with some initial
θ,
and repeatedly performs the
update:
θ
j
:=
θ
j
α
J(θ).
∂θ
j
(This update is simultaneously performed for all values of
j
= 0,
. . . , n.)
Here,
α
is called the
learning rate.
This is a very natural algorithm that
repeatedly takes a step in the direction of steepest decrease of
J.
In order to implement this algorithm, we have to work out what is the
partial derivative term on the right hand side. Let’s first work it out for the
case of if we have only one training example (x,
y),
so that we can neglect
the sum in the definition of
J.
We have:
1
J(θ)
=
(h
θ
(x)
y)
2
∂θ
j
∂θ
j
2
1
= 2
·
(h
θ
(x)
y)
·
(h
θ
(x)
y)
2
∂θ
j
= (h
θ
(x)
y)
·
∂θ
j
= (h
θ
(x)
y) x
j
n
i=0
θ
i
x
i
y
5
For a single training example, this gives the update rule:
1
θ
j
:=
θ
j
+
α y
(i)
h
θ
(x
(i)
)
x
j
.
The rule is called the
LMS
update rule (LMS stands for “least mean squares”),
and is also known as the
Widrow-Hoff
learning rule. This rule has several
properties that seem natural and intuitive. For instance, the magnitude of
the update is proportional to the
error
term (y
(i)
h
θ
(x
(i)
)); thus, for in-
stance, if we are encountering a training example on which our prediction
nearly matches the actual value of
y
(i)
, then we find that there is little need
to change the parameters; in contrast, a larger change to the parameters will
be made if our prediction
h
θ
(x
(i)
) has a large error (i.e., if it is very far from
y
(i)
).
We’d derived the LMS rule for when there was only a single training
example. There are two ways to modify this method for a training set of
more than one example. The first is replace it with the following algorithm:
Repeat until convergence
{
θ
j
:=
θ
j
+
α
}
The reader can easily verify that the quantity in the summation in the update
rule above is just
∂J(θ)/∂θ
j
(for the original definition of
J).
So, this is
simply gradient descent on the original cost function
J.
This method looks
at every example in the entire training set on every step, and is called
batch
gradient descent.
Note that, while gradient descent can be susceptible
to local minima in general, the optimization problem we have posed here
for linear regression has only one global, and no other local, optima; thus
gradient descent always converges (assuming the learning rate
α
is not too
large) to the global minimum. Indeed,
J
is a convex quadratic function.
Here is an example of gradient descent as it is run to minimize a quadratic
function.
We use the notation “a :=
b”
to denote an operation (in a computer program) in
which we
set
the value of a variable
a
to be equal to the value of
b.
In other words, this
operation overwrites
a
with the value of
b.
In contrast, we will write “a =
b”
when we are
asserting a statement of fact, that the value of
a
is equal to the value of
b.
1
(i)
m
i=1
y
(i)
h
θ
(x
(i)
)
x
j
(i)
(for every
j).
展开预览

猜您喜欢

评论

登录/注册

意见反馈

求资源

回顶部

推荐内容

热门活动

热门器件

随便看看

 
EEWorld订阅号

 
EEWorld服务号

 
汽车开发圈

电子工程世界版权所有 京B2-20211791 京ICP备10001474号-1 电信业务审批[2006]字第258号函 京公网安备 11010802033920号 Copyright © 2005-2024 EEWORLD.com.cn, Inc. All rights reserved
×