首页
网站开发
桌面应用
管理软件
微信开发
App开发
嵌入式软件
工具软件
数据采集与分析
其他
首页
>
> 详细
CIS5200代写、代做c++,Java编程
项目预算:
开发周期:
发布时间:
要求地区:
CIS5200: Machine Learning Fall 2023
Homework 1
Release Date: September 20, 2023 Due Date: October 4, 2023
• You will submit your solution for the written part of HW1 as a single PDF file via Gradescope.
The deadline is 11:59 PM ET. Contact TAs on Ed if you face any issues uploading your
homeworks.
• Collaboration is permitted and encouraged for this homework, though each student must
understand, write, and hand in their own submission. In particular, it is acceptable for
students to discuss problems with each other; it is not acceptable for students to look at
another student’s written answers when writing their own. It is also not acceptable to publicly
post your (partial) solution on Ed, but you are encouraged to ask public questions on Ed. If
you choose to collaborate, you must indicate on each homework with whom you collaborated.
Please refer to the notes and videos posted on the website if you need to recall the material discussed
in the lectures.
1 Written Questions (46 points)
Problem 1: Margin Perceptron (15 points)
Recall the Perceptron algorithm we saw in the lecture. The Perceptron algorithm terminates
once it classifies all points correctly. It does not guarantee that the hyperplane that that it finds
has large margin (γ) despite our assumption that the true hyperplane w∗ has margin γ where
γ = mini∈{1,...,m}
In this problem, we will consider the following simple modification to the Perceptron algorithm:
Algorithm 1: Margin Perceptron
Initialize w1 = 0 ∈ R
d
for t = 1, 2, . . . do
if ∃i ∈ {1, . . . , m} s.t. yi ̸= sign
w
⊤
t xi
or |w
⊤
t xi
| ≤ 1 then update wt+1 = wt + yixi
else output wt
end
We will show that Margin Perceptron stops after 3/γ2
steps and returns a hyperplane w such that
min
i∈{1,...,m}
∥w∥2
≥ γ/3.
Note that the margin is the distance of the closest point to the hyperplane, and since ∥w∥2 is not
necessarily norm 1, this quantity is given by mini∈{1,...,m}
|w⊤xi|
∥w∥2
.
As in the lecture, we will assume that ∥xi∥2 ≤ 1 for all i ∈ {1, . . . , m} and ∥w∗∥
2
2 = 1.
1.1 (2 points) Show that after every round t, we have
w
⊤
∗ wt+1 ≥ w
⊤
∗ wt + γ.
1.2 (4 points) Show that after every round t, we have
∥wt+1∥
2
2 ≤ ∥wt∥
2
2 + 3.
1.3 (3 points) Using the above two parts, show that after T rounds,
γT ≤ ∥wT +1∥2 ≤
√
3T .
Hint: Use the Cauchy-Schwarz Inequality: a
⊤b ≤ ∥a∥∥b∥.
1.4 (1 point) Use 1.3, to conclude that T ≤ 3/γ2
.
1.5 (4 points) Show that the output hyperplane w satisfies
min
i
|w
⊤xi
|
∥w∥21
≥
γ
3
.
Hint: You will need to use the results in 1.2 and 1.3 plus the stopping condition of the algorithm.
1.6 (1 point) Why is it desirable to learn a predictor that has margin?
Problem 2: Bayes Optimal Classifier (15 points)
Let η(x) denote the conditional probability of the label being 1 given a point x under the distribution
D. That is
η(x) = Pr[y = 1|x].
Recall that the true risk, under the 0/1 loss, for any classifier f is
R(f) = E
x,y
[1[f(x) ̸= y]] .
The Bayes optimal classfier w.r.t. D is the classifier f∗ that achieves the minimum risk among all
possible classifiers. In this problem, we will work out what the Bayes optimal classifier is.
2.1 (3 points) Show that
R(f) = Ex [η(x)1[f(x) = −1] + (1 − η(x))1[f(x) = 1]] .
Hint: Use the fact that Ex,y[·] = Ex Ey|x[·].
2
2.2 (3 points) Use the above to show that the minimum risk possible is
R(h∗) = min
f
R(f) = Ex [min(η(x), 1 − η(x))]
Hint: For a fixed x, think about what the minimum loss is using 2.1.
2.3 (2 points) Show that the Bayes optimal classifier that achieves the above loss is
f∗(x) = (
1 if η(x) ≥ 1/2,
−1 otherwise..
2.4 (1 point) Derive the Bayes optimal classifier under the logistic model
η(x) = 1
1 + exp(−w⊤x)
.
2.5 (6 points) Now suppose we modify the loss function from 0/1 to the following cost-based loss
function
ℓc(ˆy, y) =
c if y = 1, yˆ = −1
1 − c if y = −1, yˆ = 1
0 if y = ˆy.
Here the loss penalizes false negative with cost c and false positive with cost 1 − c, penalizing
different types of mistakes differently.1
Note that the true risk under this loss is
Rc(f) = E
x,y
[ℓc(f(x), y)] .
Find the Bayes optimal classifier in this setting.
Hint: Follow the same ideas you used to solve 2.1-2.3 using ℓc instead of 0/1 loss.
2 Programming Questions (16 points)
Use the link here to access the Google Colaboratory (Colab) file for this homework. Be sure to
make a copy by going to ”File”, and ”Save a copy in Drive”. This assignment uses the PennGrader
system for students to receive immediate feedback. As noted on the notebook, please be sure to
change the student ID from the default ’99999999’ to your 8-digit PennID.
Instructions for how to submit the programming component of HW 1 to Gradescope are included
in the Colab notebook. You may find this PyTorch reference to be helpful - if you get stuck, it may
be helpful to review some of the PyTorch documentation and functions.
1Let us see why this is a useful loss function. Consider the case of medical diagnosis, high false negative rate
means that we are predicting that patients do not have the disease when they actually do. Such a prediction could
lead to the patient not getting the care they need. In such a setting, you would want c to be closer to 1.
3
软件开发、广告设计客服
QQ:99515681
邮箱:99515681@qq.com
工作时间:8:00-23:00
微信:codinghelp
热点项目
更多
urba6006代写、java/c++编程语...
2024-12-26
代做program、代写python编程语...
2024-12-26
代写dts207tc、sql编程语言代做
2024-12-25
cs209a代做、java程序设计代写
2024-12-25
cs305程序代做、代写python程序...
2024-12-25
代写csc1001、代做python设计程...
2024-12-24
代写practice test preparatio...
2024-12-24
代写bre2031 – environmental...
2024-12-24
代写ece5550: applied kalman ...
2024-12-24
代做conmgnt 7049 – measurem...
2024-12-24
代写ece3700j introduction to...
2024-12-24
代做adad9311 designing the e...
2024-12-24
代做comp5618 - applied cyber...
2024-12-24
热点标签
mktg2509
csci 2600
38170
lng302
csse3010
phas3226
77938
arch1162
engn4536/engn6536
acx5903
comp151101
phl245
cse12
comp9312
stat3016/6016
phas0038
comp2140
6qqmb312
xjco3011
rest0005
ematm0051
5qqmn219
lubs5062m
eee8155
cege0100
eap033
artd1109
mat246
etc3430
ecmm462
mis102
inft6800
ddes9903
comp6521
comp9517
comp3331/9331
comp4337
comp6008
comp9414
bu.231.790.81
man00150m
csb352h
math1041
eengm4100
isys1002
08
6057cem
mktg3504
mthm036
mtrx1701
mth3241
eeee3086
cmp-7038b
cmp-7000a
ints4010
econ2151
infs5710
fins5516
fin3309
fins5510
gsoe9340
math2007
math2036
soee5010
mark3088
infs3605
elec9714
comp2271
ma214
comp2211
infs3604
600426
sit254
acct3091
bbt405
msin0116
com107/com113
mark5826
sit120
comp9021
eco2101
eeen40700
cs253
ece3114
ecmm447
chns3000
math377
itd102
comp9444
comp(2041|9044)
econ0060
econ7230
mgt001371
ecs-323
cs6250
mgdi60012
mdia2012
comm221001
comm5000
ma1008
engl642
econ241
com333
math367
mis201
nbs-7041x
meek16104
econ2003
comm1190
mbas902
comp-1027
dpst1091
comp7315
eppd1033
m06
ee3025
msci231
bb113/bbs1063
fc709
comp3425
comp9417
econ42915
cb9101
math1102e
chme0017
fc307
mkt60104
5522usst
litr1-uc6201.200
ee1102
cosc2803
math39512
omp9727
int2067/int5051
bsb151
mgt253
fc021
babs2202
mis2002s
phya21
18-213
cege0012
mdia1002
math38032
mech5125
07
cisc102
mgx3110
cs240
11175
fin3020s
eco3420
ictten622
comp9727
cpt111
de114102d
mgm320h5s
bafi1019
math21112
efim20036
mn-3503
fins5568
110.807
bcpm000028
info6030
bma0092
bcpm0054
math20212
ce335
cs365
cenv6141
ftec5580
math2010
ec3450
comm1170
ecmt1010
csci-ua.0480-003
econ12-200
ib3960
ectb60h3f
cs247—assignment
tk3163
ics3u
ib3j80
comp20008
comp9334
eppd1063
acct2343
cct109
isys1055/3412
math350-real
math2014
eec180
stat141b
econ2101
msinm014/msing014/msing014b
fit2004
comp643
bu1002
cm2030
联系我们
- QQ: 9951568
© 2021
www.rj363.com
软件定制开发网!