You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/homeworks/hw3.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,17 +9,17 @@ layout: home
9
9
10
10
In this homework, you will be implementing a few popular computer vision models, and training them on both CIFAR-10 and on a custom dataset we created. You will be using PyTorch for this homework.
11
11
12
-
You will be using a medium-sized repository which mimics that of a standard codebase which you might find for modern projects. Don't be intimidated! We will walk you through all of the parts of it, and hopefully after this homework you will be more confident working with codebases like this. We believe this is a realistic representation of what you may do in the future, and we hope you will find it useful.
12
+
You will be using a medium-sized [repository](https://github.com/mlberkeley/fa25-nmep-hw3/) which mimics that of a standard codebase which you might find for modern projects. Don't be intimidated! We will walk you through all of the parts of it, and hopefully after this homework you will be more confident working with codebases like this. We believe this is a realistic representation of what you may do in the future, and we hope you will find it useful.
13
13
14
14
We would recommend you first set up the repository ASAP on honeydew and try running it out of the box to see how it trains, and only afterwards focus on understanding all parts of the code. For your benefit, the codebase works out of the box, and you should be able to train a model on CIFAR-10 with no changes. Throughout the assignment, you will need to make some changes to models/alexnet.py and models/resnet.py, for which you will find the provided implementations of other models in models/ to be helpful.
15
15
16
-
All of the assignment details are provided in WORKSHEET.md - you will need to fill in some answers and make code changes. We recommend getting through this README.md file first and then doing the worksheet.
16
+
All of the assignment details are provided in this spec, you will need to fill in some answers on Gradescope and make code changes.
17
17
18
18
Best of luck, and we hope you enjoy it!
19
19
20
20
## Setup
21
21
22
-
To get started, you will need to fork and clone the repository (clone on honeydew!) and install the dependencies, preferably in a conda environment. Standard instructions are provided below.
22
+
To get started, you will need to fork and clone [the repository](https://github.com/mlberkeley/fa25-nmep-hw3/) (clone on honeydew!) and install the dependencies, preferably in a conda or uv environment. Standard instructions are provided below.
23
23
24
24
```bash
25
25
ssh honeydew
@@ -119,7 +119,7 @@ Linear with 4096 output units
119
119
Linear with num_classes output units
120
120
```
121
121
122
-
> ReLU activation after every Conv and Linear layer. DO **NOT** Forget to add activatioons after every layer. Do not apply activation after the last layer.
122
+
> ReLU activation after every Conv and Linear layer. DO **NOT** Forget to add activations after every layer. Do not apply activation after the last layer.
123
123
124
124
### 2.1.1
125
125
How many parameters does AlexNet have? How does it compare to LeNet? With the same batch size, how much memory do LeNet and AlexNet take up while training?
0 commit comments