-
Notifications
You must be signed in to change notification settings - Fork 91
Expand file tree
/
Copy pathmnist_simple_coot.cpp
More file actions
188 lines (163 loc) · 7.33 KB
/
mnist_simple_coot.cpp
File metadata and controls
188 lines (163 loc) · 7.33 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
/**
* An example of using Feed Forward Neural Network (FFN) for
* solving Digit Recognizer problem from Kaggle website.
*
* The full description of a problem as well as datasets for training
* and testing are available here https://www.kaggle.com/c/digit-recognizer
*
* mlpack is free software; you may redistribute it and/or modify it under the
* terms of the 3-clause BSD license. You should have received a copy of the
* 3-clause BSD license along with mlpack. If not, see
* http://www.opensource.org/licenses/BSD-3-Clause for more information.
*
* @author Eugene Freyman
* @author Omar Shrit
*/
#define MLPACK_ENABLE_ANN_SERIALIZATION
#define MLPACK_HAS_COOT
#include <bandicoot>
#include <mlpack.hpp>
#if ((ENS_VERSION_MAJOR < 2) || \
((ENS_VERSION_MAJOR == 2) && (ENS_VERSION_MINOR < 13)))
#error "need ensmallen version 2.13.0 or later"
#endif
using namespace mlpack;
using namespace std;
coot::Row<size_t> getLabels(coot::mat predOut)
{
coot::Row<size_t> predLabels(predOut.n_cols);
for (coot::uword i = 0; i < predOut.n_cols; ++i)
{
// predLabels(i) = predOut.col(i).index_max();
}
return predLabels;
}
int main()
{
// Dataset is randomly split into validation
// and training parts in the following ratio.
constexpr double RATIO = 0.1;
// The number of neurons in the first layer.
constexpr int H1 = 200;
// The number of neurons in the second layer.
constexpr int H2 = 100;
// Step size of the optimizer.
const double STEP_SIZE = 5e-3;
// Number of data points in each iteration of SGD
const size_t BATCH_SIZE = 64;
// Allow up to 50 epochs, unless we are stopped early by EarlyStopAtMinLoss.
const int EPOCHS = 50;
// Labeled dataset that contains data for training is loaded from CSV file,
// rows represent features, columns represent data points.
arma::mat dataset;
data::Load("../data/mnist_train.csv", dataset, true);
// Originally on Kaggle dataset CSV file has header, so it's necessary to
// get rid of the this row, in Armadillo representation it's the first column.
arma::mat headerLessDataset =
dataset.submat(0, 1, dataset.n_rows - 1, dataset.n_cols - 1);
// Splitting the training dataset on training and validation parts.
arma::mat train, valid;
data::Split(headerLessDataset, train, valid, RATIO);
// Getting training and validating dataset with features only and then
// normalising
const coot::mat trainX =
coot::conv_to<coot::mat>::from(train.submat(1, 0, train.n_rows - 1, train.n_cols - 1) / 255.0);
const coot::mat validX =
coot::conv_to<coot::mat>::from(valid.submat(1, 0, valid.n_rows - 1, valid.n_cols - 1) / 255.0);
// Labels should specify the class of a data point and be in the interval [0,
// numClasses).
// Creating labels for training and validating dataset.
const coot::mat trainY = coot::conv_to<coot::mat>::from(train.row(0));
const coot::mat validY = coot::conv_to<coot::mat>::from(valid.row(0));
// Specifying the NN model. NegativeLogLikelihood is the output layer that
// is used for classification problem. GlorotInitialization means that
// initial weights in neurons are a uniform gaussian distribution.
FFN<NegativeLogLikelihoodType<coot::mat>, GlorotInitialization, coot::mat> model;
// This is intermediate layer that is needed for connection between input
// data and relu layer. Parameters specify the number of input features
// and number of neurons in the next layer.
model.Add<LinearType<coot::mat>>(H1);
// The first relu layer.
model.Add<ReLUType<coot::mat>>();
// Intermediate layer between relu layers.
model.Add<LinearType<coot::mat>>(H2);
// The second relu layer.
model.Add<ReLUType<coot::mat>>();
// Dropout layer for regularization. First parameter is the probability of
// setting a specific value to 0.
model.Add<DropoutType<coot::mat>>(0.2);
// Intermediate layer.
model.Add<LinearType<coot::mat>>(10);
// LogSoftMax layer is used together with NegativeLogLikelihood for mapping
// output values to log of probabilities of being a specific class.
model.Add<LogSoftMaxType<coot::mat>>();
cout << "Start training ..." << endl;
// Set parameters for the Adam optimizer.
ens::Adam optimizer(
STEP_SIZE, // Step size of the optimizer.
BATCH_SIZE, // Batch size. Number of data points that are used in each
// iteration.
0.9, // Exponential decay rate for the first moment estimates.
0.999, // Exponential decay rate for the weighted infinity norm estimates.
1e-8, // Value used to initialise the mean squared gradient parameter.
EPOCHS * trainX.n_cols, // Max number of iterations.
1e-8, // Tolerance.
true);
// Declare callback to store best training weights.
ens::StoreBestCoordinates<coot::mat> bestCoordinates;
// Train neural network. If this is the first iteration, weights are
// random, using current values as starting point otherwise.
model.Train(trainX,
trainY,
optimizer,
ens::PrintLoss(),
ens::ProgressBar(),
// Stop the training using Early Stop at min loss.
ens::EarlyStopAtMinLossType<coot::mat>(
[&](const coot::mat& /* param */)
{
double validationLoss = model.Evaluate(validX, validY);
cout << "Validation loss: " << validationLoss << "."
<< endl;
return validationLoss;
}),
// Store best coordinates (neural network weights)
bestCoordinates);
// Save the best training weights into the model.
model.Parameters() = bestCoordinates.BestCoordinates();
coot::mat predOut;
// Getting predictions on training data points.
model.Predict(trainX, predOut);
// Calculating accuracy on training data points.
coot::Row<size_t> predLabels = getLabels(predOut);
double trainAccuracy =
coot::accu(predLabels == trainY) / (double) trainY.n_elem * 100;
// Getting predictions on validating data points.
model.Predict(validX, predOut);
// Calculating accuracy on validating data points.
predLabels = getLabels(predOut);
double validAccuracy =
coot::accu(predLabels == validY) / (double) validY.n_elem * 100;
cout << "Accuracy: train = " << trainAccuracy << "%,"
<< "\t valid = " << validAccuracy << "%" << endl;
data::Save("model.bin", "model", model, false);
// Loading test dataset (the one whose predicted labels
// should be sent to kaggle website).
//data::Load("../data/mnist_test.csv", dataset, true);
//coot::mat testY = dataset.row(0);
//dataset.shed_row(0); // Strip labels before predicting.
//dataset /= 255.0; // Apply the same normalization as to the training data.
//cout << "Predicting on test set..." << endl;
//coot::mat testPredOut;
//// Getting predictions on test data points.
//model.Predict(dataset, testPredOut);
//// Generating labels for the test dataset.
//coot::Row<size_t> testPred = getLabels(testPredOut);
//double testAccuracy = coot::accu(testPred == testY) /
//(double) testY.n_elem * 100;
//cout << "Accuracy: test = " << testAccuracy << "%" << endl;
//cout << "Saving predicted labels to \"results.csv\" ..." << endl;
//testPred.save("results.csv", coot::csv_ascii);
//cout << "Neural network model is saved to \"model.bin\"" << endl;
cout << "Finished" << endl;
}