Skip to content

Commit fef78e7

Browse files
committed
Last
Fixes #10
1 parent ae56c31 commit fef78e7

7 files changed

Lines changed: 7 additions & 954 deletions

Project Report/ProjectTemplate.pdf renamed to Project Report/MICHELUCCI_867644_SINGLAN_867646_Report.pdf

215 KB
Binary file not shown.

Project Report/ProjectTemplate.tex renamed to Project Report/MICHELUCCI_867644_SINGLAN_867646_Report.tex

Lines changed: 7 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -278,6 +278,7 @@ \section{The Methodological Approach}
278278
As expected, the dataset obtained is more balanced, which allows more constant results.
279279

280280
\section{Results and Evaluation}
281+
281282
For each experiment, the results are variable. So, the best results obtained are presented below.
282283

283284
Using the second pre-trained model (finetuned on bacterial spectra) to predict on raw dataset, an average accuracy score of 66\% with a standard deviation of 1.11 is obtained. (Figure 1)
@@ -287,27 +288,31 @@ \section{Results and Evaluation}
287288
\includegraphics[width=10cm, height=10cm]{raw_model_results}
288289
\end{figure}
289290

291+
\newpage
290292
Adding fine-tuning method described in the previous section to the same model resulted in an accuracy of 72.8$\pm$ 0.04. These results are illustrated by the confusion matrix below. (Figure 2)
291293
\begin{figure}[H]
292294
\centering
293295
\caption{Confusion-matrix resulting from the pre-trained model, with fine-tuning method}
294296
\includegraphics[width=10cm, height=10cm]{finetuning_model_results}
295297
\end{figure}
296298

299+
\newpage
297300
Whith the data augmentation, the accuracy obtained is sligthly better: 74.4 $\pm$ 0.01. (Figure 3)
298301
\begin{figure}[H]
299302
\centering
300303
\caption{Confusion-matrix resulting from the pre-trained model, with fine-tuning and data augmentation}
301304
\includegraphics[width=10cm, height=10cm]{finetuning_data_augmentation_model_results}
302305
\end{figure}
303306

307+
\newpage
304308
About the features extraction, the best performances are computed using the same pre-trained model and cuting its last block. The accuracy obtained is 72.6\% which is slightly lower than the one with the fine-tuning process. This result is attained using a Deep ML model as predictor. (Figure 4) Applying the classical model the accuracy is better, but overfitting occurs.
305309
\begin{figure}[H]
306310
\centering
307311
\caption{Confusion-matrix resulting from the pre-trained model as features extractor}
308312
\includegraphics[width=10cm, height=10cm]{extractor_model_results}
309313
\end{figure}
310314

315+
\newpage
311316
When data augmentation is applied to the previous extractor model, the results slightly improved. The accuracy obtained is about 77\%. (Figure 5)
312317
\begin{figure}[H]
313318
\centering
@@ -317,7 +322,7 @@ \section{Results and Evaluation}
317322

318323

319324
\section{Discussion}
320-
%The discussion section aims at interpreting the results in light of the project's objectives. The most important goal of this section is to interpret the results so that the reader is informed of the insight or answers that the results provide. This section should also present an evaluation of the particular approach taken by the group. For example: Based on the results, how could the experimental procedure be improved? What additional, future work may be warranted? What recommendations can be drawn?
325+
321326
As mentionned in the previous section the results chosen are the best simulated after multiple runs and are quite variable. Depending on the distribution of the patients (and the samples) in the finetuning and features extractor tests, performances can drop.
322327
It is quite problematic as it is a medical concern.
323328
In fact, if a healthy patient is diagnosed and treated for an illness it can be dangerous.
@@ -330,7 +335,7 @@ \section{Discussion}
330335
\end{itemize}
331336

332337
\section{Conclusions}
333-
%Conclusions should summarize the central points made in the Discussion section, reinforcing for the reader the value and implications of the work. If the results were not definitive, specific future work that may be needed can be (briefly) described. The conclusions should never contain ``surprises''. Therefore, any conclusions should be based on observations and data already discussed. It is considered extremely bad form to introduce new data in the conclusions.
338+
334339
The Transfer Learning approach is a good approach for the analysis of the Raman spectra, especially in case of data scarsity. In addition, data augmentation techniques are helpful to make more robust models.
335340

336341
A future work could be to find some better way to augment spectral data like Generative Adversarial Networks (GANs). The generative models could demonstrate an improvement in the accuracy of ML classification and also a solution to the problem of requiring a large amount of training data.

Project Report/ProjectTemplate.aux

Lines changed: 0 additions & 42 deletions
This file was deleted.

0 commit comments

Comments
 (0)