You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As expected, the dataset obtained is more balanced, which allows more constant results.
279
279
280
280
\section{Results and Evaluation}
281
+
281
282
For each experiment, the results are variable. So, the best results obtained are presented below.
282
283
283
284
Using the second pre-trained model (finetuned on bacterial spectra) to predict on raw dataset, an average accuracy score of 66\% with a standard deviation of 1.11 is obtained. (Figure 1)
@@ -287,27 +288,31 @@ \section{Results and Evaluation}
Adding fine-tuning method described in the previous section to the same model resulted in an accuracy of 72.8$\pm$ 0.04. These results are illustrated by the confusion matrix below. (Figure 2)
291
293
\begin{figure}[H]
292
294
\centering
293
295
\caption{Confusion-matrix resulting from the pre-trained model, with fine-tuning method}
About the features extraction, the best performances are computed using the same pre-trained model and cuting its last block. The accuracy obtained is 72.6\% which is slightly lower than the one with the fine-tuning process. This result is attained using a Deep ML model as predictor. (Figure 4) Applying the classical model the accuracy is better, but overfitting occurs.
305
309
\begin{figure}[H]
306
310
\centering
307
311
\caption{Confusion-matrix resulting from the pre-trained model as features extractor}
When data augmentation is applied to the previous extractor model, the results slightly improved. The accuracy obtained is about 77\%. (Figure 5)
312
317
\begin{figure}[H]
313
318
\centering
@@ -317,7 +322,7 @@ \section{Results and Evaluation}
317
322
318
323
319
324
\section{Discussion}
320
-
%The discussion section aims at interpreting the results in light of the project's objectives. The most important goal of this section is to interpret the results so that the reader is informed of the insight or answers that the results provide. This section should also present an evaluation of the particular approach taken by the group. For example: Based on the results, how could the experimental procedure be improved? What additional, future work may be warranted? What recommendations can be drawn?
325
+
321
326
As mentionned in the previous section the results chosen are the best simulated after multiple runs and are quite variable. Depending on the distribution of the patients (and the samples) in the finetuning and features extractor tests, performances can drop.
322
327
It is quite problematic as it is a medical concern.
323
328
In fact, if a healthy patient is diagnosed and treated for an illness it can be dangerous.
@@ -330,7 +335,7 @@ \section{Discussion}
330
335
\end{itemize}
331
336
332
337
\section{Conclusions}
333
-
%Conclusions should summarize the central points made in the Discussion section, reinforcing for the reader the value and implications of the work. If the results were not definitive, specific future work that may be needed can be (briefly) described. The conclusions should never contain ``surprises''. Therefore, any conclusions should be based on observations and data already discussed. It is considered extremely bad form to introduce new data in the conclusions.
338
+
334
339
The Transfer Learning approach is a good approach for the analysis of the Raman spectra, especially in case of data scarsity. In addition, data augmentation techniques are helpful to make more robust models.
335
340
336
341
A future work could be to find some better way to augment spectral data like Generative Adversarial Networks (GANs). The generative models could demonstrate an improvement in the accuracy of ML classification and also a solution to the problem of requiring a large amount of training data.
0 commit comments