Guest User

Untitled

a guest
Jun 23rd, 2018
76
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 9.12 KB | None | 0 0
  1. documentclass[12pt,a4paper,twoside,openany]{report}
  2. usepackage[utf8]{inputenc}
  3.  
  4.  
  5.  
  6. usepackage{tabu}
  7.  
  8. usepackage{array}
  9.  
  10.  
  11. usepackage{diagbox}
  12. usepackage{moreverb}
  13. usepackage{commath}
  14.  
  15. usepackage{textcomp}
  16. usepackage{lmodern}
  17. usepackage{helvet}
  18. usepackage[T1]{fontenc}
  19. usepackage[english]{babel}
  20. usepackage[utf8]{inputenc}
  21. usepackage{amsmath}
  22. usepackage{amssymb}
  23. usepackage{graphicx}
  24. usepackage{subfig}
  25. numberwithin{equation}{chapter}
  26. numberwithin{figure}{chapter}
  27. numberwithin{table}{chapter}
  28. usepackage{listings}
  29.  
  30. usepackage[top=3cm, bottom=3cm,
  31. inner=3cm, outer=3cm]{geometry}
  32. usepackage{eso-pic}
  33. newcommand{backgroundpic}[3]{
  34. put(#1,#2){
  35. parbox[b][paperheight]{paperwidth}{
  36. centering
  37. includegraphics[width=paperwidth,height=paperheight,keepaspectratio]{#3}}}}
  38. usepackage{float}
  39. usepackage{parskip}
  40. setlength{parindent}{0cm}
  41.  
  42. usepackage{hyperref}
  43. hypersetup{colorlinks, citecolor=black,
  44. filecolor=black, linkcolor=black,
  45. urlcolor=black}
  46.  
  47.  
  48. setcounter{tocdepth}{5}
  49. setcounter{secnumdepth}{5}
  50.  
  51.  
  52. usepackage{titlesec}
  53. titleformat{chapter}[display]
  54. {Hugebfseriesfilcenter}
  55. {{fontsize{50pt}{1em}vspace{-4.2ex}selectfont textnormal{thechapter}}}{1ex}{}[]
  56.  
  57.  
  58. usepackage{fancyhdr}
  59. pagestyle{fancy}
  60. renewcommand{chaptermark}[1]{markboth{thechapter.space#1}{}}
  61.  
  62.  
  63. deflayout{2}
  64.  
  65.  
  66. ifnumlayout=2
  67. fancyhf{}
  68. fancyhead[LE,RO]{nouppercase{ leftmark}}
  69. fancyfoot[LE,RO]{thepage}
  70. fancypagestyle{plain}{
  71. fancyhf{}
  72. renewcommand{headrulewidth}{0pt}
  73. fancyfoot[LE,RO]{thepage}}
  74. else
  75. fancyhf{}
  76. fancyhead[C]{nouppercase{ leftmark}}
  77. fancyfoot[C]{thepage}
  78. fi
  79.  
  80.  
  81. usepackage[textsize=tiny]{todonotes}
  82.  
  83. setlength{marginparwidth}{2.5cm}
  84.  
  85. setlength{headheight}{15pt}
  86.  
  87.  
  88.  
  89.  
  90. begin{document}
  91.  
  92.  
  93. section{Network classification performance}
  94.  
  95. In this section the prediction accuracy results for the baseline model and the three different variations are presented. subsection{Baseline}
  96.  
  97.  
  98. The obtained results for the single-label classification task using the UrbanSound8K dataset are shown in Table ref{tab:results_baseline}. Following the same procedure as cite{dataset2}, accuracy was calculated as the average of the individual prediction accuracies across the 10 folders. The results reveal a high degree of overfitting to the training data, measured by the difference between training and test accuracy. The different explicit regularization techniques tried out during the hyperparameter search, i.e. dropout and weight decay, did not help to reduce this generalization gap, suggesting that generalization is a difficult task on this particular dataset. It is prone to overfitting due to the low intra-class variance of some of the classes. The main reason for this low variance is the fact that many of the sound excerpts of each class were extracted from the same audio file when the dataset was created.
  99.  
  100. This bias is more pronounced for those classes whose original sound recordings had longer duration. It is less likely to find a sound recording of 20 seconds duration for the class textit{Gun shot}, whereas for the classes textit{Jackhammer} or textit{Air conditioner} that is many times the case.
  101.  
  102. begin{table}[H] centering captionsetup{singlelinecheck = false, justification = raggedright} begin{tabular}{ |
  103. >{centeringarraybackslash} m{4cm}| >{centeringarraybackslash} m{4cm} |} hline hline Training accuracy & Test accuracy \ hline
  104. 90.60 % & 64.82 % \ hline hline end{tabular} caption{Training and test accuracy results (10-fold cross validated).} label{tab:results_baseline} end{table}
  105.  
  106.  
  107. In Figure ref{fig:cnf_baseline} the normalized confusion matrices for the trainingsets can be seen. This gives a clearer picture about the overfitting distribution over the different classes. The differences between training and test accuracies reveal the degree of overfitting present in the model for each class. A big difference the network is memorizing the training samples, but is not able to generalize well to new unseen test samples. textit{Air conditioner}, textit{Engine idling} and textit{Jackhammer} are the classes where the highest d of overfitting is observed. textit{Gun shot}, textit{Dog bark} and textit{Street music} are the classes where the lowest degree of overfitting is observed.
  108.  
  109. begin{figure}[H]
  110. centering
  111. begin{minipage}[b]{0.45textwidth}
  112. includegraphics[width=textwidth]{cnf_original_train_10fold.PNG}
  113. % caption{Training set confusion matrix}
  114. % label{fig:cnf_train_baseline}
  115. end{minipage}
  116. hfill
  117. begin{minipage}[b]{0.45textwidth}
  118. includegraphics[width=textwidth]{cnf_original_test_10fold.PNG}
  119. % caption{Test set confusion matrix}
  120. % label{fig:cnf_test_baseline}
  121. end{minipage}
  122. caption{Training (left) and test (right) confusion matrices of the baseline model.}
  123. label{fig:cnf_baseline}
  124. end{figure}
  125.  
  126.  
  127. The following observations can be drawn from these For the three classes with the highest degree of :
  128.  
  129. begin{itemize}
  130. item According to the aforementioned observations about the original duration of the audio files, these three are the classes presenting the lowest intra-class variance.
  131. item All of them present a noisy sound nature (see Figure ref{fig:spectrogram_images}), making it difficult for the network to learn meaningful patterns.
  132. item The highest confusions when classifying each of these classes occur between these three classes and also the class textit{Drilling}. This is a proof of the similarity of these four classes and the consequent difficulty of differentiating between them.
  133. end{itemize}
  134.  
  135. As for the classes with the lowest degree of overfitting:
  136.  
  137. begin{itemize}
  138. item These are either transient sounds or classes with high degree of intra-class variance, such as textit{Street music}.
  139. item The instances of the class textit{Street music} might have been extracted from the same audio file. However, they will still present a high degree of variance since music is a non-stationary process.
  140. item A high variance among class instances forces the network to learn meaningful features of each class, since it cannot rely on any particular characteristic common only to the training data. Therefore the lower degree of overfitting for the textit{Street music} class.
  141. end{itemize}
  142.  
  143. By looking at Figure ref{fig:UrbanSound8K_slices_FGBG} it can be seen that the classes textit{Siren} and textit{Car horn} are the only classes where a higher number of background instances are present, compared to foreground instances.
  144. This explains why, besides being a transient (textit{Car horn}) and, in theory (see Figure ref{fig:spectrogram_images}), an easily identifiable sound (textit{Siren}), the network occasionally confuses them with sounds like textit{Street music} or textit{Children playing}, which are common background city noises.
  145.  
  146. subsection{Dataset variations}
  147.  
  148. The results obtained when training the model on the three dataset variations introduced in Section 3.7 are presented in this subsection.
  149.  
  150.  
  151. subsubsection{Undersampling}
  152.  
  153.  
  154. In Table ref{tab:reduced_accuracies2} the changes in prediction accuracy per class when performing undersampling of the dataset can be seen. The values shadowed in color were calculated as: $accuracy: version_i - baseline:accuracy$, where $i in (A,E)$.
  155. begin{table}[H]
  156. centering
  157. includegraphics[scale=0.6]{reduced_class_variations_goodone.png}
  158. caption{Change in prediction accuracy (%) per class for different variations of the training set.}
  159. label{tab:reduced_accuracies2}
  160. end{table}
  161.  
  162.  
  163.  
  164. The study reveals that undersampling is not procedure for this particular dataset. Some classes remained unaltered to the variations, like textit{Dog bark}, textit{Siren} or textit{Street music}, whereas other classes like textit{Air conditioner} were singnificantly. None of the variations had an overall ince in prediction accura.
  165.  
  166. An example of how reducing the training data can have dangerous consequences is shown in Figure ref{fig:comp_original_reduced5}. When the amount of samples of the class textit{Engine} is reduced, the accuracy of the class textit{Air conditioner} remains unaltered. However, it can be seen than now the network is confusing the textit{Engine} sounds with the class textit{Air conditioner}. This gives an intuition about how close some classes are to others and the consequent difficulty for the network to differentiate between them. Thus, reducing the amount of information about one of these classes makes the network more likely to get confused when trying to identify other classes that are similar to the class whose samples were reduced.
  167.  
  168.  
  169. end{document}
Add Comment
Please, Sign In to add comment