aboutsummaryrefslogtreecommitdiff
path: root/report/paper.md
diff options
context:
space:
mode:
authornunzip <np.scarh@gmail.com>2018-11-20 14:50:42 +0000
committernunzip <np.scarh@gmail.com>2018-11-20 14:50:42 +0000
commit35046978650c590b36b3160c87bd3c56ce9e19ec (patch)
tree5f74fa750a59bd62de06715b02a9a6e265fd5a89 /report/paper.md
parentab88e84c237d01181672279fcb08e24c51cb2e7d (diff)
parenta419a8a6cc6df8d98ddd4f38f5a98491863b741e (diff)
downloadvz215_np1915-35046978650c590b36b3160c87bd3c56ce9e19ec.tar.gz
vz215_np1915-35046978650c590b36b3160c87bd3c56ce9e19ec.tar.bz2
vz215_np1915-35046978650c590b36b3160c87bd3c56ce9e19ec.zip
Merge branch 'master' of skozl.com:e4-pattern
Diffstat (limited to 'report/paper.md')
-rwxr-xr-xreport/paper.md26
1 files changed, 22 insertions, 4 deletions
diff --git a/report/paper.md b/report/paper.md
index 0d91bfe..961937b 100755
--- a/report/paper.md
+++ b/report/paper.md
@@ -432,7 +432,16 @@ The optimal number of constant and random eigenvectors to use is therefore an in
The optimal randomness after doing an exhaustive search as seen on figure \label{fig:opti-rand}peaks at
95 randomised eigenvectors out of 155 total eigenvectors, or 60 static and 95 random eigenvectors. The values of $M_{\textrm{lda}}$ in the figures is the maximum of 51.
-The red peaks on the 3d-plot represent the proportion of randomised eigenvectors which achieve the optimal accuracy, which have been further plotted in figure \label{opt-2d}
+The red peaks on the 3d-plot represent the proportion of randomised eigenvectors which achieve the optimal accuracy, which have been further plotted in figure \ref{opt-2d}. We found that for our data, the optimal ratio of random eigenvectors for a given $M$ is between $0.6$ and $0.9$.
+
+\begin{figure}
+\begin{center}
+\includegraphics[width=19em]{fig/nunzplot1.pdf}
+\caption{Optimal randomness ratio}
+\label{fig:opt-2d}
+\end{center}
+\end{figure}
+
### Ensemble Confusion Matrix
@@ -444,11 +453,20 @@ The red peaks on the 3d-plot represent the proportion of randomised eigenvectors
\end{center}
\end{figure}
-We can compute an ensemble confusion matrix before the committee machines as shown in figure \ref{fig:ens-cm}. This confusion matrix combines the output of all the models in the ensemble. As can be seen from the figure, different models make different mistakes.
-
+We can compute an ensemble confusion matrix before the committee machines as shown in figure \ref{fig:ens-cm}. This confusion matrix combines the output of all the models in the ensemble. As can be seen from the figure, models in the ensemble usually make more mistakes than an individual model. When the ensemble size is large enough, the errors are rectified by the committee machine, resulting in low error as observed in figure \ref{fig:random-e}.
## Comparison
-Combining bagging and feature space randomization we are able to achieve higher test accuracy than the individual models. Here is a comparison for various splits.
+Combining bagging and feature space randomization we are able to consistently achieve higher test accuracy than the individual models. In table \ref{tab:compare} $70/30$ splits.
+
+\begin{table}[]
+\begin{tabular}{lrr} \hline
+Seed & Individual$(M=120)$ & Bag + Feature Ens.$(M=60+95)$\\ \hline
+0 & 0.916 & 0.923 \\
+1 & 0.929 & 0.942 \\
+5 & 0.897 & 0.910 \\ \hline
+\end{tabular}
+\label{tab:compare}
+\end{table}
# References