aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authornunzip <np.scarh@gmail.com>2018-11-20 14:55:18 +0000
committernunzip <np.scarh@gmail.com>2018-11-20 14:55:18 +0000
commit8fb95b6acaf059b4689067573194bcd3fdadc659 (patch)
tree5a9873f61f5c46e6f4b8b9ac6af4ace6b6a77ca1
parent35046978650c590b36b3160c87bd3c56ce9e19ec (diff)
downloadvz215_np1915-8fb95b6acaf059b4689067573194bcd3fdadc659.tar.gz
vz215_np1915-8fb95b6acaf059b4689067573194bcd3fdadc659.tar.bz2
vz215_np1915-8fb95b6acaf059b4689067573194bcd3fdadc659.zip
Fix reference section 3. Improve appendix table
-rwxr-xr-xreport/paper.md9
1 files changed, 5 insertions, 4 deletions
diff --git a/report/paper.md b/report/paper.md
index 961937b..71b4f40 100755
--- a/report/paper.md
+++ b/report/paper.md
@@ -432,7 +432,7 @@ The optimal number of constant and random eigenvectors to use is therefore an in
The optimal randomness after doing an exhaustive search as seen on figure \label{fig:opti-rand}peaks at
95 randomised eigenvectors out of 155 total eigenvectors, or 60 static and 95 random eigenvectors. The values of $M_{\textrm{lda}}$ in the figures is the maximum of 51.
-The red peaks on the 3d-plot represent the proportion of randomised eigenvectors which achieve the optimal accuracy, which have been further plotted in figure \ref{opt-2d}. We found that for our data, the optimal ratio of random eigenvectors for a given $M$ is between $0.6$ and $0.9$.
+The red peaks on the 3d-plot represent the proportion of randomised eigenvectors which achieve the optimal accuracy, which have been further plotted in figure \ref{fig:opt-2d}. We found that for our data, the optimal ratio of random eigenvectors for a given $M$ is between $0.6$ and $0.9$.
\begin{figure}
\begin{center}
@@ -454,6 +454,7 @@ The red peaks on the 3d-plot represent the proportion of randomised eigenvectors
\end{figure}
We can compute an ensemble confusion matrix before the committee machines as shown in figure \ref{fig:ens-cm}. This confusion matrix combines the output of all the models in the ensemble. As can be seen from the figure, models in the ensemble usually make more mistakes than an individual model. When the ensemble size is large enough, the errors are rectified by the committee machine, resulting in low error as observed in figure \ref{fig:random-e}.
+
## Comparison
Combining bagging and feature space randomization we are able to consistently achieve higher test accuracy than the individual models. In table \ref{tab:compare} $70/30$ splits.
@@ -478,8 +479,8 @@ Seed & Individual$(M=120)$ & Bag + Feature Ens.$(M=60+95)$\\ \hline
\begin{table}[ht]
\centering
-\begin{tabular}[t]{cc}
-PCA &Fast PCA\\
+\begin{tabular}[t]{cc} \hline
+PCA &Fast PCA\\ \hline
2.9755E+05 &2.9828E+05\\
1.4873E+05 &1.4856E+05\\
1.2286E+05 &1.2259E+05\\
@@ -489,7 +490,7 @@ PCA &Fast PCA\\
3.7118E+04 &3.7030E+04\\
3.2101E+04 &3.2046E+04\\
2.7871E+04 &2.7814E+04\\
-2.4396E+04 &2.4339E+04\\
+2.4396E+04 &2.4339E+04\\ \hline
\end{tabular}
\caption{Comparison of eigenvalues obtain with the two computation methods}
\label{tab:eigen}