From e3e713a66b0a1e85714d764663823c92ffbd1f2d Mon Sep 17 00:00:00 2001 From: nunzip Date: Fri, 15 Feb 2019 16:59:27 +0000 Subject: Section II grammar fix --- report/paper.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/report/paper.md b/report/paper.md index e44444b..885f27d 100644 --- a/report/paper.md +++ b/report/paper.md @@ -42,7 +42,7 @@ Initializing K-means is an expensive process, based on sequential attempts of ce # RF classifier -We use a random forest classifier to label images based on the bag-of-words histograms. Random forests are an ensemble of randomly generated decision trees, who's performance depends on the ensemble size, tree depth, randomness and weak learner used. +We use a random forest classifier to label images based on the bag-of-words histograms. Random forests are an ensemble of randomly generated decision trees, whose performance depends on the ensemble size, tree depth, randomness and weak learner used. ## Hyperparameters tuning @@ -59,7 +59,7 @@ We expect a large tree depth to lead into overfitting. However for the data anal \end{center} \end{figure} -Random forests will select a random number of features on which to apply a weak learner (such as axis aligned split) and then choose the best feature of the sampled ones to perform the split on, based on a given criteria (our results use the *Gini index*). The fewer features that are compared for each split the quicker the trees are built and the more random they are. Therefore the randomness parameter can be considered as the number of features used when making splits. We evaluate accuracy given different randomness when using a K-means vocabulary of size 100 in figure \ref{fig:kmeanrandom}. The results in the figure \ref{fig:kmeanrandom} use a forest size of 100 as we infered that this is the estimatator count for which performance gains tend to plateau (when selecting $\sqrt{n}$ random features). +Random forests will select a random number of features on which to apply a weak learner (such as axis aligned split) and then choose the best feature of the sampled ones to perform the split on, based on a given criteria (our results use the *Gini index*). The fewer features that are compared for each split the quicker the trees are built and the more random they are. Therefore the randomness parameter can be considered as the number of features used when making splits. We evaluate accuracy given different randomness when using a K-means vocabulary of size 100 in figure \ref{fig:kmeanrandom}. The results in the figure \ref{fig:kmeanrandom} also use a forest size of 100 as we infered that this is the estimatator count for which performance gains tend to plateau (when selecting $\sqrt{n}$ random features). This parameter also affects correlation between trees. We expect trees to be more correlated when using a large number of features for splits. \begin{figure} -- cgit v1.2.3 From d2f8f7376a4a785f11f062dfd81ba83b9fb83cd3 Mon Sep 17 00:00:00 2001 From: nunzip Date: Fri, 15 Feb 2019 17:41:16 +0000 Subject: Rewrite comparison --- report/paper.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/report/paper.md b/report/paper.md index 885f27d..36259c6 100644 --- a/report/paper.md +++ b/report/paper.md @@ -148,7 +148,7 @@ In many applications the increase in training time would not justify the small i For the Caltech_101 dataset, a RF codebook seems to be the most suitable method to perform RF classification. -The `water_lilly` is the most misclassified class, both for K-means and RF codebook (refer to figures \ref{fig:km_cm} and \ref{fig:p3_cm}). This indicates that the features obtained from the class do not provide for very discriminative splits, resulting in the prioritsation of other features in the first nodes of the decision trees. +The `water_lilly` is the most misclassified class, both for K-means and RF codebook (refer to figures \ref{fig:km_cm} and \ref{fig:p3_cm}). This indicates that the quantised descriptors obtained from the class do not provide for very discriminative splits, resulting in the prioritsation of other features in the first nodes of the decision trees. # References -- cgit v1.2.3