aboutsummaryrefslogtreecommitdiff
path: root/report
diff options
context:
space:
mode:
authornunzip <np.scarh@gmail.com>2019-02-12 20:49:28 +0000
committernunzip <np.scarh@gmail.com>2019-02-12 20:49:28 +0000
commit2503cfe68a47131c364edec96ac66dd00ef928f7 (patch)
tree7bdb282102082accda676c538087d07895a5cf7b /report
parent7150ceab64658c2154b3b3f73fc99fdba27b2473 (diff)
downloade4-vision-2503cfe68a47131c364edec96ac66dd00ef928f7.tar.gz
e4-vision-2503cfe68a47131c364edec96ac66dd00ef928f7.tar.bz2
e4-vision-2503cfe68a47131c364edec96ac66dd00ef928f7.zip
Small corrections
Diffstat (limited to 'report')
-rw-r--r--report/paper.md6
1 files changed, 4 insertions, 2 deletions
diff --git a/report/paper.md b/report/paper.md
index 4ab5924..f7d623a 100644
--- a/report/paper.md
+++ b/report/paper.md
@@ -6,7 +6,9 @@ A common technique for codebook generation involves utilising K-means clustering
image descriptors. In this way descriptors may be mapped to *visual* words which lend themselves to
binning and therefore the creation of bag-of-words histograms for the use of classification.
-In this courseworok 100-thousand random SIFT descriptors of the Caltech_101 dataset are used to build the K-means visual vocabulary.
+In this courseworok 100-thousand random SIFT descriptors (size=128) of the Caltech_101 dataset are used to build the K-means visual vocabulary.
+
+Both training and testing use 15 randomly selected images from the 10 available classess.
## Vocabulary size
@@ -172,7 +174,7 @@ and in many cases the increase in training time would not justify the minimum in
For Caltech_101 RF-codebook seems to be the most suitable method to perform RF-classification.
It is observable that for the particular dataset we are analysing the class *water_lilly*
-is the one that gets misclassified the most, both in k-means and RF-codebook (refer to figures \ref{fig:km_cm} and \ref{fig:p3_cm}. This means that the features obtained
+is the one that gets misclassified the most, both in k-means and RF-codebook (refer to figures \ref{fig:km_cm} and \ref{fig:p3_cm}). This means that the features obtained
from this class do not guarantee very discriminative splits, hence the first splits in the trees
will prioritize features taken from other classes.