aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorVasil Zlatanov <v@skozl.com>2019-02-12 17:22:55 +0000
committerVasil Zlatanov <v@skozl.com>2019-02-12 17:22:55 +0000
commit801b30cc67dee6071101320bc9ac1f6edde655f9 (patch)
tree05acc405c7d96c0c380b48bb8403e746da9728e1
parentf8f7f8f692bc960ef5c5be74936a6d1de5a5caac (diff)
downloade4-vision-801b30cc67dee6071101320bc9ac1f6edde655f9.tar.gz
e4-vision-801b30cc67dee6071101320bc9ac1f6edde655f9.tar.bz2
e4-vision-801b30cc67dee6071101320bc9ac1f6edde655f9.zip
Remove word to improve spacing
-rw-r--r--report/paper.md2
1 files changed, 1 insertions, 1 deletions
diff --git a/report/paper.md b/report/paper.md
index 3e5ec48..bae8979 100644
--- a/report/paper.md
+++ b/report/paper.md
@@ -15,7 +15,7 @@ The number of clusters or the number of centroids determine the vocabulary size
## Bag-of-words histogram quantisation of descriptor vectors
-An example histogram for training image shown on figure \ref{fig:histo_tr}, computed with a vocubulary size of 100. A corresponding testing image of the same class is shown in figure \ref{fig:histo_te}. The histograms appear to have similar counts for the same words, demonstrating they had a descriptors which matched the *keywowrds* in similar proportions. We later look at the effect of the vocubalary size (as determined by the number of K-mean centroids) on the classificaiton accuracy in figure \ref{fig:km_vocsize}.
+An example histogram for training image shown on figure \ref{fig:histo_tr}, computed with a vocubulary size of 100. A corresponding testing image of the same class is shown in figure \ref{fig:histo_te}. The histograms appear to have similar counts for the same words, demonstrating they had a descriptors which matched the *keywowrds* in similar proportions. We later look at the effect of the vocubalary size (as determined by the number of centroids) on the classificaiton accuracy in figure \ref{fig:km_vocsize}.
The time complexity of quantisation with a K-means codebooks is $O(n^{dk+1})$ , where n is the number of entities to be clustered, d is the dimension and k is the cluster count [@km-complexity]. As the computation time is high, the tests we use a subsample of descriptors to compute the centroids. An alternative method is NUNZIO PUCCI WRITE HERE