From 2685022dbcaeeafdbb0025a8737edbaa5eeab425 Mon Sep 17 00:00:00 2001 From: nunzip Date: Thu, 13 Dec 2018 23:21:00 +0000 Subject: Fix repetitions and grammar mistakes --- report/metadata.yaml | 8 +++++--- report/paper.md | 39 +++++++++++++++++++++++++++++---------- 2 files changed, 34 insertions(+), 13 deletions(-) diff --git a/report/metadata.yaml b/report/metadata.yaml index e4d4470..74732a6 100755 --- a/report/metadata.yaml +++ b/report/metadata.yaml @@ -13,9 +13,11 @@ nocite: | abstract: | This report analyses distance metrics learning techniques with regards to identification accuracy for the dataset CUHK03. The baseline method used for - identification is Eucdidian based Nearest Neighbors based on Euclidean distance. + identification is Nearest Neighbors based on Euclidean distance. The improved approach evaluated utilises Jaccardian metrics to rearrange the NN - ranklist based on reciprocal neighbours. While this approach is more complex and introduced new hyperparameter, significant accuracy improvements are observed - - approximately 10% increased Top-1 identifications, and good improvements for Top-$N$ accuracy with low $N$. + ranklist based on reciprocal neighbours. While this approach is more complex and introduces new hyperparameters, + significant accuracy improvements are observed - + approximately 10% higher Top-1 identification, and good improvements for Top-$N$ + accuracy with low $N$. ... diff --git a/report/paper.md b/report/paper.md index 73c49de..83cde10 100755 --- a/report/paper.md +++ b/report/paper.md @@ -5,9 +5,9 @@ The person re-identification problem presented in this paper requires matching pedestrian images from disjoint cameras by pedestrian detectors. This problem is challenging, as identities captured in photos are subject to various lighting, pose, -blur, background and oclusion from various camera views. This report considers +blur, background and occlusion from various camera views. This report considers features extracted from the CUHK03 dataset, following a 50 layer Residual network -(Resnet50). This paper considers distance metrics techniques which can be used to +(ResNet50). This paper considers distance metrics techniques which can be used to perform person re-identification across *disjoint* cameras, using these features. Features extracted from Neural Networks such as ResNet-50 are already highly processed @@ -72,9 +72,9 @@ Magnitude normalization is a common technique, used to equalize feature importan Applying magnitude normalization (scaling feature vectors to unit length) had a negative effect re-identification. Furthemore standartization by removing feature mean and deviation also had negative effect on performance as seen on figure \ref{fig:baselineacc}. This may -be due to the fact that we are removing feature scaling that was introduced by the Neural network, +be due to the fact that we are removing feature scaling that was introduced by the Neural Network, such that some of the features are more significant than others. By standartizing our -features at this point, we remove such scaling and may be losing using useful metrics. +features at this point, we remove such scaling and may be losing useful information. ## kMeans Clustering @@ -88,14 +88,14 @@ This method did not bring any major improvement to the baseline, as it can be se figure \ref{fig:baselineacc}. It is noticeable how the number of clusters affects performance, showing better identification accuracy for a number of clusters away from the local minimum achieved at 60 clusters (figure \ref{fig:kmeans}). This trend can likely -be explained by the number of distance comparison's performed. +be explained by the number of distance comparisons performed. We would expect clustering with $k=1$ and $k=\textrm{label count}$ to have the same performance -the baseline approach without clustering, as we are performing the same number of comparisons. +of the baseline approach without clustering, as we are performing the same number of comparisons. -Clustering is a great method of reducing computation time. Assuming 39 clusters of 39 neighbours -we would be performing only 78 distance computation for a gallery size of 1487, instead of the -original 1487. This however comes at the cost of ignoring neighbours from other clusters which may +Clustering is a great method of reducing computation time. Assuming 38 clusters of 38 neighbours +we would be performing only 76 distance computations for a gallery size of 1467, instead of the +original 1467. This however comes at the cost of ignoring neighbours from other clusters which may be closer. Since clusters do not necessarily have the same number of datapoints inside them (sizes are uneven), we find that the lowest average number of comparison happens at around 60 clusters, which also appears to be the worst performing number of clusters. @@ -268,7 +268,26 @@ as a result despite the identifiable feature being hidden in the query. An examp \end{figure} -# Conclusion +# Conclusions + +Overall the reranking method gives a significant improvement to both top $k$ accuracy and mean average precision. +The cost of this operation is an increase in computation time due to the change in complexity from +the baseline, summarized in table \ref{tab:complexity}. + +\begin{table}[] +\centering +\begin{tabular}{ll} +\hline +\textbf{Process} & \textbf{Complexity} \\ \hline +Rerank - Distances calculation & \textit{O(N\textsuperscript{2})} \\ +Rerank - Ranking & \textit{O(N\textsuperscript{2}logN)} \\ +Baseline - Distances calculation & \textit{O(N\textsuperscript{2})} \\ +Baseline - Ranking & \textit{O(N\textsuperscript{2})} \\ +\end{tabular} +\caption{Complexity evaluation} +\label{tab:complexity} +\end{table} + # References -- cgit v1.2.3