diff options
-rwxr-xr-x | report/paper.md | 17 |
1 files changed, 10 insertions, 7 deletions
diff --git a/report/paper.md b/report/paper.md index a20e943..89e260c 100755 --- a/report/paper.md +++ b/report/paper.md @@ -113,16 +113,16 @@ improve identification accuracy, and consider it an additional baseline. ## Mahalanobis Distance -We were not able to achieve significant improvements using mahalanobis for +We were not able to achieve significant improvements using Mahalanobis for original distance ranking compared to square euclidiaen metrics. -The mahalanobis distance metric was used to create the ranklist as an alternative to euclidean distance: +The Mahalanobis distance metric was used to create the ranklist as an alternative to euclidean distance: $$ d_M(p,g_i) = (p-g_i)^TM(p-g_i). $$ -When performing mahalanobis with the covariance matrix $M$ generated from the training set, reported accuracy is reduced to **38%** . +When performing Mahalanobis with the covariance matrix $M$ generated from the training set, reported accuracy is reduced to **38%** . -We also attempted to perform the same mahalanobis metric on a reduced PCA featureset. This allowed for significant execution +We also attempted to perform the same Mahalanobis metric on a reduced PCA featureset. This allowed for significant execution time improvements due to the greatly reduced computation requierments for smaller featurespace, but nevertheless demonstrated no improvements over an euclidean metric. @@ -140,7 +140,7 @@ transformations performed the the ResNet-50 convolution model the features were \end{center} \end{figure} -While we did not use mahalanobis as a primary distance metric, it is possible to use the Mahalanobis metric, together with the next investigated solution involving $k$-reciprocal re-ranking. +While we did not use Mahalanobis as a primary distance metric, it is possible to use the Mahalanobis metric, together with the next investigated solution involving $k$-reciprocal re-ranking. # Suggested Improvement @@ -193,12 +193,13 @@ be defined as $k_1$: $R^*(g_i,k_1)$. The distances obtained are then mixed, obtaining a final distance $d^*(p,g_i)$ that is used to obtain the improved rank-list: $d^*(p,g_i)=(1-\lambda)d_J(p,g_i)+\lambda d(p,g_i)$. +## Optimisation The aim is to learn optimal values for $k_1,k_2$ and $\lambda$ in the training set that improve top1 identification accuracy. This is done through a simple multi-direction search algorithm followed by exhaustive search to estimate $k_{1_{opt}}$ and $k_{2_{opt}}$ for eleven values of $\lambda$ from zero (only Jaccard distance) to one (only original distance) in steps of 0.1. The results obtained through this approach suggest: $k_{1_{opt}}=9, k_{2_{opt}}=3, 0.1\leq\lambda_{opt}\leq 0.3$. -It is possible to verify that the optimization of $k_{1_{opt}}$, $k_{2_{opt}}$ and $\lambda$ +It is possible to verify that the optimisation of $k_{1_{opt}}$, $k_{2_{opt}}$ and $\lambda$ has been successful. Figures \ref{fig:pqvals} and \ref{fig:lambda} show that the optimal values obtained from training are close to the ones for the local maximum of gallery and query. @@ -220,7 +221,6 @@ training are close to the ones for the local maximum of gallery and query. \end{center} \end{figure} - ## $k$-reciprocal Re-ranking Evaluation Re-ranking achieves better results than the other baseline methods analyzed both as top $k$ @@ -252,6 +252,9 @@ The difference between the top $k$ accuracies of the two methods gets smaller as \end{center} \end{figure} +The improved results due to $k$-reciprocal re-ranking can be explained by considering...re-ranking can be explained by considering... + + # Conclusion # References |