From a1f5db1cd15800175eb0d20e8f044bab5724cb29 Mon Sep 17 00:00:00 2001 From: Vasil Zlatanov Date: Sun, 10 Mar 2019 19:47:11 +0000 Subject: Add part 5 for report --- report/bibliography.bib | 7 +++++++ 1 file changed, 7 insertions(+) (limited to 'report/bibliography.bib') diff --git a/report/bibliography.bib b/report/bibliography.bib index 8230369..0defd2d 100644 --- a/report/bibliography.bib +++ b/report/bibliography.bib @@ -1,3 +1,10 @@ +@misc{inception-note, +Author = {Shane Barratt and Rishi Sharma}, +Title = {A Note on the Inception Score}, +Year = {2018}, +Eprint = {arXiv:1801.01973}, +} + @inproceedings{km-complexity, author = {Inaba, Mary and Katoh, Naoki and Imai, Hiroshi}, title = {Applications of Weighted Voronoi Diagrams and Randomization to Variance-based K-clustering: (Extended Abstract)}, -- cgit v1.2.3-54-g00ecf From f602d1d2488ad249bafab18b0d55f6a5436f32f4 Mon Sep 17 00:00:00 2001 From: Vasil Zlatanov Date: Mon, 11 Mar 2019 14:47:24 +0000 Subject: Add some references and descriptoins --- report/bibliography.bib | 7 +++++++ report/paper.md | 38 +++++++++++++++----------------------- 2 files changed, 22 insertions(+), 23 deletions(-) (limited to 'report/bibliography.bib') diff --git a/report/bibliography.bib b/report/bibliography.bib index 0defd2d..3ccece5 100644 --- a/report/bibliography.bib +++ b/report/bibliography.bib @@ -1,3 +1,10 @@ +@misc{improved, +Author = {Tim Salimans and Ian Goodfellow and Wojciech Zaremba and Vicki Cheung and Alec Radford and Xi Chen}, +Title = {Improved Techniques for Training GANs}, +Year = {2016}, +Eprint = {arXiv:1606.03498}, +} + @misc{inception-note, Author = {Shane Barratt and Rishi Sharma}, Title = {A Note on the Inception Score}, diff --git a/report/paper.md b/report/paper.md index dc9f95a..34e9b6a 100644 --- a/report/paper.md +++ b/report/paper.md @@ -86,16 +86,16 @@ While training the different proposed DCGAN architectures, we did not observe mo ## CGAN Architecture description -CGAN is a conditional version foa Generative adversarial network which utilises labeled data. Unlike DCGAN, CGAN is trained with explicitly provided labels which allows CGAN to associate features with specific labels. This has the intrinsic advantage of allowing us to specify the label of generated data. The baseline CGAN which we evaluate is visible in figure \ref{fig:cganrc}. The baseline GAN arhitecture presents a series blocks each contained a dense layer, ReLu layer and a Batch Normalisation layer. The baseline discriminator use Dense layers, followed by ReLu and a Droupout layer. +CGAN is a conditional version of a GAN which utilises labeled data. Unlike DCGAN, CGAN is trained with explicitly provided labels which allow CGAN to associate features with specific labels. This has the intrinsic advantage of allowing us to specify the label of generated data. The baseline CGAN which we evaluate is visible in figure \ref{fig:cganrc}. The baseline GAN arhitecture presents a series blocks each contained a dense layer, ReLu layer and a Batch Normalisation layer. The baseline discriminator uses Dense layers, followed by ReLu and a Droupout layer. We evaluate permutations of the architecture involving: -* Shallow CGAN -* Deep CGAN -* Deep Convolutional GAN -* Label Smoothing (One Sided) -* Various Dropout -* Virtual Batch Normalisation +* Shallow CGAN - 1 Dense-ReLu-BN block +* Deep CGAN - 5 Dense-ReLu-BN +* Deep Convolutional GAN - DCGAN + conditional label input +* Label Smoothing (One Sided) - Truth labels to 0 and $1-\alpha$ (0.9) +* Various Dropout - Use 0.1 and 0.5 Dropout parameters +* Virtual Batch Normalisation - Normalisation based on one batch [@improved] \begin{figure} \begin{center} @@ -143,24 +143,13 @@ The effect of dropout for the non-convolutional CGAN architecture does not affec # Inception Score -## Classifier Architecture Used - -## Results - -Measure the inception scores i.e. we use the class labels to -generate images in CGAN and compare them with the predicted labels of the generated images. - -Also report the recognition accuracies on the -MNIST real testing set (10K), in comparison to the inception scores. - -**Please measure and discuss the inception scores for the different hyper-parameters/tricks and/or -architectures in Q2.** - -We measure the performance of the considered GAN's using the Inecption score [-inception], as calculated -with L2-Net logits. +Inception score is calculated as introduced by Tim Salimans et. al [@improved]. However as we are evaluating MNIST, we use LeNet as the basis of the inceptioen score. +Inception score is calculated with the logits of the LeNet $$ \textrm{IS}(x) = \exp(\mathbb{E}_x \left( \textrm{KL} ( p(y\mid x) \| p(y) ) \right) ) $$ +## Classifier Architecture Used + \begin{table}[] \begin{tabular}{llll} & Accuracy & Inception Sc. & GAN Tr. Time \\ \hline @@ -174,10 +163,13 @@ Medium CGAN DO=0.1 & 0.761 & 3.836 & 10:36 \\ Medium CGAN DO=0.5 & 0.725 & 3.677 & 10:36 \\ Medium CGAN+VBN & ? & ? & ? \\ Medium CGAN+VBN+LS & ? & ? & ? \\ -*MNIST original & 0.9846 & 9.685 & N/A +*MNIST original & 0.9846 & 9.685 & N/A \\ \hline \end{tabular} \end{table} + +**Please measure and discuss the inception scores for the different hyper-parameters/tricks and/or + # Re-training the handwritten digit classifier ## Results -- cgit v1.2.3-54-g00ecf From 9eeba3d77e0dbd1610213c2857bc32fb3187db28 Mon Sep 17 00:00:00 2001 From: Vasil Zlatanov Date: Mon, 11 Mar 2019 17:40:58 +0000 Subject: Add lenet bibtex --- report/bibliography.bib | 8 ++++++++ report/paper.md | 4 +++- 2 files changed, 11 insertions(+), 1 deletion(-) (limited to 'report/bibliography.bib') diff --git a/report/bibliography.bib b/report/bibliography.bib index 3ccece5..430d8b5 100644 --- a/report/bibliography.bib +++ b/report/bibliography.bib @@ -1,3 +1,11 @@ +@INPROCEEDINGS{lenet, + author = {Yann Lecun and Léon Bottou and Yoshua Bengio and Patrick Haffner}, + title = {Gradient-based learning applied to document recognition}, + booktitle = {Proceedings of the IEEE}, + year = {1998}, + pages = {2278--2324} +} + @misc{improved, Author = {Tim Salimans and Ian Goodfellow and Wojciech Zaremba and Vicki Cheung and Alec Radford and Xi Chen}, Title = {Improved Techniques for Training GANs}, diff --git a/report/paper.md b/report/paper.md index 522eaed..1989472 100644 --- a/report/paper.md +++ b/report/paper.md @@ -126,11 +126,13 @@ We evaluate permutations of the architecture involving: ### Inception Score -Inception score is calculated as introduced by Tim Salimans et. al [@improved]. However as we are evaluating MNIST, we use LeNet as the basis of the inceptioen score. +Inception score is calculated as introduced by Tim Salimans et. al [@improved]. However as we are evaluating MNIST, we use LeNet-5 [@lenet] as the basis of the inceptioen score. We use the logits extracted from LeNet: $$ \textrm{IS}(x) = \exp(\mathbb{E}_x \left( \textrm{KL} ( p(y\mid x) \| p(y) ) \right) ) $$ +We further report the classification accuracy as found with LeNet. + ### Classifier Architecture Used \begin{table}[] -- cgit v1.2.3-54-g00ecf