aboutsummaryrefslogtreecommitdiff
path: root/dcgan.py
diff options
context:
space:
mode:
authornunzip <np.scarh@gmail.com>2019-03-07 20:02:21 +0000
committernunzip <np.scarh@gmail.com>2019-03-07 20:02:21 +0000
commit5c5680085208f941511c9d74447c391124e38f9e (patch)
tree574b0e5fd0db3427bd0423d3fe76626ae6280f4b /dcgan.py
parent7b288397a51633c878631c0e6d48c96ffce09a84 (diff)
downloade4-gan-5c5680085208f941511c9d74447c391124e38f9e.tar.gz
e4-gan-5c5680085208f941511c9d74447c391124e38f9e.tar.bz2
e4-gan-5c5680085208f941511c9d74447c391124e38f9e.zip
variable dropout
Diffstat (limited to 'dcgan.py')
-rw-r--r--dcgan.py11
1 files changed, 6 insertions, 5 deletions
diff --git a/dcgan.py b/dcgan.py
index 347f61e..4ee5917 100644
--- a/dcgan.py
+++ b/dcgan.py
@@ -17,7 +17,7 @@ import sys
import numpy as np
class DCGAN():
- def __init__(self, conv_layers = 1):
+ def __init__(self, conv_layers = 1, dropout=0.25):
# Input shape
self.img_rows = 28
self.img_cols = 28
@@ -25,6 +25,7 @@ class DCGAN():
self.img_shape = (self.img_rows, self.img_cols, self.channels)
self.latent_dim = 100
self.conv_layers = conv_layers
+ self.dropout = dropout
optimizer = Adam(0.002, 0.5)
@@ -88,20 +89,20 @@ class DCGAN():
model.add(Conv2D(32, kernel_size=3, strides=2, input_shape=self.img_shape, padding="same"))
model.add(LeakyReLU(alpha=0.2))
- model.add(Dropout(0.25))
+ model.add(Dropout(self.dropout))
model.add(Conv2D(64, kernel_size=3, strides=2, padding="same"))
model.add(ZeroPadding2D(padding=((0,1),(0,1))))
model.add(BatchNormalization())
model.add(LeakyReLU(alpha=0.2))
- model.add(Dropout(0.25))
+ model.add(Dropout(dropout))
model.add(Conv2D(128, kernel_size=3, strides=2, padding="same"))
model.add(BatchNormalization())
model.add(LeakyReLU(alpha=0.2))
- model.add(Dropout(0.25))
+ model.add(Dropout(dropout))
model.add(Conv2D(256, kernel_size=3, strides=1, padding="same"))
model.add(BatchNormalization())
model.add(LeakyReLU(alpha=0.2))
- model.add(Dropout(0.25))
+ model.add(Dropout(dropout))
model.add(Flatten())
model.add(Dense(1, activation='sigmoid'))