Hide keyboard shortcuts

Hot-keys on this page

r m x p   toggle line displays

j k   next/prev highlighted chunk

0   (zero) top of page

1   (one) first highlighted chunk

1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

33

34

35

36

37

38

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

58

59

60

61

62

63

64

65

66

67

68

69

70

71

72

73

74

75

76

77

78

79

80

81

82

83

84

85

86

87

88

89

90

91

92

93

94

95

96

97

98

99

100

101

102

103

104

105

106

107

108

109

110

111

112

113

114

115

116

117

118

119

120

121

122

123

124

125

126

127

128

129

130

131

132

133

134

135

136

137

138

139

140

141

142

143

144

145

146

147

148

149

150

151

152

153

154

155

156

157

158

159

160

161

162

163

164

import importlib 

import json 

import numpy as np 

import os 

import sys 

import time 

import uuid 

import copy 

 

from hypergan.discriminators import * 

from hypergan.encoders import * 

from hypergan.generators import * 

from hypergan.inputs import * 

from hypergan.samplers import * 

from hypergan.trainers import * 

 

import hyperchamber as hc 

from hyperchamber import Config 

from hypergan.ops import TensorflowOps 

import tensorflow as tf 

import hypergan as hg 

 

from hypergan.gan_component import ValidationException, GANComponent 

from .base_gan import BaseGAN 

 

from hypergan.discriminators.fully_connected_discriminator import FullyConnectedDiscriminator 

from hypergan.encoders.uniform_encoder import UniformEncoder 

from hypergan.trainers.multi_step_trainer import MultiStepTrainer 

 

class AlphaGAN(BaseGAN): 

"""  

""" 

def __init__(self, *args, **kwargs): 

BaseGAN.__init__(self, *args, **kwargs) 

self.discriminator = None 

self.encoder = None 

self.generator = None 

self.loss = None 

self.trainer = None 

self.session = None 

 

def required(self): 

return "generator discriminator z_discriminator g_encoder".split() 

 

def create(self): 

BaseGAN.create(self) 

if self.session is None: 

self.session = self.ops.new_session(self.ops_config) 

with tf.device(self.device): 

self.inputs.x = tf.identity(self.inputs.x, name='input') 

config = self.config 

ops = self.ops 

 

g_encoder = dict(config.g_encoder or config.discriminator) 

encoder = self.create_component(g_encoder) 

encoder.ops.describe("g_encoder") 

encoder.create(self.inputs.x) 

encoder.z = tf.zeros(0) 

if(len(encoder.sample.get_shape()) == 2): 

s = ops.shape(encoder.sample) 

encoder.sample = tf.reshape(encoder.sample, [s[0],s[1], 1, 1]) 

 

z_discriminator = dict(config.z_discriminator or config.discriminator) 

z_discriminator['layer_filter']=None 

 

encoder_discriminator = self.create_component(z_discriminator) 

encoder_discriminator.ops.describe("z_discriminator") 

standard_discriminator = self.create_component(config.discriminator) 

standard_discriminator.ops.describe("discriminator") 

 

#encoder.sample = ops.reshape(encoder.sample, [ops.shape(encoder.sample)[0], -1]) 

uniform_encoder_config = config.encoder 

z_size = 1 

for size in ops.shape(encoder.sample)[1:]: 

z_size *= size 

uniform_encoder_config.z = z_size // len(uniform_encoder_config.projections) 

uniform_encoder = UniformEncoder(self, uniform_encoder_config) 

uniform_encoder.create() 

 

self.generator = self.create_component(config.generator) 

 

direction = tf.random_normal(ops.shape(uniform_encoder.sample), stddev=0.3, name='direction') 

slider = tf.get_variable('slider', initializer=tf.constant_initializer(0.0), shape=[1, 1], dtype=tf.float32, trainable=False) 

x = self.inputs.x 

 

# project the output of the autoencoder 

z_hat = encoder.sample 

 

z = uniform_encoder.sample + slider * direction 

z = ops.reshape(z, ops.shape(z_hat)) 

# end encoding 

 

g = self.generator.create(z) 

sample = self.generator.sample 

self.uniform_sample = g 

x_hat = self.generator.reuse(z_hat) 

 

encoder_discriminator.create(x=z, g=z_hat) 

 

eloss = dict(config.loss) 

eloss['gradient_penalty'] = False 

eloss['gradient_locally_stable'] = False 

encoder_loss = self.create_component(eloss, discriminator = encoder_discriminator) 

encoder_loss.create() 

 

stacked_xg = ops.concat([x, x_hat, g], axis=0) 

standard_discriminator.create(stacked_xg) 

 

standard_loss = self.create_component(config.loss, discriminator = standard_discriminator) 

standard_loss.create(split=3) 

 

self.trainer = self.create_component(config.trainer) 

 

#loss terms 

distance = config.distance or ops.lookup('l1_distance') 

cycloss = tf.reduce_mean(distance(self.inputs.x,x_hat)) 

cycloss_lambda = config.cycloss_lambda 

if cycloss_lambda is None: 

cycloss_lambda = 10 

cycloss *= cycloss_lambda 

loss1=('generator', cycloss + encoder_loss.g_loss) 

loss2=('generator', cycloss + standard_loss.g_loss) 

loss3=('discriminator', standard_loss.d_loss) 

loss4=('discriminator', encoder_loss.d_loss) 

 

var_lists = [] 

var_lists.append(encoder.variables()) 

var_lists.append(self.generator.variables()) 

var_lists.append(standard_discriminator.variables()) 

var_lists.append(encoder_discriminator.variables()) 

 

metrics = [] 

metrics.append(encoder_loss.metrics) 

metrics.append(standard_loss.metrics) 

metrics.append(None) 

metrics.append(None) 

 

# trainer 

 

self.trainer = MultiStepTrainer(self, self.config.trainer, [loss1,loss2,loss3,loss4], var_lists=var_lists, metrics=metrics) 

self.trainer.create() 

 

self.session.run(tf.global_variables_initializer()) 

 

self.encoder = encoder 

self.uniform_encoder = uniform_encoder 

 

self.slider = slider 

self.direction = direction 

 

 

def step(self, feed_dict={}): 

return self.trainer.step(feed_dict) 

 

def input_nodes(self): 

"used in hypergan build" 

#return [] 

return [self.inputs.x, self.uniform_encoder.sample, self.slider, self.direction] 

 

 

def output_nodes(self): 

"used in hypergan build" 

#return [self.uniform_sample] 

return [self.encoder.sample, self.uniform_sample]