From 42da0129e365e81fac25c19ef0c595236c22160b Mon Sep 17 00:00:00 2001 From: Sean Morgan Date: Wed, 22 May 2019 08:34:43 -0400 Subject: [PATCH 1/3] FIX: update variable name in test --- tensorflow_addons/optimizers/lazy_adam_test.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tensorflow_addons/optimizers/lazy_adam_test.py b/tensorflow_addons/optimizers/lazy_adam_test.py index 9a3adab71d..c0fd2bf311 100644 --- a/tensorflow_addons/optimizers/lazy_adam_test.py +++ b/tensorflow_addons/optimizers/lazy_adam_test.py @@ -212,7 +212,7 @@ def doTestBasic(self, use_callable_params=False): self.evaluate(var0)) self.assertAllCloseAccordingToType(var1_np, self.evaluate(var1)) - self.assertEqual("var0_%d/m:0" % (i,), + self.assertEqual("Adam/var0_%d/m:0" % (i,), opt.get_slot(var0, "m").name) @test_utils.run_in_graph_and_eager_modes(reset_test=True) From 9d301e72fd5de2763e2d440a4e2bfe467054b44d Mon Sep 17 00:00:00 2001 From: Sean Morgan Date: Wed, 22 May 2019 15:19:45 -0400 Subject: [PATCH 2/3] Update optimizer init --- tensorflow_addons/optimizers/lazy_adam.py | 10 ++++++++++ tensorflow_addons/optimizers/lazy_adam_test.py | 2 +- 2 files changed, 11 insertions(+), 1 deletion(-) diff --git a/tensorflow_addons/optimizers/lazy_adam.py b/tensorflow_addons/optimizers/lazy_adam.py index 4efe4416b3..3beb0e900b 100644 --- a/tensorflow_addons/optimizers/lazy_adam.py +++ b/tensorflow_addons/optimizers/lazy_adam.py @@ -47,6 +47,16 @@ class LazyAdam(tf.keras.optimizers.Adam): False. """ + def __init__(self, + learning_rate=0.001, + beta_1=0.9, + beta_2=0.999, + epsilon=1e-7, + amsgrad=False, + name='LazyAdam', + **kwargs): + super(LazyAdam, self).__init__(name=name, **kwargs) + def _resource_apply_sparse(self, grad, var, indices): var_dtype = var.dtype.base_dtype lr_t = self._decayed_lr(var_dtype) diff --git a/tensorflow_addons/optimizers/lazy_adam_test.py b/tensorflow_addons/optimizers/lazy_adam_test.py index c0fd2bf311..cea6484df5 100644 --- a/tensorflow_addons/optimizers/lazy_adam_test.py +++ b/tensorflow_addons/optimizers/lazy_adam_test.py @@ -212,7 +212,7 @@ def doTestBasic(self, use_callable_params=False): self.evaluate(var0)) self.assertAllCloseAccordingToType(var1_np, self.evaluate(var1)) - self.assertEqual("Adam/var0_%d/m:0" % (i,), + self.assertEqual("LazyAdam/var0_%d/m:0" % (i,), opt.get_slot(var0, "m").name) @test_utils.run_in_graph_and_eager_modes(reset_test=True) From b9cc08c2649c8e6b5ef1400561620690b689d7d2 Mon Sep 17 00:00:00 2001 From: Sean Morgan Date: Wed, 22 May 2019 15:56:58 -0400 Subject: [PATCH 3/3] Update params and lint --- tensorflow_addons/optimizers/lazy_adam.py | 9 ++++++++- 1 file changed, 8 insertions(+), 1 deletion(-) diff --git a/tensorflow_addons/optimizers/lazy_adam.py b/tensorflow_addons/optimizers/lazy_adam.py index 3beb0e900b..494d108732 100644 --- a/tensorflow_addons/optimizers/lazy_adam.py +++ b/tensorflow_addons/optimizers/lazy_adam.py @@ -55,7 +55,14 @@ def __init__(self, amsgrad=False, name='LazyAdam', **kwargs): - super(LazyAdam, self).__init__(name=name, **kwargs) + super(LazyAdam, self).__init__( + learning_rate=learning_rate, + beta_1=beta_1, + beta_2=beta_2, + epsilon=epsilon, + amsgrad=amsgrad, + name=name, + **kwargs) def _resource_apply_sparse(self, grad, var, indices): var_dtype = var.dtype.base_dtype