diff --git a/keras/src/activations/activations_test.py b/keras/src/activations/activations_test.py index a36bee12861..d88c350b171 100644 --- a/keras/src/activations/activations_test.py +++ b/keras/src/activations/activations_test.py @@ -638,7 +638,7 @@ def celu(x, alpha=1.0): x = np.random.random((2, 5)) result = activations.celu(x[np.newaxis, :], alpha=0.5)[0] - expected = celu(x, True) + expected = celu(x, alpha=0.5) self.assertAllClose(result, expected, rtol=1e-05) def test_glu(self): diff --git a/keras/src/ops/nn.py b/keras/src/ops/nn.py index 954c58a1dfb..d1d1f84817c 100644 --- a/keras/src/ops/nn.py +++ b/keras/src/ops/nn.py @@ -607,7 +607,7 @@ def hard_tanh(x): Example: - >>> x = x = np.array([-2., -1., 0., 1., 2.]) + >>> x = np.array([-2., -1., 0., 1., 2.]) >>> x_hard_tanh = keras.ops.hard_tanh(x) >>> print(x_hard_tanh) array([-1. -1. 0. 1. 1.], shape=(5,), dtype=float64)