10 learning_rate: float = 0.001,
13 epsilon: float = 1e-7) ->
None:
15 Initialize the Adam optimizer.
18 learning_rate (float): The learning rate for optimization.
19 beta1 (float): Exponential decay rate for the first moment estimates.
20 beta2 (float): Exponential decay rate for the second moment estimates.
21 epsilon (float): Small constant to prevent division by zero.
24 self.beta1: float = beta1
25 self.beta2: float = beta2
26 self.epsilon: float = epsilon