tf.raw_ops.SparseApplyAdadelta
var: Should be from a Variable().
tf.raw_ops.SparseApplyAdadelta( var, accum, accum_update, lr, rho, epsilon, grad, indices, use_locking=False, name=None )
Args | |
---|---|
var
|
Tensor . Must be one of the following types:
float32 ,
float64 ,
int32 ,
uint8 ,
int16 ,
int8 ,
complex64 ,
int64 ,
qint8 ,
quint8 ,
qint32 ,
bfloat16 ,
uint16 ,
complex128 ,
half ,
uint32 ,
uint64 .
|
accum
|
Tensor . Must have the same type as
var .
|
accum_update
|
Tensor . Must have the same type as
var .
|
lr |
A Tensor . Must have the same type as var . Learning rate. Must be a scalar. |
rho |
A Tensor . Must have the same type as var . Decay factor. Must be a scalar. |
epsilon |
A Tensor . Must have the same type as var . Constant factor. Must be a scalar. |
grad |
A Tensor . Must have the same type as var . The gradient. |
indices |
A Tensor . Must be one of the following types: int32 , int64 . A vector of indices into the first dimension of var and accum. |
use_locking |
An optional bool . Defaults to False . If True, updating of the var and accum tensors will be protected by a lock; otherwise the behavior is undefined, but may exhibit less contention. |
name |
A name for the operation (optional). |
Returns | |
---|---|
A mutable Tensor . Has the same type as var . |
© 2022 The TensorFlow Authors. All rights reserved.
Licensed under the Creative Commons Attribution License 4.0.
Code samples licensed under the Apache 2.0 License.
https://www.tensorflow.org/versions/r2.9/api_docs/python/tf/raw_ops/SparseApplyAdadelta