public interface Trainable
Modifier and Type | Method and Description |
---|---|
TrainingConfig |
getConfig() |
INDArray |
getGradientsViewArray() |
long |
numParams() |
INDArray |
params() |
Map<String,INDArray> |
paramTable(boolean backpropOnly) |
boolean |
updaterDivideByMinibatch(String paramName)
DL4J layers typically produce the sum of the gradients during the backward pass for each layer, and if required
(if minibatch=true) then divide by the minibatch size.
However, there are some exceptions, such as the batch norm mean/variance estimate parameters: these "gradients" are actually not gradients, but are updates to be applied directly to the parameter vector. |
TrainingConfig getConfig()
long numParams()
INDArray params()
Map<String,INDArray> paramTable(boolean backpropOnly)
backpropOnly
- If true: return only parameters that are not exclusively used for layerwise pretrainingboolean updaterDivideByMinibatch(String paramName)
paramName
- Name of the parameterINDArray getGradientsViewArray()
Copyright © 2021. All rights reserved.