public class SoftmaxBp extends DynamicCustomOp
DynamicCustomOp.DynamicCustomOpsBuilderaxis, bArguments, dArguments, iArguments, inplaceCall, inputArguments, outputArguments, outputVariables, tArgumentsdimensions, extraArgs, inPlace, ownName, ownNameSetWithDefault, sameDiff, scalarValue| Constructor and Description |
|---|
SoftmaxBp() |
SoftmaxBp(@NonNull INDArray input,
@NonNull INDArray grad,
INDArray output,
Integer dimension) |
SoftmaxBp(@NonNull INDArray input,
@NonNull INDArray grad,
Integer dimension) |
SoftmaxBp(SameDiff sd,
SDVariable input,
SDVariable grad,
Integer dimension) |
| Modifier and Type | Method and Description |
|---|---|
List<DataType> |
calculateOutputDataTypes(List<DataType> dataTypes)
Calculate the data types for the output arrays.
|
List<SDVariable> |
doDiff(List<SDVariable> grad)
The actual implementation for automatic differentiation.
|
String |
opName()
This method returns op opName as string
|
addBArgument, addDArgument, addIArgument, addIArgument, addInputArgument, addOutputArgument, addTArgument, assertValidForExecution, bArgs, builder, calculateOutputShape, calculateOutputShape, clearArrays, dArgs, getBArgument, getDescriptor, getIArgument, getInputArgument, getOutputArgument, getTArgument, iArgs, initFromOnnx, initFromTensorFlow, inputArguments, numBArguments, numDArguments, numIArguments, numInputArguments, numOutputArguments, numTArguments, onnxName, opHash, opNum, opType, outputArguments, outputVariables, outputVariables, removeIArgument, removeInputArgument, removeOutputArgument, removeTArgument, setInputArgument, setInputArguments, setOutputArgument, tArgs, tensorflowName, toString, wrapFilterNull, wrapOrNull, wrapOrNullarg, arg, argNames, args, attributeAdaptersForFunction, configFieldName, diff, dup, equals, getNumOutputs, getValue, hashCode, isConfigProperties, larg, mappingsForFunction, onnxNames, outputs, outputVariable, outputVariablesNames, propertiesForFunction, rarg, replaceArg, setInstanceId, setPropertiesForFunction, setValueFor, tensorflowNamesclone, finalize, getClass, notify, notifyAll, wait, wait, waitisInplaceCallpublic SoftmaxBp()
public SoftmaxBp(SameDiff sd, SDVariable input, SDVariable grad, Integer dimension)
public SoftmaxBp(@NonNull
@NonNull INDArray input,
@NonNull
@NonNull INDArray grad,
Integer dimension)
public String opName()
DynamicCustomOpopName in interface CustomOpopName in class DynamicCustomOppublic List<SDVariable> doDiff(List<SDVariable> grad)
DifferentialFunctiondoDiff in class DynamicCustomOppublic List<DataType> calculateOutputDataTypes(List<DataType> dataTypes)
DifferentialFunctionDifferentialFunction.calculateOutputShape(), this method differs in that it does not
require the input arrays to be populated.
This is important as it allows us to do greedy datatype inference for the entire net - even if arrays are not
available.calculateOutputDataTypes in class DifferentialFunctiondataTypes - The data types of the inputsCopyright © 2021. All rights reserved.