Class CuDNNFunctionOptimizations.CudnnConv2dNCHWtoNHWCConversion
- java.lang.Object
-
- org.nd4j.autodiff.samediff.optimize.optimizations.CuDNNFunctionOptimizations.CudnnConv2dNCHWtoNHWCConversion
-
- All Implemented Interfaces:
Optimizer
- Enclosing class:
- CuDNNFunctionOptimizations
public static class CuDNNFunctionOptimizations.CudnnConv2dNCHWtoNHWCConversion extends Object implements Optimizer
https://docs.nvidia.com/deeplearning/sdk/dl-performance-guide/index.html#tensor-layout For tensor cores: we want NHWC layout: Section 7.3.1 "Layout choice has an effect on performance, as convolutions implemented for Tensor Cores require NHWC layout and are fastest when input tensors are laid out in NHWC." "To maximize performance, we recommend using NHWC tensor layout." As for weights format: cuDNN docs are vague - but TF uses NCHW+OIHW or NHWC+OHWI
-
-
Constructor Summary
Constructors Constructor Description CudnnConv2dNCHWtoNHWCConversion()
-
Method Summary
All Methods Instance Methods Concrete Methods Modifier and Type Method Description booleancheckAndApply(SameDiff sd, OptimizationHelper helper, SameDiffOp op, ArrayHolder constantArrays, ArrayHolder variablesArrays)
-
-
-
Method Detail
-
checkAndApply
public boolean checkAndApply(SameDiff sd, OptimizationHelper helper, SameDiffOp op, ArrayHolder constantArrays, ArrayHolder variablesArrays)
- Specified by:
checkAndApplyin interfaceOptimizer- Parameters:
sd- Current SameDiff instance to optimizehelper- Helper class for optimizationop- Operation to check for optimizationconstantArrays- Array holder for constant arraysvariablesArrays- Array holder for variable arrays- Returns:
- True if the optimization was applied
-
-