diff --git a/deeplearning4j/deeplearning4j-core/src/test/java/org/deeplearning4j/gradientcheck/LossFunctionGradientCheck.java b/deeplearning4j/deeplearning4j-core/src/test/java/org/deeplearning4j/gradientcheck/LossFunctionGradientCheck.java index 45e98246e449..99b568939166 100644 --- a/deeplearning4j/deeplearning4j-core/src/test/java/org/deeplearning4j/gradientcheck/LossFunctionGradientCheck.java +++ b/deeplearning4j/deeplearning4j-core/src/test/java/org/deeplearning4j/gradientcheck/LossFunctionGradientCheck.java @@ -51,6 +51,8 @@ import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertTrue; +import static org.nd4j.linalg.indexing.NDArrayIndex.all; +import static org.nd4j.linalg.indexing.NDArrayIndex.point; /** * Created by Alex on 12/09/2016. @@ -411,7 +413,9 @@ public static INDArray[] getFeaturesAndLabels(ILossFunction l, long[] featuresSh case "LossKLD": //KL divergence: should be a probability distribution for labels?? ret[1] = Nd4j.rand(labelsShape); - Nd4j.getExecutioner().exec(new OldSoftMax(ret[1]), 1); + for(int i=0; i