Repository: incubator-systemml
Updated Branches:
  refs/heads/master 19e21744c -> 6c215e700


[SYSTEMML-1573] Incorporate ALLOW_OPERATOR_FUSION in ConvolutionOp for 
developer testing

Closes #482.


Project: http://git-wip-us.apache.org/repos/asf/incubator-systemml/repo
Commit: 
http://git-wip-us.apache.org/repos/asf/incubator-systemml/commit/6c215e70
Tree: http://git-wip-us.apache.org/repos/asf/incubator-systemml/tree/6c215e70
Diff: http://git-wip-us.apache.org/repos/asf/incubator-systemml/diff/6c215e70

Branch: refs/heads/master
Commit: 6c215e700c1855074228972f952663663f6eabaa
Parents: 19e2174
Author: Niketan Pansare <[email protected]>
Authored: Sun May 7 20:17:21 2017 -0800
Committer: Niketan Pansare <[email protected]>
Committed: Sun May 7 21:17:21 2017 -0700

----------------------------------------------------------------------
 src/main/java/org/apache/sysml/hops/ConvolutionOp.java | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-systemml/blob/6c215e70/src/main/java/org/apache/sysml/hops/ConvolutionOp.java
----------------------------------------------------------------------
diff --git a/src/main/java/org/apache/sysml/hops/ConvolutionOp.java 
b/src/main/java/org/apache/sysml/hops/ConvolutionOp.java
index cb67d65..7ab2988 100644
--- a/src/main/java/org/apache/sysml/hops/ConvolutionOp.java
+++ b/src/main/java/org/apache/sysml/hops/ConvolutionOp.java
@@ -153,15 +153,15 @@ public class ConvolutionOp extends Hop  implements 
MultiThreadedHop
                // RELU_MAX_POOLING and RELU_MAX_POOLING_BACKWARD is extremely 
useful for CP backend 
                // by reducing unnecessary sparse-to-dense-to-sparse conversion.
                // For other backends, this operators is not necessary as it 
reduces an additional relu operator.
-               if(et == ExecType.CP && op == ConvOp.MAX_POOLING && 
isInputReLU(inputs.get(0))) {
+               if(OptimizerUtils.ALLOW_OPERATOR_FUSION && et == ExecType.CP && 
op == ConvOp.MAX_POOLING && isInputReLU(inputs.get(0))) {
                        in = inputs.get(0).getInput().get(0).constructLops();
                        lopOp = OperationTypes.RELU_MAX_POOLING;
                }
-               else if(et == ExecType.CP && op == ConvOp.MAX_POOLING_BACKWARD 
&& isInputReLU(inputs.get(0))) {
+               else if(OptimizerUtils.ALLOW_OPERATOR_FUSION && et == 
ExecType.CP && op == ConvOp.MAX_POOLING_BACKWARD && isInputReLU(inputs.get(0))) 
{
                        in = inputs.get(0).getInput().get(0).constructLops();
                        lopOp = OperationTypes.RELU_MAX_POOLING_BACKWARD;
                }
-               else if(op == ConvOp.BIAS_ADD && isInputConv2d(inputs.get(0))) {
+               else if(OptimizerUtils.ALLOW_OPERATOR_FUSION && op == 
ConvOp.BIAS_ADD && isInputConv2d(inputs.get(0))) {
                        lopOp = OperationTypes.DIRECT_CONV2D_BIAS_ADD;
                        
                        // the first lop is image 

Reply via email to