Skip to content

Commit

Permalink
Blacklist _convolution op (#1048)
Browse files Browse the repository at this point in the history
* Blacklist _convolution op in LTC

* Removed duplicate Torch_AtenSelectScatterOp instance from autogen .td

* Removed duplicate Torch_AtenSliceScatterOp instance from autogen .td
  • Loading branch information
henrytwo authored and antoniojkim committed Jul 22, 2022
1 parent 71aec98 commit 86fa58a
Show file tree
Hide file tree
Showing 4 changed files with 7 additions and 57 deletions.
3 changes: 2 additions & 1 deletion build_tools/autogen_ltc_backend.yaml
Original file line number Diff line number Diff line change
@@ -1,13 +1,14 @@
blacklist:
# List of unsupported ops in LTC autogen because of some error
- _index_put_impl_ # Error: TODO not sure if there are other valid types to handle here
- empty_like # Error: TODO add support for type BaseType(name=<BaseTy.MemoryFormat: 12>)
- index.Tensor # Error: TODO not sure if there are other valid types to handle here
- index_put # Error: TODO not sure if there are other valid types to handle here
- index_put_ # Error: TODO not sure if there are other valid types to handle here
- _index_put_impl_ # Error: TODO not sure if there are other valid types to handle here
- stack # Error: TODO not sure if there are other valid types to handle here

# Additional ops which autogen is supported for but don't compile yet
- _convolution
- detach
- item
- size
Expand Down
5 changes: 5 additions & 0 deletions e2e_testing/torchscript/xfail_sets.py
Original file line number Diff line number Diff line change
Expand Up @@ -181,6 +181,11 @@
}

LTC_XFAIL_SET = {
"_Convolution2DAllFalseModule_basic",
"_Convolution2DBenchmarkModule_basic",
"_Convolution2DCudnnModule_basic",
"_Convolution2DDeterministicModule_basic",
"_Convolution2DTF32Module_basic",
"AdaptiveAvgPool2dNonUnitOutputSizeDynamicModule_basic",
"AdaptiveAvgPool2dNonUnitOutputSizeStaticModule_basic",
"AddIntModule_basic",
Expand Down
54 changes: 0 additions & 54 deletions include/torch-mlir/Dialect/Torch/IR/GeneratedTorchOps.td
Original file line number Diff line number Diff line change
Expand Up @@ -5341,32 +5341,6 @@ def Torch_AtenSelectIntOp : Torch_Op<"aten.select.int", [
}];
}

def Torch_AtenSelectScatterOp : Torch_Op<"aten.select_scatter", [
AllowsTypeRefinement,
HasValueSemantics,
ReadOnly
]> {
let summary = "Generated op for `aten::select_scatter : (Tensor, Tensor, int, int) -> (Tensor)`";
let arguments = (ins
AnyTorchTensorType:$self,
AnyTorchTensorType:$src,
Torch_IntType:$dim,
Torch_IntType:$index
);
let results = (outs
AnyTorchTensorType:$result
);
let hasCustomAssemblyFormat = 1;
let extraClassDefinition = [{
ParseResult AtenSelectScatterOp::parse(OpAsmParser &parser, OperationState &result) {
return parseDefaultTorchOp(parser, result, 4, 1);
}
void AtenSelectScatterOp::print(OpAsmPrinter &printer) {
printDefaultTorchOp(printer, *this, 4, 1);
}
}];
}

def Torch_AtenSizeIntOp : Torch_Op<"aten.size.int", [
AllowsTypeRefinement,
HasValueSemantics,
Expand Down Expand Up @@ -5822,34 +5796,6 @@ def Torch_AtenSliceTensorOp : Torch_Op<"aten.slice.Tensor", [
}];
}

def Torch_AtenSliceScatterOp : Torch_Op<"aten.slice_scatter", [
AllowsTypeRefinement,
HasValueSemantics,
ReadOnly
]> {
let summary = "Generated op for `aten::slice_scatter : (Tensor, Tensor, int, int?, int?, int) -> (Tensor)`";
let arguments = (ins
AnyTorchTensorType:$self,
AnyTorchTensorType:$src,
Torch_IntType:$dim,
AnyTorchOptionalIntType:$start,
AnyTorchOptionalIntType:$end,
Torch_IntType:$step
);
let results = (outs
AnyTorchTensorType:$result
);
let hasCustomAssemblyFormat = 1;
let extraClassDefinition = [{
ParseResult AtenSliceScatterOp::parse(OpAsmParser &parser, OperationState &result) {
return parseDefaultTorchOp(parser, result, 6, 1);
}
void AtenSliceScatterOp::print(OpAsmPrinter &printer) {
printDefaultTorchOp(printer, *this, 6, 1);
}
}];
}

def Torch_AtenLenTensorOp : Torch_Op<"aten.len.Tensor", [
AllowsTypeRefinement,
HasValueSemantics,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -438,7 +438,6 @@ def emit_with_mutating_variants(key, **kwargs):
emit("aten::_reshape_alias : (Tensor, int[], int[]) -> (Tensor)")
emit("aten::resize_ : (Tensor, int[], int?) -> (Tensor)")
emit("aten::select.int : (Tensor, int, int) -> (Tensor)")
emit("aten::select_scatter : (Tensor, Tensor, int, int) -> (Tensor)")
emit("aten::size.int : (Tensor, int) -> (int)", has_folder=True)
emit("aten::stack : (Tensor[], int) -> (Tensor)")
emit("aten::sum : (Tensor, int?) -> (Tensor)")
Expand All @@ -457,7 +456,6 @@ def emit_with_mutating_variants(key, **kwargs):
emit("aten::where.ScalarOther : (Tensor, Tensor, Scalar) -> (Tensor)")
emit("aten::where.ScalarSelf : (Tensor, Scalar, Tensor) -> (Tensor)")
emit("aten::slice.Tensor : (Tensor, int, int?, int?, int) -> (Tensor)")
emit("aten::slice_scatter : (Tensor, Tensor, int, int?, int?, int) -> (Tensor)")
emit("aten::len.Tensor : (Tensor) -> (int)")
emit("aten::cpu : (Tensor) -> (Tensor)")
emit("aten::gather : (Tensor, int, Tensor, bool) -> (Tensor)")
Expand Down

0 comments on commit 86fa58a

Please sign in to comment.