Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 9 additions & 1 deletion backends/arm/scripts/collect_testname_resources.py
Original file line number Diff line number Diff line change
Expand Up @@ -61,6 +61,7 @@
"upsample_nearest2d.vec",
"index_put.default",
"conv_transpose2d.default",
"index_copy.default",
]
_ALL_EDGE_OPS = _SAMPLE_INPUT.keys() | _CUSTOM_EDGE_OPS

Expand Down Expand Up @@ -138,9 +139,16 @@ def _collect_arm_models(models_md: pathlib.Path) -> set[str]:
def _normalize_op_name(edge_name: str) -> str:
op, overload = edge_name.split(".")

# There are ops where we want to keep "copy" in the name
# Add them in this list as we encounter them
ignore_copy_list = {"index_copy"}

op = op.lower()
op = op.removeprefix("_")
op = op.removesuffix("_copy")

if op not in ignore_copy_list:
op = op.removesuffix("_copy")

op = op.removesuffix("_with_indices")

overload = overload.lower()
Expand Down
181 changes: 181 additions & 0 deletions backends/arm/test/ops/test_index_copy.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,181 @@
# Copyright 2026 Arm Limited and/or its affiliates.
#
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree.
from typing import Tuple

import torch
from executorch.backends.arm._passes import InsertInt32CastsAfterInt64PlaceholdersPass
from executorch.backends.arm.test import common
from executorch.backends.arm.test.tester.test_pipeline import (
EthosU85PipelineINT,
OpNotSupportedPipeline,
TosaPipelineFP,
TosaPipelineINT,
VgfPipeline,
)

input_t = Tuple[int, torch.Tensor, torch.LongTensor, torch.Tensor] # dim, x, index, y


class IndexCopyModule(torch.nn.Module):
base_test_data = {
"rand_1d": lambda: (
0,
torch.rand((6,), dtype=torch.float32),
torch.LongTensor([0, 2, 5]),
torch.tensor([10.0, 20.0, 30.0], dtype=torch.float32),
),
"rand_3d": lambda: (
0,
torch.rand((4, 2, 3), dtype=torch.float32),
torch.LongTensor([0, 3]),
torch.ones((2, 2, 3), dtype=torch.float32),
),
"rand_3d_dim_1": lambda: (
1,
torch.rand((4, 2, 3), dtype=torch.float32),
torch.LongTensor([0, 1]),
torch.ones((4, 2, 3), dtype=torch.float32),
),
"rand_3d_dim_2": lambda: (
2,
torch.rand((4, 2, 3), dtype=torch.float32),
torch.LongTensor([0]),
torch.ones((4, 2, 1), dtype=torch.float32),
),
"rand_single_index": lambda: (
0,
torch.rand((4, 5), dtype=torch.float32),
torch.LongTensor([0]),
torch.zeros((1, 5), dtype=torch.float32),
),
"rand_single_index_not_zero": lambda: (
0,
torch.rand((4, 5), dtype=torch.float32),
torch.LongTensor([2]),
torch.zeros((1, 5), dtype=torch.float32),
),
"rand_all_rows": lambda: (
0,
torch.rand((3, 4), dtype=torch.float32),
torch.LongTensor([0, 1, 2]),
torch.ones((3, 4), dtype=torch.float32),
),
}

test_data = {
f"{name}_{variant}": (
lambda test_case=test_case, inplace=inplace: (test_case(), inplace)
)
for name, test_case in base_test_data.items()
for variant, inplace in (
("out_of_place", False),
("in_place", True),
)
}

aten_ops = {
False: ["torch.ops.aten.index_put.default"],
True: ["torch.ops.aten.index_put_.default"],
}
exir_op = "executorch_exir_dialects_edge__ops_aten_index_put_default"

def __init__(self, inplace: bool = False):
super().__init__()
self.inplace = inplace

def forward(
self, dim: int, x: torch.Tensor, index: torch.LongTensor, y: torch.Tensor
):
if self.inplace:
return x.index_copy_(dim, index, y)
return x.index_copy(dim, index, y)


xfails_u85 = {
"rand_single_index_not_zero_out_of_place": "MLETORCH-1949: index_copy (SCATTER/INDEX_PUT) produces incorrect results for non-zero indices on U85",
"rand_single_index_not_zero_in_place": "MLETORCH-1949: index_copy (SCATTER/INDEX_PUT) produces incorrect results for non-zero indices on U85",
}


@common.parametrize("test_data", IndexCopyModule.test_data)
def test_index_copy_tosa_FP(test_data):
inputs, inplace = test_data()
module = IndexCopyModule(inplace=inplace)
pipeline = TosaPipelineFP(
module=module,
test_data=inputs,
aten_op=[],
transform_passes=[InsertInt32CastsAfterInt64PlaceholdersPass()],
)
pipeline.run()


@common.parametrize("test_data", IndexCopyModule.test_data)
def test_index_copy_tosa_INT(test_data):
inputs, inplace = test_data()
module = IndexCopyModule(inplace=inplace)
pipeline = TosaPipelineINT(
module=module,
test_data=inputs,
aten_op=IndexCopyModule.aten_ops[inplace],
)
pipeline.run()


@common.parametrize("test_data", IndexCopyModule.test_data)
def test_index_copy_u55_INT(test_data):
inputs, inplace = test_data()
# SCATTER (index_put) is not supported on U55
pipeline = OpNotSupportedPipeline[input_t](
IndexCopyModule(inplace=inplace),
inputs,
{IndexCopyModule.exir_op: 1},
quantize=True,
u55_subset=True,
n_expected_delegates=0,
)
pipeline.run()


@common.parametrize("test_data", IndexCopyModule.test_data, xfails=xfails_u85)
@common.XfailIfNoCorstone320
def test_index_copy_u85_INT(test_data):
inputs, inplace = test_data()
pipeline = EthosU85PipelineINT[input_t](
IndexCopyModule(inplace=inplace),
inputs,
aten_ops=IndexCopyModule.aten_ops[inplace],
)
# int64 index inputs need to be cast to int32; _to_dim_order_copy is not delegated
pipeline.tester.use_portable_ops = True
pipeline.run()


@common.parametrize("test_data", IndexCopyModule.test_data)
@common.SkipIfNoModelConverter
def test_index_copy_vgf_no_quant(test_data):
inputs, inplace = test_data()
pipeline = VgfPipeline[input_t](
IndexCopyModule(inplace=inplace),
inputs,
aten_op=[],
transform_passes=[InsertInt32CastsAfterInt64PlaceholdersPass()],
quantize=False,
)
pipeline.run()


@common.parametrize("test_data", IndexCopyModule.test_data)
@common.SkipIfNoModelConverter
def test_index_copy_vgf_quant(test_data):
inputs, inplace = test_data()
pipeline = VgfPipeline[input_t](
IndexCopyModule(inplace=inplace),
inputs,
aten_op=IndexCopyModule.aten_ops[inplace],
quantize=True,
tosa_spec="TOSA-1.0+INT",
)
pipeline.run()
Loading