From 1b7d47ebb10e557eec35f662d8a6fa855e6e4748 Mon Sep 17 00:00:00 2001 From: IvanKobzarev Date: Thu, 5 Feb 2026 09:26:33 -0800 Subject: [PATCH 1/2] [lint] fix mypy unnecessary dict comprehension [ghstack-poisoned] --- autoparallel/shardings/ordered_sharding.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/autoparallel/shardings/ordered_sharding.py b/autoparallel/shardings/ordered_sharding.py index 5298a505..e5a8f5da 100644 --- a/autoparallel/shardings/ordered_sharding.py +++ b/autoparallel/shardings/ordered_sharding.py @@ -186,7 +186,7 @@ def compute_optimal_placement_order_for_parameters(module, sharding_placement): user_src_placement = list(d.values())[0][0] mesh_ndim = len(user_src_placement) - param_grad_map = {p: g for p, g in param_and_grad_nodes} + param_grad_map = dict(param_and_grad_nodes) aligned_pg = [] for param_or_grad_node in redistribution_map.keys(): # just allow for arbitrary execution order if both param and grad From 1c0e87ba40bb49c059643e530e82295ba0182c35 Mon Sep 17 00:00:00 2001 From: IvanKobzarev Date: Thu, 5 Feb 2026 12:17:29 -0800 Subject: [PATCH 2/2] Update on "[lint] fix mypy unnecessary dict comprehension" [ghstack-poisoned] --- autoparallel/shardings/ordered_sharding.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/autoparallel/shardings/ordered_sharding.py b/autoparallel/shardings/ordered_sharding.py index e5a8f5da..60df8909 100644 --- a/autoparallel/shardings/ordered_sharding.py +++ b/autoparallel/shardings/ordered_sharding.py @@ -94,7 +94,7 @@ def get_redistributed_input_placements( x for x in tree_flatten(node.args)[0] if isinstance(x, torch.fx.Node) ] num_input_nodes = len(all_input_nodes) - curr_specs: list[Union[DTensorSpec, tuple[Optional[DTensorSpec], ...]]] = [ + curr_specs: list[Union[DTensorSpec, tuple[Optional[DTensorSpec], ...], None]] = [ sharding_placement[n].output_specs for n in all_input_nodes ] # FIXME ? if node.target == operator.getitem: