struct mlx5_flow_group *send_to_vport_grp;
struct mlx5_flow_group *send_to_vport_meta_grp;
struct mlx5_flow_group *peer_miss_grp;
- struct mlx5_flow_handle **peer_miss_rules[MLX5_MAX_PORTS];
+ struct xarray peer_miss_rules;
struct mlx5_flow_group *miss_grp;
struct mlx5_flow_handle **send_to_vport_meta_rules;
struct mlx5_flow_handle *miss_rule_uni;
struct mlx5_flow_handle *flow;
struct mlx5_vport *peer_vport;
struct mlx5_flow_spec *spec;
- int err, pfindex;
+ int err;
unsigned long i;
void *misc;
}
}
- pfindex = mlx5_get_dev_index(peer_dev);
- if (pfindex >= MLX5_MAX_PORTS) {
- esw_warn(esw->dev, "Peer dev index(%d) is over the max num defined(%d)\n",
- pfindex, MLX5_MAX_PORTS);
- err = -EINVAL;
+ err = xa_insert(&esw->fdb_table.offloads.peer_miss_rules,
+ MLX5_CAP_GEN(peer_dev, vhca_id), flows, GFP_KERNEL);
+ if (err)
goto add_ec_vf_flow_err;
- }
- esw->fdb_table.offloads.peer_miss_rules[pfindex] = flows;
kvfree(spec);
return 0;
struct mlx5_core_dev *peer_dev)
{
struct mlx5_eswitch *peer_esw = peer_dev->priv.eswitch;
- u16 peer_index = mlx5_get_dev_index(peer_dev);
+ u16 peer_vhca_id = MLX5_CAP_GEN(peer_dev, vhca_id);
struct mlx5_flow_handle **flows;
struct mlx5_vport *peer_vport;
unsigned long i;
- flows = esw->fdb_table.offloads.peer_miss_rules[peer_index];
+ flows = xa_erase(&esw->fdb_table.offloads.peer_miss_rules,
+ peer_vhca_id);
if (!flows)
return;
}
kvfree(flows);
- esw->fdb_table.offloads.peer_miss_rules[peer_index] = NULL;
}
static int esw_add_fdb_miss_rule(struct mlx5_eswitch *esw)
return;
xa_init(&esw->paired);
+ xa_init(&esw->fdb_table.offloads.peer_miss_rules);
esw->num_peers = 0;
esw->devcom = mlx5_devcom_register_component(esw->dev->priv.devc,
MLX5_DEVCOM_ESW_OFFLOADS,
mlx5_devcom_unregister_component(esw->devcom);
xa_destroy(&esw->paired);
+ xa_destroy(&esw->fdb_table.offloads.peer_miss_rules);
esw->devcom = NULL;
}