Back to home page

OSCL-LXR

 
 

    


0001 // SPDX-License-Identifier: GPL-2.0 or Linux-OpenIB
0002 /* Copyright (c) 2015 - 2021 Intel Corporation */
0003 #include "main.h"
0004 #include "../../../net/ethernet/intel/ice/ice.h"
0005 
0006 MODULE_ALIAS("i40iw");
0007 MODULE_AUTHOR("Intel Corporation, <e1000-rdma@lists.sourceforge.net>");
0008 MODULE_DESCRIPTION("Intel(R) Ethernet Protocol Driver for RDMA");
0009 MODULE_LICENSE("Dual BSD/GPL");
0010 
0011 static struct notifier_block irdma_inetaddr_notifier = {
0012     .notifier_call = irdma_inetaddr_event
0013 };
0014 
0015 static struct notifier_block irdma_inetaddr6_notifier = {
0016     .notifier_call = irdma_inet6addr_event
0017 };
0018 
0019 static struct notifier_block irdma_net_notifier = {
0020     .notifier_call = irdma_net_event
0021 };
0022 
0023 static struct notifier_block irdma_netdevice_notifier = {
0024     .notifier_call = irdma_netdevice_event
0025 };
0026 
0027 static void irdma_register_notifiers(void)
0028 {
0029     register_inetaddr_notifier(&irdma_inetaddr_notifier);
0030     register_inet6addr_notifier(&irdma_inetaddr6_notifier);
0031     register_netevent_notifier(&irdma_net_notifier);
0032     register_netdevice_notifier(&irdma_netdevice_notifier);
0033 }
0034 
0035 static void irdma_unregister_notifiers(void)
0036 {
0037     unregister_netevent_notifier(&irdma_net_notifier);
0038     unregister_inetaddr_notifier(&irdma_inetaddr_notifier);
0039     unregister_inet6addr_notifier(&irdma_inetaddr6_notifier);
0040     unregister_netdevice_notifier(&irdma_netdevice_notifier);
0041 }
0042 
0043 static void irdma_prep_tc_change(struct irdma_device *iwdev)
0044 {
0045     iwdev->vsi.tc_change_pending = true;
0046     irdma_sc_suspend_resume_qps(&iwdev->vsi, IRDMA_OP_SUSPEND);
0047 
0048     /* Wait for all qp's to suspend */
0049     wait_event_timeout(iwdev->suspend_wq,
0050                !atomic_read(&iwdev->vsi.qp_suspend_reqs),
0051                IRDMA_EVENT_TIMEOUT);
0052     irdma_ws_reset(&iwdev->vsi);
0053 }
0054 
0055 static void irdma_log_invalid_mtu(u16 mtu, struct irdma_sc_dev *dev)
0056 {
0057     if (mtu < IRDMA_MIN_MTU_IPV4)
0058         ibdev_warn(to_ibdev(dev), "MTU setting [%d] too low for RDMA traffic. Minimum MTU is 576 for IPv4\n", mtu);
0059     else if (mtu < IRDMA_MIN_MTU_IPV6)
0060         ibdev_warn(to_ibdev(dev), "MTU setting [%d] too low for RDMA traffic. Minimum MTU is 1280 for IPv6\\n", mtu);
0061 }
0062 
0063 static void irdma_fill_qos_info(struct irdma_l2params *l2params,
0064                 struct iidc_qos_params *qos_info)
0065 {
0066     int i;
0067 
0068     l2params->num_tc = qos_info->num_tc;
0069     l2params->vsi_prio_type = qos_info->vport_priority_type;
0070     l2params->vsi_rel_bw = qos_info->vport_relative_bw;
0071     for (i = 0; i < l2params->num_tc; i++) {
0072         l2params->tc_info[i].egress_virt_up =
0073             qos_info->tc_info[i].egress_virt_up;
0074         l2params->tc_info[i].ingress_virt_up =
0075             qos_info->tc_info[i].ingress_virt_up;
0076         l2params->tc_info[i].prio_type = qos_info->tc_info[i].prio_type;
0077         l2params->tc_info[i].rel_bw = qos_info->tc_info[i].rel_bw;
0078         l2params->tc_info[i].tc_ctx = qos_info->tc_info[i].tc_ctx;
0079     }
0080     for (i = 0; i < IIDC_MAX_USER_PRIORITY; i++)
0081         l2params->up2tc[i] = qos_info->up2tc[i];
0082     if (qos_info->pfc_mode == IIDC_DSCP_PFC_MODE) {
0083         l2params->dscp_mode = true;
0084         memcpy(l2params->dscp_map, qos_info->dscp_map, sizeof(l2params->dscp_map));
0085     }
0086 }
0087 
0088 static void irdma_iidc_event_handler(struct ice_pf *pf, struct iidc_event *event)
0089 {
0090     struct irdma_device *iwdev = dev_get_drvdata(&pf->adev->dev);
0091     struct irdma_l2params l2params = {};
0092 
0093     if (*event->type & BIT(IIDC_EVENT_AFTER_MTU_CHANGE)) {
0094         ibdev_dbg(&iwdev->ibdev, "CLNT: new MTU = %d\n", iwdev->netdev->mtu);
0095         if (iwdev->vsi.mtu != iwdev->netdev->mtu) {
0096             l2params.mtu = iwdev->netdev->mtu;
0097             l2params.mtu_changed = true;
0098             irdma_log_invalid_mtu(l2params.mtu, &iwdev->rf->sc_dev);
0099             irdma_change_l2params(&iwdev->vsi, &l2params);
0100         }
0101     } else if (*event->type & BIT(IIDC_EVENT_BEFORE_TC_CHANGE)) {
0102         if (iwdev->vsi.tc_change_pending)
0103             return;
0104 
0105         irdma_prep_tc_change(iwdev);
0106     } else if (*event->type & BIT(IIDC_EVENT_AFTER_TC_CHANGE)) {
0107         struct iidc_qos_params qos_info = {};
0108 
0109         if (!iwdev->vsi.tc_change_pending)
0110             return;
0111 
0112         l2params.tc_changed = true;
0113         ibdev_dbg(&iwdev->ibdev, "CLNT: TC Change\n");
0114         ice_get_qos_params(pf, &qos_info);
0115         irdma_fill_qos_info(&l2params, &qos_info);
0116         if (iwdev->rf->protocol_used != IRDMA_IWARP_PROTOCOL_ONLY)
0117             iwdev->dcb_vlan_mode = qos_info.num_tc > 1 && !l2params.dscp_mode;
0118         irdma_change_l2params(&iwdev->vsi, &l2params);
0119     } else if (*event->type & BIT(IIDC_EVENT_CRIT_ERR)) {
0120         ibdev_warn(&iwdev->ibdev, "ICE OICR event notification: oicr = 0x%08x\n",
0121                event->reg);
0122         if (event->reg & IRDMAPFINT_OICR_PE_CRITERR_M) {
0123             u32 pe_criterr;
0124 
0125             pe_criterr = readl(iwdev->rf->sc_dev.hw_regs[IRDMA_GLPE_CRITERR]);
0126 #define IRDMA_Q1_RESOURCE_ERR 0x0001024d
0127             if (pe_criterr != IRDMA_Q1_RESOURCE_ERR) {
0128                 ibdev_err(&iwdev->ibdev, "critical PE Error, GLPE_CRITERR=0x%08x\n",
0129                       pe_criterr);
0130                 iwdev->rf->reset = true;
0131             } else {
0132                 ibdev_warn(&iwdev->ibdev, "Q1 Resource Check\n");
0133             }
0134         }
0135         if (event->reg & IRDMAPFINT_OICR_HMC_ERR_M) {
0136             ibdev_err(&iwdev->ibdev, "HMC Error\n");
0137             iwdev->rf->reset = true;
0138         }
0139         if (event->reg & IRDMAPFINT_OICR_PE_PUSH_M) {
0140             ibdev_err(&iwdev->ibdev, "PE Push Error\n");
0141             iwdev->rf->reset = true;
0142         }
0143         if (iwdev->rf->reset)
0144             iwdev->rf->gen_ops.request_reset(iwdev->rf);
0145     }
0146 }
0147 
0148 /**
0149  * irdma_request_reset - Request a reset
0150  * @rf: RDMA PCI function
0151  */
0152 static void irdma_request_reset(struct irdma_pci_f *rf)
0153 {
0154     struct ice_pf *pf = rf->cdev;
0155 
0156     ibdev_warn(&rf->iwdev->ibdev, "Requesting a reset\n");
0157     ice_rdma_request_reset(pf, IIDC_PFR);
0158 }
0159 
0160 /**
0161  * irdma_lan_register_qset - Register qset with LAN driver
0162  * @vsi: vsi structure
0163  * @tc_node: Traffic class node
0164  */
0165 static int irdma_lan_register_qset(struct irdma_sc_vsi *vsi,
0166                    struct irdma_ws_node *tc_node)
0167 {
0168     struct irdma_device *iwdev = vsi->back_vsi;
0169     struct ice_pf *pf = iwdev->rf->cdev;
0170     struct iidc_rdma_qset_params qset = {};
0171     int ret;
0172 
0173     qset.qs_handle = tc_node->qs_handle;
0174     qset.tc = tc_node->traffic_class;
0175     qset.vport_id = vsi->vsi_idx;
0176     ret = ice_add_rdma_qset(pf, &qset);
0177     if (ret) {
0178         ibdev_dbg(&iwdev->ibdev, "WS: LAN alloc_res for rdma qset failed.\n");
0179         return ret;
0180     }
0181 
0182     tc_node->l2_sched_node_id = qset.teid;
0183     vsi->qos[tc_node->user_pri].l2_sched_node_id = qset.teid;
0184 
0185     return 0;
0186 }
0187 
0188 /**
0189  * irdma_lan_unregister_qset - Unregister qset with LAN driver
0190  * @vsi: vsi structure
0191  * @tc_node: Traffic class node
0192  */
0193 static void irdma_lan_unregister_qset(struct irdma_sc_vsi *vsi,
0194                       struct irdma_ws_node *tc_node)
0195 {
0196     struct irdma_device *iwdev = vsi->back_vsi;
0197     struct ice_pf *pf = iwdev->rf->cdev;
0198     struct iidc_rdma_qset_params qset = {};
0199 
0200     qset.qs_handle = tc_node->qs_handle;
0201     qset.tc = tc_node->traffic_class;
0202     qset.vport_id = vsi->vsi_idx;
0203     qset.teid = tc_node->l2_sched_node_id;
0204 
0205     if (ice_del_rdma_qset(pf, &qset))
0206         ibdev_dbg(&iwdev->ibdev, "WS: LAN free_res for rdma qset failed.\n");
0207 }
0208 
0209 static void irdma_remove(struct auxiliary_device *aux_dev)
0210 {
0211     struct iidc_auxiliary_dev *iidc_adev = container_of(aux_dev,
0212                                 struct iidc_auxiliary_dev,
0213                                 adev);
0214     struct ice_pf *pf = iidc_adev->pf;
0215     struct irdma_device *iwdev = auxiliary_get_drvdata(aux_dev);
0216 
0217     irdma_ib_unregister_device(iwdev);
0218     ice_rdma_update_vsi_filter(pf, iwdev->vsi_num, false);
0219 
0220     pr_debug("INIT: Gen2 PF[%d] device remove success\n", PCI_FUNC(pf->pdev->devfn));
0221 }
0222 
0223 static void irdma_fill_device_info(struct irdma_device *iwdev, struct ice_pf *pf,
0224                    struct ice_vsi *vsi)
0225 {
0226     struct irdma_pci_f *rf = iwdev->rf;
0227 
0228     rf->cdev = pf;
0229     rf->gen_ops.register_qset = irdma_lan_register_qset;
0230     rf->gen_ops.unregister_qset = irdma_lan_unregister_qset;
0231     rf->hw.hw_addr = pf->hw.hw_addr;
0232     rf->pcidev = pf->pdev;
0233     rf->msix_count =  pf->num_rdma_msix;
0234     rf->pf_id = pf->hw.pf_id;
0235     rf->msix_entries = &pf->msix_entries[pf->rdma_base_vector];
0236     rf->default_vsi.vsi_idx = vsi->vsi_num;
0237     rf->protocol_used = pf->rdma_mode & IIDC_RDMA_PROTOCOL_ROCEV2 ?
0238                 IRDMA_ROCE_PROTOCOL_ONLY : IRDMA_IWARP_PROTOCOL_ONLY;
0239     rf->rdma_ver = IRDMA_GEN_2;
0240     rf->rsrc_profile = IRDMA_HMC_PROFILE_DEFAULT;
0241     rf->rst_to = IRDMA_RST_TIMEOUT_HZ;
0242     rf->gen_ops.request_reset = irdma_request_reset;
0243     rf->limits_sel = 7;
0244     rf->iwdev = iwdev;
0245     mutex_init(&iwdev->ah_tbl_lock);
0246     iwdev->netdev = vsi->netdev;
0247     iwdev->vsi_num = vsi->vsi_num;
0248     iwdev->init_state = INITIAL_STATE;
0249     iwdev->roce_cwnd = IRDMA_ROCE_CWND_DEFAULT;
0250     iwdev->roce_ackcreds = IRDMA_ROCE_ACKCREDS_DEFAULT;
0251     iwdev->rcv_wnd = IRDMA_CM_DEFAULT_RCV_WND_SCALED;
0252     iwdev->rcv_wscale = IRDMA_CM_DEFAULT_RCV_WND_SCALE;
0253     if (rf->protocol_used == IRDMA_ROCE_PROTOCOL_ONLY)
0254         iwdev->roce_mode = true;
0255 }
0256 
0257 static int irdma_probe(struct auxiliary_device *aux_dev, const struct auxiliary_device_id *id)
0258 {
0259     struct iidc_auxiliary_dev *iidc_adev = container_of(aux_dev,
0260                                 struct iidc_auxiliary_dev,
0261                                 adev);
0262     struct ice_pf *pf = iidc_adev->pf;
0263     struct ice_vsi *vsi = ice_get_main_vsi(pf);
0264     struct iidc_qos_params qos_info = {};
0265     struct irdma_device *iwdev;
0266     struct irdma_pci_f *rf;
0267     struct irdma_l2params l2params = {};
0268     int err;
0269 
0270     if (!vsi)
0271         return -EIO;
0272     iwdev = ib_alloc_device(irdma_device, ibdev);
0273     if (!iwdev)
0274         return -ENOMEM;
0275     iwdev->rf = kzalloc(sizeof(*rf), GFP_KERNEL);
0276     if (!iwdev->rf) {
0277         ib_dealloc_device(&iwdev->ibdev);
0278         return -ENOMEM;
0279     }
0280 
0281     irdma_fill_device_info(iwdev, pf, vsi);
0282     rf = iwdev->rf;
0283 
0284     err = irdma_ctrl_init_hw(rf);
0285     if (err)
0286         goto err_ctrl_init;
0287 
0288     l2params.mtu = iwdev->netdev->mtu;
0289     ice_get_qos_params(pf, &qos_info);
0290     irdma_fill_qos_info(&l2params, &qos_info);
0291     if (iwdev->rf->protocol_used != IRDMA_IWARP_PROTOCOL_ONLY)
0292         iwdev->dcb_vlan_mode = l2params.num_tc > 1 && !l2params.dscp_mode;
0293 
0294     err = irdma_rt_init_hw(iwdev, &l2params);
0295     if (err)
0296         goto err_rt_init;
0297 
0298     err = irdma_ib_register_device(iwdev);
0299     if (err)
0300         goto err_ibreg;
0301 
0302     ice_rdma_update_vsi_filter(pf, iwdev->vsi_num, true);
0303 
0304     ibdev_dbg(&iwdev->ibdev, "INIT: Gen2 PF[%d] device probe success\n", PCI_FUNC(rf->pcidev->devfn));
0305     auxiliary_set_drvdata(aux_dev, iwdev);
0306 
0307     return 0;
0308 
0309 err_ibreg:
0310     irdma_rt_deinit_hw(iwdev);
0311 err_rt_init:
0312     irdma_ctrl_deinit_hw(rf);
0313 err_ctrl_init:
0314     kfree(iwdev->rf);
0315     ib_dealloc_device(&iwdev->ibdev);
0316 
0317     return err;
0318 }
0319 
0320 static const struct auxiliary_device_id irdma_auxiliary_id_table[] = {
0321     {.name = "ice.iwarp", },
0322     {.name = "ice.roce", },
0323     {},
0324 };
0325 
0326 MODULE_DEVICE_TABLE(auxiliary, irdma_auxiliary_id_table);
0327 
0328 static struct iidc_auxiliary_drv irdma_auxiliary_drv = {
0329     .adrv = {
0330         .id_table = irdma_auxiliary_id_table,
0331         .probe = irdma_probe,
0332         .remove = irdma_remove,
0333     },
0334     .event_handler = irdma_iidc_event_handler,
0335 };
0336 
0337 static int __init irdma_init_module(void)
0338 {
0339     int ret;
0340 
0341     ret = auxiliary_driver_register(&i40iw_auxiliary_drv);
0342     if (ret) {
0343         pr_err("Failed i40iw(gen_1) auxiliary_driver_register() ret=%d\n",
0344                ret);
0345         return ret;
0346     }
0347 
0348     ret = auxiliary_driver_register(&irdma_auxiliary_drv.adrv);
0349     if (ret) {
0350         auxiliary_driver_unregister(&i40iw_auxiliary_drv);
0351         pr_err("Failed irdma auxiliary_driver_register() ret=%d\n",
0352                ret);
0353         return ret;
0354     }
0355 
0356     irdma_register_notifiers();
0357 
0358     return 0;
0359 }
0360 
0361 static void __exit irdma_exit_module(void)
0362 {
0363     irdma_unregister_notifiers();
0364     auxiliary_driver_unregister(&irdma_auxiliary_drv.adrv);
0365     auxiliary_driver_unregister(&i40iw_auxiliary_drv);
0366 }
0367 
0368 module_init(irdma_init_module);
0369 module_exit(irdma_exit_module);