This adds thresholds to be specified in the VendorPortCounters PMA operation.
The reserved fields are now defined in the appropriate header file. Signed-off-by: Mike Marciniszyn <mike.marcinis...@qlogic.com> --- drivers/infiniband/hw/qib/qib_mad.c | 97 +++++++++++++++++++++++------------ drivers/infiniband/hw/qib/qib_mad.h | 5 +- 2 files changed, 66 insertions(+), 36 deletions(-) diff --git a/drivers/infiniband/hw/qib/qib_mad.c b/drivers/infiniband/hw/qib/qib_mad.c index 94b0d1f..83305ac 100644 --- a/drivers/infiniband/hw/qib/qib_mad.c +++ b/drivers/infiniband/hw/qib/qib_mad.c @@ -1135,14 +1135,16 @@ static int pma_get_classportinfo(struct ib_perf *pmp, p->class_version = 1; p->cap_mask = IB_PMA_CLASS_CAP_EXT_WIDTH; /* - * Set the most significant bit of CM2 to indicate support for - * congestion statistics - */ - p->reserved[0] = dd->psxmitwait_supported << 7; - /* * Expected response time is 4.096 usec. * 2^18 == 1.073741824 sec. */ - p->resp_time_value = 18; + p->cap_mask2_resp_time_value = cpu_to_be32(18); + if (dd->psxmitwait_supported) + /* + * Set the most significant two bits of CM2 to indicate + * support for congestion statistics + */ + p->cap_mask2_resp_time_value |= + IB_PMA_CLASS_CAP_PORT_CONGS; return reply((struct ib_smp *) pmp); } @@ -1709,8 +1711,14 @@ static int pma_set_portcounters(struct ib_perf *pmp, } static int pma_set_portcounters_cong(struct ib_perf *pmp, - struct ib_device *ibdev, u8 port) + struct ib_device *ibdev, u8 port, + struct ib_perf *pmp_in) { + /* Congestion PMA packets start at offset 24 not 64 */ + struct ib_pma_portcounters_cong *p_out = + (struct ib_pma_portcounters_cong *)pmp->reserved; + struct ib_pma_portcounters_cong *p_in = + (struct ib_pma_portcounters_cong *)pmp_in->reserved; struct qib_ibport *ibp = to_iport(ibdev, port); struct qib_pportdata *ppd = ppd_from_ibp(ibp); struct qib_devdata *dd = dd_from_ppd(ppd); @@ -1724,36 +1732,56 @@ static int pma_set_portcounters_cong(struct ib_perf *pmp, ret = pma_get_portcounters_cong(pmp, ibdev, port); if (counter_select & IB_PMA_SEL_CONG_XMIT) { - spin_lock_irqsave(&ppd->ibport_data.lock, flags); - ppd->cong_stats.counter = 0; - dd->f_set_cntr_sample(ppd, QIB_CONG_TIMER_PSINTERVAL, - 0x0); - spin_unlock_irqrestore(&ppd->ibport_data.lock, flags); + if (p_out->port_xmit_wait > p_in->port_xmit_wait) { + spin_lock_irqsave(&ppd->ibport_data.lock, flags); + ppd->cong_stats.counter = 0; + dd->f_set_cntr_sample(ppd, QIB_CONG_TIMER_PSINTERVAL, + 0x0); + spin_unlock_irqrestore(&ppd->ibport_data.lock, flags); + } } if (counter_select & IB_PMA_SEL_CONG_PORT_DATA) { - ibp->z_port_xmit_data = cntrs.port_xmit_data; - ibp->z_port_rcv_data = cntrs.port_rcv_data; - ibp->z_port_xmit_packets = cntrs.port_xmit_packets; - ibp->z_port_rcv_packets = cntrs.port_rcv_packets; + if (p_out->port_xmit_data > p_in->port_xmit_data) + ibp->z_port_xmit_data = cntrs.port_xmit_data; + if (p_out->port_rcv_data > p_in->port_rcv_data) + ibp->z_port_rcv_data = cntrs.port_rcv_data; + if (p_out->port_xmit_packets > p_in->port_xmit_packets) + ibp->z_port_xmit_packets = cntrs.port_xmit_packets; + if (p_out->port_rcv_packets > p_in->port_rcv_packets) + ibp->z_port_rcv_packets = cntrs.port_rcv_packets; } if (counter_select & IB_PMA_SEL_CONG_ALL) { - ibp->z_symbol_error_counter = - cntrs.symbol_error_counter; - ibp->z_link_error_recovery_counter = - cntrs.link_error_recovery_counter; - ibp->z_link_downed_counter = - cntrs.link_downed_counter; - ibp->z_port_rcv_errors = cntrs.port_rcv_errors; - ibp->z_port_rcv_remphys_errors = - cntrs.port_rcv_remphys_errors; - ibp->z_port_xmit_discards = - cntrs.port_xmit_discards; - ibp->z_local_link_integrity_errors = - cntrs.local_link_integrity_errors; - ibp->z_excessive_buffer_overrun_errors = - cntrs.excessive_buffer_overrun_errors; - ibp->n_vl15_dropped = 0; - ibp->z_vl15_dropped = cntrs.vl15_dropped; + if (p_out->symbol_error_counter > p_in->symbol_error_counter) + ibp->z_symbol_error_counter = + cntrs.symbol_error_counter; + if (p_out->link_error_recovery_counter > + p_in->link_error_recovery_counter) + ibp->z_link_error_recovery_counter = + cntrs.link_error_recovery_counter; + if (p_out->link_downed_counter > p_in->link_downed_counter) + ibp->z_link_downed_counter = + cntrs.link_downed_counter; + if (p_out->port_rcv_errors > p_in->port_rcv_errors) + ibp->z_port_rcv_errors = cntrs.port_rcv_errors; + if (p_out->port_rcv_remphys_errors > + p_in->port_rcv_remphys_errors) + ibp->z_port_rcv_remphys_errors = + cntrs.port_rcv_remphys_errors; + if (p_out->port_xmit_discards > p_in->port_xmit_discards) + ibp->z_port_xmit_discards = + cntrs.port_xmit_discards; + if ((p_out->lli_ebor_errors & 0xf0) > + (p_in->lli_ebor_errors & 0xf0)) + ibp->z_local_link_integrity_errors = + cntrs.local_link_integrity_errors; + if ((p_out->lli_ebor_errors & 0x0f) > + (p_in->lli_ebor_errors & 0x0f)) + ibp->z_excessive_buffer_overrun_errors = + cntrs.excessive_buffer_overrun_errors; + if (p_out->vl15_dropped > p_in->vl15_dropped) { + ibp->n_vl15_dropped = 0; + ibp->z_vl15_dropped = cntrs.vl15_dropped; + } } return ret; @@ -2004,7 +2032,8 @@ static int process_perf(struct ib_device *ibdev, u8 port, ret = pma_set_portcounters_ext(pmp, ibdev, port); goto bail; case IB_PMA_PORT_COUNTERS_CONG: - ret = pma_set_portcounters_cong(pmp, ibdev, port); + ret = pma_set_portcounters_cong(pmp, ibdev, port, + (struct ib_perf *)in_mad); goto bail; default: pmp->status |= IB_SMP_UNSUP_METH_ATTR; diff --git a/drivers/infiniband/hw/qib/qib_mad.h b/drivers/infiniband/hw/qib/qib_mad.h index 147aff9..b1c7b21 100644 --- a/drivers/infiniband/hw/qib/qib_mad.h +++ b/drivers/infiniband/hw/qib/qib_mad.h @@ -187,6 +187,8 @@ struct ib_vl_weight_elem { #define IB_PMA_CLASS_CAP_EXT_WIDTH cpu_to_be16(1 << 9) #define IB_PMA_CLASS_CAP_XMIT_WAIT cpu_to_be16(1 << 12) +#define IB_PMA_CLASS_CAP_PORT_CONGS cpu_to_be32(3 << 30) + #define IB_PMA_CLASS_PORT_INFO cpu_to_be16(0x0001) #define IB_PMA_PORT_SAMPLES_CONTROL cpu_to_be16(0x0010) #define IB_PMA_PORT_SAMPLES_RESULT cpu_to_be16(0x0011) @@ -214,8 +216,7 @@ struct ib_pma_classportinfo { u8 base_version; u8 class_version; __be16 cap_mask; - u8 reserved[3]; - u8 resp_time_value; /* only lower 5 bits */ + __be32 cap_mask2_resp_time_value; /* 27, 5 bits respectively */ union ib_gid redirect_gid; __be32 redirect_tc_sl_fl; /* 8, 4, 20 bits respectively */ __be16 redirect_lid; -- To unsubscribe from this list: send the line "unsubscribe linux-rdma" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html