svn commit: r342288 - head/sys/dev/cxgbe/iw_cxgbe
Navdeep Parhar
np at FreeBSD.org
Thu Dec 20 22:31:09 UTC 2018
Author: np
Date: Thu Dec 20 22:31:07 2018
New Revision: 342288
URL: https://svnweb.freebsd.org/changeset/base/342288
Log:
cxgbe/iw_cxgbe: Do not terminate CTRx messages with \n.
Modified:
head/sys/dev/cxgbe/iw_cxgbe/cm.c
head/sys/dev/cxgbe/iw_cxgbe/cq.c
head/sys/dev/cxgbe/iw_cxgbe/qp.c
head/sys/dev/cxgbe/iw_cxgbe/t4.h
Modified: head/sys/dev/cxgbe/iw_cxgbe/cm.c
==============================================================================
--- head/sys/dev/cxgbe/iw_cxgbe/cm.c Thu Dec 20 22:26:54 2018 (r342287)
+++ head/sys/dev/cxgbe/iw_cxgbe/cm.c Thu Dec 20 22:31:07 2018 (r342288)
@@ -429,7 +429,7 @@ static void process_timeout(struct c4iw_ep *ep)
abort = 0;
break;
default:
- CTR4(KTR_IW_CXGBE, "%s unexpected state ep %p tid %u state %u\n"
+ CTR4(KTR_IW_CXGBE, "%s unexpected state ep %p tid %u state %u"
, __func__, ep, ep->hwtid, ep->com.state);
abort = 0;
}
@@ -1020,7 +1020,7 @@ process_newconn(struct c4iw_listen_ep *master_lep, str
ret = soaccept(new_so, (struct sockaddr **)&remote);
if (ret != 0) {
CTR4(KTR_IW_CXGBE,
- "%s:listen sock:%p, new sock:%p, ret:%d\n",
+ "%s:listen sock:%p, new sock:%p, ret:%d",
__func__, master_lep->com.so, new_so, ret);
if (remote != NULL)
free(remote, M_SONAME);
@@ -2309,7 +2309,7 @@ process_mpa_request(struct c4iw_ep *ep)
MPA_V2_IRD_ORD_MASK;
ep->ord = min_t(u32, ep->ord,
cur_max_read_depth(ep->com.dev));
- CTR3(KTR_IW_CXGBE, "%s initiator ird %u ord %u\n",
+ CTR3(KTR_IW_CXGBE, "%s initiator ird %u ord %u",
__func__, ep->ird, ep->ord);
if (ntohs(mpa_v2_params->ird) & MPA_V2_PEER2PEER_MODEL)
if (peer2peer) {
@@ -2463,7 +2463,7 @@ int c4iw_accept_cr(struct iw_cm_id *cm_id, struct iw_c
ep->ird = 1;
}
- CTR4(KTR_IW_CXGBE, "%s %d ird %d ord %d\n", __func__, __LINE__,
+ CTR4(KTR_IW_CXGBE, "%s %d ird %d ord %d", __func__, __LINE__,
ep->ird, ep->ord);
ep->com.cm_id = cm_id;
Modified: head/sys/dev/cxgbe/iw_cxgbe/cq.c
==============================================================================
--- head/sys/dev/cxgbe/iw_cxgbe/cq.c Thu Dec 20 22:26:54 2018 (r342287)
+++ head/sys/dev/cxgbe/iw_cxgbe/cq.c Thu Dec 20 22:31:07 2018 (r342288)
@@ -671,7 +671,7 @@ proc_cqe:
BUG_ON(wq->sq.in_use <= 0 && wq->sq.in_use >= wq->sq.size);
wq->sq.cidx = (uint16_t)idx;
- CTR2(KTR_IW_CXGBE, "%s completing sq idx %u\n",
+ CTR2(KTR_IW_CXGBE, "%s completing sq idx %u",
__func__, wq->sq.cidx);
*cookie = wq->sq.sw_sq[wq->sq.cidx].wr_id;
t4_sq_consume(wq);
Modified: head/sys/dev/cxgbe/iw_cxgbe/qp.c
==============================================================================
--- head/sys/dev/cxgbe/iw_cxgbe/qp.c Thu Dec 20 22:26:54 2018 (r342287)
+++ head/sys/dev/cxgbe/iw_cxgbe/qp.c Thu Dec 20 22:31:07 2018 (r342288)
@@ -576,7 +576,7 @@ static void free_qp_work(struct work_struct *work)
ucontext = qhp->ucontext;
rhp = qhp->rhp;
- CTR3(KTR_IW_CXGBE, "%s qhp %p ucontext %p\n", __func__,
+ CTR3(KTR_IW_CXGBE, "%s qhp %p ucontext %p", __func__,
qhp, ucontext);
destroy_qp(&rhp->rdev, &qhp->wq,
ucontext ? &ucontext->uctx : &rhp->rdev.uctx);
@@ -1878,10 +1878,10 @@ c4iw_create_qp(struct ib_pd *pd, struct ib_qp_init_att
qhp->ibqp.qp_num = qhp->wq.sq.qid;
init_timer(&(qhp->timer));
- CTR5(KTR_IW_CXGBE, "%s sq id %u size %u memsize %zu num_entries %u\n",
+ CTR5(KTR_IW_CXGBE, "%s sq id %u size %u memsize %zu num_entries %u",
__func__, qhp->wq.sq.qid,
qhp->wq.sq.size, qhp->wq.sq.memsize, attrs->cap.max_send_wr);
- CTR5(KTR_IW_CXGBE, "%s rq id %u size %u memsize %zu num_entries %u\n",
+ CTR5(KTR_IW_CXGBE, "%s rq id %u size %u memsize %zu num_entries %u",
__func__, qhp->wq.rq.qid,
qhp->wq.rq.size, qhp->wq.rq.memsize, attrs->cap.max_recv_wr);
return &qhp->ibqp;
Modified: head/sys/dev/cxgbe/iw_cxgbe/t4.h
==============================================================================
--- head/sys/dev/cxgbe/iw_cxgbe/t4.h Thu Dec 20 22:26:54 2018 (r342287)
+++ head/sys/dev/cxgbe/iw_cxgbe/t4.h Thu Dec 20 22:31:07 2018 (r342288)
@@ -490,13 +490,13 @@ t4_ring_sq_db(struct t4_wq *wq, u16 inc, union t4_wr *
/* Flush host queue memory writes. */
wmb();
if (wc && inc == 1 && wq->sq.bar2_qid == 0 && wqe) {
- CTR2(KTR_IW_CXGBE, "%s: WC wq->sq.pidx = %d\n",
+ CTR2(KTR_IW_CXGBE, "%s: WC wq->sq.pidx = %d",
__func__, wq->sq.pidx);
pio_copy((u64 __iomem *)
((u64)wq->sq.bar2_va + SGE_UDB_WCDOORBELL),
(u64 *)wqe);
} else {
- CTR2(KTR_IW_CXGBE, "%s: DB wq->sq.pidx = %d\n",
+ CTR2(KTR_IW_CXGBE, "%s: DB wq->sq.pidx = %d",
__func__, wq->sq.pidx);
writel(V_PIDX_T5(inc) | V_QID(wq->sq.bar2_qid),
(void __iomem *)((u64)wq->sq.bar2_va +
@@ -515,12 +515,12 @@ t4_ring_rq_db(struct t4_wq *wq, u16 inc, union t4_recv
/* Flush host queue memory writes. */
wmb();
if (wc && inc == 1 && wq->rq.bar2_qid == 0 && wqe) {
- CTR2(KTR_IW_CXGBE, "%s: WC wq->rq.pidx = %d\n",
+ CTR2(KTR_IW_CXGBE, "%s: WC wq->rq.pidx = %d",
__func__, wq->rq.pidx);
pio_copy((u64 __iomem *)((u64)wq->rq.bar2_va +
SGE_UDB_WCDOORBELL), (u64 *)wqe);
} else {
- CTR2(KTR_IW_CXGBE, "%s: DB wq->rq.pidx = %d\n",
+ CTR2(KTR_IW_CXGBE, "%s: DB wq->rq.pidx = %d",
__func__, wq->rq.pidx);
writel(V_PIDX_T5(inc) | V_QID(wq->rq.bar2_qid),
(void __iomem *)((u64)wq->rq.bar2_va +
@@ -604,7 +604,7 @@ static inline void t4_swcq_produce(struct t4_cq *cq)
{
cq->sw_in_use++;
if (cq->sw_in_use == cq->size) {
- CTR2(KTR_IW_CXGBE, "%s cxgb4 sw cq overflow cqid %u\n",
+ CTR2(KTR_IW_CXGBE, "%s cxgb4 sw cq overflow cqid %u",
__func__, cq->cqid);
cq->error = 1;
BUG_ON(1);
@@ -676,7 +676,7 @@ static inline int t4_next_hw_cqe(struct t4_cq *cq, str
static inline struct t4_cqe *t4_next_sw_cqe(struct t4_cq *cq)
{
if (cq->sw_in_use == cq->size) {
- CTR2(KTR_IW_CXGBE, "%s cxgb4 sw cq overflow cqid %u\n",
+ CTR2(KTR_IW_CXGBE, "%s cxgb4 sw cq overflow cqid %u",
__func__, cq->cqid);
cq->error = 1;
BUG_ON(1);
More information about the svn-src-head
mailing list