From: Amir Vadai Date: Thu, 26 Nov 2009 15:28:58 +0000 (+0200) Subject: sdp: cleaned up debug prints X-Git-Tag: v4.1.12-92~264^2~5^2~238 X-Git-Url: https://www.infradead.org/git/?a=commitdiff_plain;h=9019e4057b825e22ce4c8fbaac0fe1fd3dd27b49;p=users%2Fjedix%2Flinux-maple.git sdp: cleaned up debug prints Signed-off-by: Amir Vadai --- diff --git a/drivers/infiniband/ulp/sdp/sdp_cma.c b/drivers/infiniband/ulp/sdp/sdp_cma.c index 258ec0575cf10..b3776218f8875 100644 --- a/drivers/infiniband/ulp/sdp/sdp_cma.c +++ b/drivers/infiniband/ulp/sdp/sdp_cma.c @@ -331,7 +331,7 @@ int sdp_cma_handler(struct rdma_cm_id *id, struct rdma_cm_event *event) rc = rdma_resolve_route(id, SDP_ROUTE_TIMEOUT); break; case RDMA_CM_EVENT_ADDR_ERROR: - sdp_warn(sk, "RDMA_CM_EVENT_ADDR_ERROR\n"); + sdp_dbg(sk, "RDMA_CM_EVENT_ADDR_ERROR\n"); rc = -ENETUNREACH; break; case RDMA_CM_EVENT_ROUTE_RESOLVED: @@ -371,7 +371,7 @@ int sdp_cma_handler(struct rdma_cm_id *id, struct rdma_cm_event *event) sdp_dbg(sk, "RDMA_CM_EVENT_CONNECT_REQUEST\n"); rc = sdp_connect_handler(sk, id, event); if (rc) { - sdp_warn(sk, "Destroy qp !!!!\n"); + sdp_dbg(sk, "Destroying qp\n"); rdma_reject(id, NULL, 0); break; } @@ -407,7 +407,7 @@ int sdp_cma_handler(struct rdma_cm_id *id, struct rdma_cm_event *event) sdp_dbg(sk, "RDMA_CM_EVENT_CONNECT_RESPONSE\n"); rc = sdp_response_handler(sk, id, event); if (rc) { - sdp_warn(sk, "Destroy qp !!!!\n"); + sdp_dbg(sk, "Destroying qp\n"); rdma_reject(id, NULL, 0); } else rc = rdma_accept(id, NULL); @@ -461,7 +461,7 @@ int sdp_cma_handler(struct rdma_cm_id *id, struct rdma_cm_event *event) rc = sdp_disconnected_handler(sk); break; case RDMA_CM_EVENT_DEVICE_REMOVAL: - sdp_warn(sk, "RDMA_CM_EVENT_DEVICE_REMOVAL\n"); + sdp_dbg(sk, "RDMA_CM_EVENT_DEVICE_REMOVAL\n"); rc = -ENETRESET; break; default: diff --git a/drivers/infiniband/ulp/sdp/sdp_main.c b/drivers/infiniband/ulp/sdp/sdp_main.c index 9db3e166905c6..bb8a15be71085 100644 --- a/drivers/infiniband/ulp/sdp/sdp_main.c +++ b/drivers/infiniband/ulp/sdp/sdp_main.c @@ -169,7 +169,7 @@ static int sdp_get_port(struct sock *sk, unsigned short snum) rc = rdma_bind_addr(ssk->id, (struct sockaddr *)&addr); if (rc) { - sdp_warn(sk, "Destroy qp !!!!\n"); + sdp_dbg(sk, "Destroying qp\n"); rdma_destroy_id(ssk->id); ssk->id = NULL; return rc; @@ -742,7 +742,7 @@ static int sdp_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len) rc = rdma_resolve_addr(ssk->id, (struct sockaddr *)&src_addr, uaddr, SDP_RESOLVE_TIMEOUT); if (rc) { - sdp_warn(sk, "rdma_resolve_addr failed: %d\n", rc); + sdp_dbg(sk, "rdma_resolve_addr failed: %d\n", rc); return rc; } @@ -1005,7 +1005,7 @@ void sdp_dreq_wait_timeout_work(struct work_struct *work) goto out; } - sdp_warn(sk, "timed out waiting for FIN/DREQ. " + sdp_dbg(sk, "timed out waiting for FIN/DREQ. " "going into abortive close.\n"); sdp_sk(sk)->dreq_wait_timeout = 0; @@ -1018,7 +1018,7 @@ void sdp_dreq_wait_timeout_work(struct work_struct *work) release_sock(sk); if (sdp_sk(sk)->id) { - sdp_warn(sk, "Destroyed QP!!!!\n"); + sdp_dbg(sk, "Destroyed QP\n"); sdp_sk(sk)->qp_active = 0; rdma_disconnect(sdp_sk(sk)->id); } else @@ -2031,7 +2031,7 @@ static int sdp_recvmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg, } if (rx_sa->flags & RX_SA_ABORTED) { - sdp_warn(sk, "rx_sa aborted. not rdmaing\n"); + sdp_dbg_data(sk, "rx_sa aborted. not rdmaing\n"); goto skb_cleanup; } @@ -2161,7 +2161,7 @@ static int sdp_recvmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg, err = sdp_rdma_to_iovec(sk, msg->msg_iov, skb, used); if (err == -EAGAIN) { - sdp_warn(sk, "RDMA Read aborted\n"); + sdp_dbg_data(sk, "RDMA Read aborted\n"); used = 0; goto skb_cleanup; } diff --git a/drivers/infiniband/ulp/sdp/sdp_rx.c b/drivers/infiniband/ulp/sdp/sdp_rx.c index da5bd38d0d939..3e65b2608cb65 100644 --- a/drivers/infiniband/ulp/sdp/sdp_rx.c +++ b/drivers/infiniband/ulp/sdp/sdp_rx.c @@ -320,7 +320,7 @@ static inline struct sk_buff *sdp_sock_queue_rcv_skb(struct sock *sk, rx_sa->flags = 0; if (ssk->tx_sa) { - sdp_warn(&ssk->isk.sk, "got RX SrcAvail while waiting " + sdp_dbg_data(&ssk->isk.sk, "got RX SrcAvail while waiting " "for TX SrcAvail. waking up TX SrcAvail" "to be aborted\n"); wake_up(sk->sk_sleep); @@ -457,11 +457,11 @@ static int sdp_process_rx_ctl_skb(struct sdp_sock *ssk, struct sk_buff *skb) case SDP_MID_SRCAVAIL: WARN_ON(!(sk->sk_shutdown & RCV_SHUTDOWN)); - sdp_warn(sk, "DATA after socket rcv was shutdown\n"); + sdp_dbg(sk, "DATA after socket rcv was shutdown\n"); /* got data in RCV_SHUTDOWN */ if (sk->sk_state == TCP_FIN_WAIT1) { - sdp_warn(sk, "RX data when state = FIN_WAIT1\n"); + sdp_dbg(sk, "RX data when state = FIN_WAIT1\n"); /* go into abortive close */ sdp_exch_state(sk, TCPF_FIN_WAIT1, TCP_TIME_WAIT); @@ -493,7 +493,7 @@ static int sdp_process_rx_ctl_skb(struct sdp_sock *ssk, struct sk_buff *skb) the dirty logic from recvmsg */ sdp_post_sendsm(sk); } else { - sdp_warn(sk, "Got SrcAvailCancel - " + sdp_dbg(sk, "Got SrcAvailCancel - " "but no SrcAvail in process\n"); } break; @@ -583,7 +583,7 @@ static int sdp_process_rx_skb(struct sdp_sock *ssk, struct sk_buff *skb) unlikely(sk->sk_shutdown & RCV_SHUTDOWN)) { sdp_prf(sk, NULL, "Control skb - queing to control queue"); if (h->mid == SDP_MID_SRCAVAIL_CANCEL) { - sdp_warn(sk, "Got SrcAvailCancel. " + sdp_dbg_data(sk, "Got SrcAvailCancel. " "seq: 0x%d seq_ack: 0x%d\n", ntohl(h->mseq), ntohl(h->mseq_ack)); ssk->srcavail_cancel_mseq = ntohl(h->mseq); @@ -803,7 +803,7 @@ static void sdp_rx_irq(struct ib_cq *cq, void *cq_context) int credits_before; if (cq != ssk->rx_ring.cq) { - sdp_warn(sk, "cq = %p, ssk->cq = %p\n", cq, ssk->rx_ring.cq); + sdp_dbg(sk, "cq = %p, ssk->cq = %p\n", cq, ssk->rx_ring.cq); return; } @@ -812,7 +812,7 @@ static void sdp_rx_irq(struct ib_cq *cq, void *cq_context) sdp_prf(sk, NULL, "rx irq"); if (!rx_ring_trylock(&ssk->rx_ring, &flags)) { - sdp_warn(&ssk->isk.sk, "ring destroyed. not polling it\n"); + sdp_dbg(&ssk->isk.sk, "ring destroyed. not polling it\n"); return; } diff --git a/drivers/infiniband/ulp/sdp/sdp_tx.c b/drivers/infiniband/ulp/sdp/sdp_tx.c index 386ab40f03aa9..573373a4c7681 100644 --- a/drivers/infiniband/ulp/sdp/sdp_tx.c +++ b/drivers/infiniband/ulp/sdp/sdp_tx.c @@ -85,7 +85,7 @@ void sdp_post_send(struct sdp_sock *ssk, struct sk_buff *skb) if (unlikely(h->mid == SDP_MID_SRCAVAIL)) { struct tx_srcavail_state *tx_sa = TX_SRCAVAIL_STATE(skb); if (ssk->tx_sa != tx_sa) { - sdp_warn(&ssk->isk.sk, "SrcAvail cancelled " + sdp_dbg_data(&ssk->isk.sk, "SrcAvail cancelled " "before being sent!\n"); WARN_ON(1); __kfree_skb(skb); @@ -209,7 +209,7 @@ static int sdp_handle_send_comp(struct sdp_sock *ssk, struct ib_wc *wc) struct sock *sk = &ssk->isk.sk; sdp_prf(sk, skb, "Send completion with error. " "Status %d", wc->status); - sdp_warn(sk, "Send completion with error. " + sdp_dbg_data(sk, "Send completion with error. " "Status %d\n", wc->status); sdp_set_error(sk, -ECONNRESET); wake_up(&ssk->wq); @@ -241,12 +241,12 @@ static inline void sdp_process_tx_wc(struct sdp_sock *ssk, struct ib_wc *wc) sdp_prf1(sk, NULL, "TX comp: RDMA read"); if (!ssk->tx_ring.rdma_inflight) { - sdp_warn(sk, "ERROR: unexpected RDMA read\n"); + sdp_dbg(sk, "ERROR: unexpected RDMA read\n"); return; } if (!ssk->tx_ring.rdma_inflight->busy) { - sdp_warn(sk, "ERROR: too many RDMA read completions\n"); + sdp_dbg(sk, "ERROR: too many RDMA read completions\n"); return; } @@ -284,7 +284,7 @@ static int sdp_process_tx_cq(struct sdp_sock *ssk) int wc_processed = 0; if (!ssk->tx_ring.cq) { - sdp_warn(&ssk->isk.sk, "tx irq on destroyed tx_cq\n"); + sdp_dbg(&ssk->isk.sk, "tx irq on destroyed tx_cq\n"); return 0; } diff --git a/drivers/infiniband/ulp/sdp/sdp_zcopy.c b/drivers/infiniband/ulp/sdp/sdp_zcopy.c index a712f3b61e0f2..8754c1b5be252 100644 --- a/drivers/infiniband/ulp/sdp/sdp_zcopy.c +++ b/drivers/infiniband/ulp/sdp/sdp_zcopy.c @@ -100,7 +100,7 @@ static int sdp_post_srcavail_cancel(struct sock *sk) struct sdp_sock *ssk = sdp_sk(sk); struct sk_buff *skb; - sdp_warn(&ssk->isk.sk, "Posting srcavail cancel\n"); + sdp_dbg_data(&ssk->isk.sk, "Posting srcavail cancel\n"); skb = sdp_alloc_skb_srcavail_cancel(sk, 0); skb_entail(sk, ssk, skb); @@ -121,7 +121,7 @@ void srcavail_cancel_timeout(struct work_struct *work) lock_sock(sk); - sdp_warn(sk, "both SrcAvail and SrcAvailCancel timedout." + sdp_dbg_data(sk, "both SrcAvail and SrcAvailCancel timedout." " closing connection\n"); sdp_set_error(sk, -ECONNRESET); wake_up(&ssk->wq); @@ -173,7 +173,7 @@ static int sdp_wait_rdmardcompl(struct sdp_sock *ssk, long *timeo_p, int len, } if (ssk->rx_sa) { - sdp_warn(sk, "Crossing SrcAvail - aborting this\n"); + sdp_dbg_data(sk, "Crossing SrcAvail - aborting this\n"); tx_sa->abort_flags |= TX_SA_CROSS_SEND; err = -ETIME; break ; @@ -211,7 +211,7 @@ static int sdp_wait_rdmardcompl(struct sdp_sock *ssk, long *timeo_p, int len, tx_sa->bytes_acked, tx_sa->bytes_sent, tx_sa->abort_flags); if (!ssk->qp_active) { - sdp_warn(sk, "QP destroyed while waiting\n"); + sdp_dbg(sk, "QP destroyed while waiting\n"); return -EINVAL; } return err; @@ -230,19 +230,19 @@ static int sdp_wait_rdma_wr_finished(struct sdp_sock *ssk, long *timeo_p) if (unlikely(!ssk->qp_active)) { err = -EPIPE; - sdp_warn(sk, "socket closed\n"); + sdp_dbg(sk, "socket closed\n"); break; } if (unlikely(!*timeo_p)) { err = -EAGAIN; - sdp_warn(sk, "timedout\n"); + sdp_dbg(sk, "timedout\n"); break; } if (unlikely(signal_pending(current))) { err = sock_intr_errno(*timeo_p); - sdp_warn(sk, "signalled\n"); + sdp_dbg_data(sk, "signalled\n"); break; } @@ -341,19 +341,19 @@ void sdp_handle_sendsm(struct sdp_sock *ssk, u32 mseq_ack) } if (mseq_ack < ssk->tx_sa->mseq) { - sdp_warn(sk, "SendSM arrived for old SrcAvail. " + sdp_dbg_data(sk, "SendSM arrived for old SrcAvail. " "SendSM mseq_ack: 0x%x, SrcAvail mseq: 0x%x\n", mseq_ack, ssk->tx_sa->mseq); goto out; } - sdp_warn(sk, "Got SendSM - aborting SrcAvail\n"); + sdp_dbg_data(sk, "Got SendSM - aborting SrcAvail\n"); ssk->tx_sa->abort_flags |= TX_SA_SENDSM; cancel_delayed_work(&ssk->srcavail_cancel_work); wake_up(sk->sk_sleep); - sdp_warn(sk, "woke up sleepers\n"); + sdp_dbg_data(sk, "woke up sleepers\n"); out: spin_unlock_irqrestore(&ssk->tx_sa_lock, flags); @@ -373,12 +373,12 @@ void sdp_handle_rdma_read_compl(struct sdp_sock *ssk, u32 mseq_ack, BUG_ON(!ssk); if (!ssk->tx_sa) { - sdp_warn(sk, "Got RdmaRdCompl for aborted SrcAvail\n"); + sdp_dbg_data(sk, "Got RdmaRdCompl for aborted SrcAvail\n"); goto out; } if (ssk->tx_sa->mseq < mseq_ack) { - sdp_warn(sk, "RdmaRdCompl arrived for old SrcAvail. " + sdp_dbg_data(sk, "RdmaRdCompl arrived for old SrcAvail. " "SendSM mseq_ack: 0x%x, SrcAvail mseq: 0x%x\n", mseq_ack, ssk->tx_sa->mseq); goto out; @@ -453,7 +453,7 @@ int sdp_get_pages(struct sock *sk, struct page **pages, int page_cnt, return 0; err: - sdp_warn(sk, "Error getting pages. done_pages: %d page_cnt: %d\n", + sdp_dbg(sk, "Error getting pages. done_pages: %d page_cnt: %d\n", done_pages, page_cnt); for (; done_pages > 0; done_pages--) page_cache_release(pages[done_pages - 1]); @@ -718,7 +718,7 @@ int sdp_rdma_to_iovec(struct sock *sk, struct iovec *iov, struct sk_buff *skb, sge += sge_cnt; if (unlikely(ssk->srcavail_cancel_mseq > rx_sa->mseq)) { - sdp_warn(sk, "got SrcAvailCancel - Aborting RDMA\n"); + sdp_dbg_data(sk, "got SrcAvailCancel - Aborting RDMA\n"); rc = -EAGAIN; } } while (!rc && sge_left > 0); @@ -741,7 +741,7 @@ int sdp_rdma_to_iovec(struct sock *sk, struct iovec *iov, struct sk_buff *skb, if (rc && ssk->qp_active) { /* post rdma, wait_for_compl or post rdma_rd_comp failed - * post sendsm */ - sdp_warn(sk, "post rdma, wait_for_compl " + sdp_dbg_data(sk, "post rdma, wait_for_compl " "or post rdma_rd_comp failed - post sendsm\n"); rx_sa->flags |= RX_SA_ABORTED; ssk->rx_sa = NULL; /* TODO: change it into SDP_MID_DATA and get @@ -811,7 +811,7 @@ static int sdp_rdma_adv_single(struct sock *sk, rc = sdp_post_srcavail(sk, tx_sa, 0, offset, len); if (rc) { - sdp_warn(sk, "Error posting SrcAvail\n"); + sdp_dbg(sk, "Error posting SrcAvail\n"); goto err_abort_send; } @@ -820,15 +820,15 @@ static int sdp_rdma_adv_single(struct sock *sk, enum tx_sa_flag f = tx_sa->abort_flags; if (f & TX_SA_SENDSM) { - sdp_warn(sk, "got SendSM. use SEND verb.\n"); + sdp_dbg_data(sk, "got SendSM. use SEND verb.\n"); } else if (f & TX_SA_ERROR) { - sdp_warn(sk, "SrcAvail error completion\n"); + sdp_dbg_data(sk, "SrcAvail error completion\n"); sdp_reset(sk); } else if (ssk->qp_active) { if (f & TX_SA_INTRRUPTED) sdp_dbg_data(sk, "SrcAvail error completion\n"); else - sdp_warn(sk, "abort_flag = 0x%x.\n", f); + sdp_dbg_data(sk, "abort_flag = 0x%x.\n", f); sdp_post_srcavail_cancel(sk); @@ -839,7 +839,7 @@ static int sdp_rdma_adv_single(struct sock *sk, sdp_wait_rdmardcompl(ssk, &timeo, len, 1); sdp_dbg_data(sk, "finished waiting\n"); } else { - sdp_warn(sk, "QP was destroyed while waiting\n"); + sdp_dbg_data(sk, "QP was destroyed while waiting\n"); } goto err_abort_send; @@ -882,7 +882,7 @@ int sdp_sendmsg_zcopy(struct kiocb *iocb, struct sock *sk, struct msghdr *msg, sdp_dbg_data(sk, "%s\n", __func__); sdp_prf1(sk, NULL, "sdp_sendmsg_zcopy start"); if (ssk->rx_sa) { - sdp_warn(sk, "Deadlock prevent: crossing SrcAvail\n"); + sdp_dbg_data(sk, "Deadlock prevent: crossing SrcAvail\n"); return -EAGAIN; } @@ -955,7 +955,7 @@ int sdp_sendmsg_zcopy(struct kiocb *iocb, struct sock *sk, struct msghdr *msg, rc = sdp_rdma_adv_single(sk, tx_sa, iov, page_cnt, off, len); if (rc) - sdp_warn(sk, "Error sending SrcAvail. rc = %d\n", rc); + sdp_dbg_data(sk, "Error sending SrcAvail. rc = %d\n", rc); if (tx_sa->addrs)