rc = rdma_resolve_route(id, SDP_ROUTE_TIMEOUT);
break;
case RDMA_CM_EVENT_ADDR_ERROR:
- sdp_warn(sk, "RDMA_CM_EVENT_ADDR_ERROR\n");
+ sdp_dbg(sk, "RDMA_CM_EVENT_ADDR_ERROR\n");
rc = -ENETUNREACH;
break;
case RDMA_CM_EVENT_ROUTE_RESOLVED:
sdp_dbg(sk, "RDMA_CM_EVENT_CONNECT_REQUEST\n");
rc = sdp_connect_handler(sk, id, event);
if (rc) {
- sdp_warn(sk, "Destroy qp !!!!\n");
+ sdp_dbg(sk, "Destroying qp\n");
rdma_reject(id, NULL, 0);
break;
}
sdp_dbg(sk, "RDMA_CM_EVENT_CONNECT_RESPONSE\n");
rc = sdp_response_handler(sk, id, event);
if (rc) {
- sdp_warn(sk, "Destroy qp !!!!\n");
+ sdp_dbg(sk, "Destroying qp\n");
rdma_reject(id, NULL, 0);
} else
rc = rdma_accept(id, NULL);
rc = sdp_disconnected_handler(sk);
break;
case RDMA_CM_EVENT_DEVICE_REMOVAL:
- sdp_warn(sk, "RDMA_CM_EVENT_DEVICE_REMOVAL\n");
+ sdp_dbg(sk, "RDMA_CM_EVENT_DEVICE_REMOVAL\n");
rc = -ENETRESET;
break;
default:
rc = rdma_bind_addr(ssk->id, (struct sockaddr *)&addr);
if (rc) {
- sdp_warn(sk, "Destroy qp !!!!\n");
+ sdp_dbg(sk, "Destroying qp\n");
rdma_destroy_id(ssk->id);
ssk->id = NULL;
return rc;
rc = rdma_resolve_addr(ssk->id, (struct sockaddr *)&src_addr,
uaddr, SDP_RESOLVE_TIMEOUT);
if (rc) {
- sdp_warn(sk, "rdma_resolve_addr failed: %d\n", rc);
+ sdp_dbg(sk, "rdma_resolve_addr failed: %d\n", rc);
return rc;
}
goto out;
}
- sdp_warn(sk, "timed out waiting for FIN/DREQ. "
+ sdp_dbg(sk, "timed out waiting for FIN/DREQ. "
"going into abortive close.\n");
sdp_sk(sk)->dreq_wait_timeout = 0;
release_sock(sk);
if (sdp_sk(sk)->id) {
- sdp_warn(sk, "Destroyed QP!!!!\n");
+ sdp_dbg(sk, "Destroyed QP\n");
sdp_sk(sk)->qp_active = 0;
rdma_disconnect(sdp_sk(sk)->id);
} else
}
if (rx_sa->flags & RX_SA_ABORTED) {
- sdp_warn(sk, "rx_sa aborted. not rdmaing\n");
+ sdp_dbg_data(sk, "rx_sa aborted. not rdmaing\n");
goto skb_cleanup;
}
err = sdp_rdma_to_iovec(sk, msg->msg_iov, skb,
used);
if (err == -EAGAIN) {
- sdp_warn(sk, "RDMA Read aborted\n");
+ sdp_dbg_data(sk, "RDMA Read aborted\n");
used = 0;
goto skb_cleanup;
}
rx_sa->flags = 0;
if (ssk->tx_sa) {
- sdp_warn(&ssk->isk.sk, "got RX SrcAvail while waiting "
+ sdp_dbg_data(&ssk->isk.sk, "got RX SrcAvail while waiting "
"for TX SrcAvail. waking up TX SrcAvail"
"to be aborted\n");
wake_up(sk->sk_sleep);
case SDP_MID_SRCAVAIL:
WARN_ON(!(sk->sk_shutdown & RCV_SHUTDOWN));
- sdp_warn(sk, "DATA after socket rcv was shutdown\n");
+ sdp_dbg(sk, "DATA after socket rcv was shutdown\n");
/* got data in RCV_SHUTDOWN */
if (sk->sk_state == TCP_FIN_WAIT1) {
- sdp_warn(sk, "RX data when state = FIN_WAIT1\n");
+ sdp_dbg(sk, "RX data when state = FIN_WAIT1\n");
/* go into abortive close */
sdp_exch_state(sk, TCPF_FIN_WAIT1,
TCP_TIME_WAIT);
the dirty logic from recvmsg */
sdp_post_sendsm(sk);
} else {
- sdp_warn(sk, "Got SrcAvailCancel - "
+ sdp_dbg(sk, "Got SrcAvailCancel - "
"but no SrcAvail in process\n");
}
break;
unlikely(sk->sk_shutdown & RCV_SHUTDOWN)) {
sdp_prf(sk, NULL, "Control skb - queing to control queue");
if (h->mid == SDP_MID_SRCAVAIL_CANCEL) {
- sdp_warn(sk, "Got SrcAvailCancel. "
+ sdp_dbg_data(sk, "Got SrcAvailCancel. "
"seq: 0x%d seq_ack: 0x%d\n",
ntohl(h->mseq), ntohl(h->mseq_ack));
ssk->srcavail_cancel_mseq = ntohl(h->mseq);
int credits_before;
if (cq != ssk->rx_ring.cq) {
- sdp_warn(sk, "cq = %p, ssk->cq = %p\n", cq, ssk->rx_ring.cq);
+ sdp_dbg(sk, "cq = %p, ssk->cq = %p\n", cq, ssk->rx_ring.cq);
return;
}
sdp_prf(sk, NULL, "rx irq");
if (!rx_ring_trylock(&ssk->rx_ring, &flags)) {
- sdp_warn(&ssk->isk.sk, "ring destroyed. not polling it\n");
+ sdp_dbg(&ssk->isk.sk, "ring destroyed. not polling it\n");
return;
}
if (unlikely(h->mid == SDP_MID_SRCAVAIL)) {
struct tx_srcavail_state *tx_sa = TX_SRCAVAIL_STATE(skb);
if (ssk->tx_sa != tx_sa) {
- sdp_warn(&ssk->isk.sk, "SrcAvail cancelled "
+ sdp_dbg_data(&ssk->isk.sk, "SrcAvail cancelled "
"before being sent!\n");
WARN_ON(1);
__kfree_skb(skb);
struct sock *sk = &ssk->isk.sk;
sdp_prf(sk, skb, "Send completion with error. "
"Status %d", wc->status);
- sdp_warn(sk, "Send completion with error. "
+ sdp_dbg_data(sk, "Send completion with error. "
"Status %d\n", wc->status);
sdp_set_error(sk, -ECONNRESET);
wake_up(&ssk->wq);
sdp_prf1(sk, NULL, "TX comp: RDMA read");
if (!ssk->tx_ring.rdma_inflight) {
- sdp_warn(sk, "ERROR: unexpected RDMA read\n");
+ sdp_dbg(sk, "ERROR: unexpected RDMA read\n");
return;
}
if (!ssk->tx_ring.rdma_inflight->busy) {
- sdp_warn(sk, "ERROR: too many RDMA read completions\n");
+ sdp_dbg(sk, "ERROR: too many RDMA read completions\n");
return;
}
int wc_processed = 0;
if (!ssk->tx_ring.cq) {
- sdp_warn(&ssk->isk.sk, "tx irq on destroyed tx_cq\n");
+ sdp_dbg(&ssk->isk.sk, "tx irq on destroyed tx_cq\n");
return 0;
}
struct sdp_sock *ssk = sdp_sk(sk);
struct sk_buff *skb;
- sdp_warn(&ssk->isk.sk, "Posting srcavail cancel\n");
+ sdp_dbg_data(&ssk->isk.sk, "Posting srcavail cancel\n");
skb = sdp_alloc_skb_srcavail_cancel(sk, 0);
skb_entail(sk, ssk, skb);
lock_sock(sk);
- sdp_warn(sk, "both SrcAvail and SrcAvailCancel timedout."
+ sdp_dbg_data(sk, "both SrcAvail and SrcAvailCancel timedout."
" closing connection\n");
sdp_set_error(sk, -ECONNRESET);
wake_up(&ssk->wq);
}
if (ssk->rx_sa) {
- sdp_warn(sk, "Crossing SrcAvail - aborting this\n");
+ sdp_dbg_data(sk, "Crossing SrcAvail - aborting this\n");
tx_sa->abort_flags |= TX_SA_CROSS_SEND;
err = -ETIME;
break ;
tx_sa->bytes_acked, tx_sa->bytes_sent, tx_sa->abort_flags);
if (!ssk->qp_active) {
- sdp_warn(sk, "QP destroyed while waiting\n");
+ sdp_dbg(sk, "QP destroyed while waiting\n");
return -EINVAL;
}
return err;
if (unlikely(!ssk->qp_active)) {
err = -EPIPE;
- sdp_warn(sk, "socket closed\n");
+ sdp_dbg(sk, "socket closed\n");
break;
}
if (unlikely(!*timeo_p)) {
err = -EAGAIN;
- sdp_warn(sk, "timedout\n");
+ sdp_dbg(sk, "timedout\n");
break;
}
if (unlikely(signal_pending(current))) {
err = sock_intr_errno(*timeo_p);
- sdp_warn(sk, "signalled\n");
+ sdp_dbg_data(sk, "signalled\n");
break;
}
}
if (mseq_ack < ssk->tx_sa->mseq) {
- sdp_warn(sk, "SendSM arrived for old SrcAvail. "
+ sdp_dbg_data(sk, "SendSM arrived for old SrcAvail. "
"SendSM mseq_ack: 0x%x, SrcAvail mseq: 0x%x\n",
mseq_ack, ssk->tx_sa->mseq);
goto out;
}
- sdp_warn(sk, "Got SendSM - aborting SrcAvail\n");
+ sdp_dbg_data(sk, "Got SendSM - aborting SrcAvail\n");
ssk->tx_sa->abort_flags |= TX_SA_SENDSM;
cancel_delayed_work(&ssk->srcavail_cancel_work);
wake_up(sk->sk_sleep);
- sdp_warn(sk, "woke up sleepers\n");
+ sdp_dbg_data(sk, "woke up sleepers\n");
out:
spin_unlock_irqrestore(&ssk->tx_sa_lock, flags);
BUG_ON(!ssk);
if (!ssk->tx_sa) {
- sdp_warn(sk, "Got RdmaRdCompl for aborted SrcAvail\n");
+ sdp_dbg_data(sk, "Got RdmaRdCompl for aborted SrcAvail\n");
goto out;
}
if (ssk->tx_sa->mseq < mseq_ack) {
- sdp_warn(sk, "RdmaRdCompl arrived for old SrcAvail. "
+ sdp_dbg_data(sk, "RdmaRdCompl arrived for old SrcAvail. "
"SendSM mseq_ack: 0x%x, SrcAvail mseq: 0x%x\n",
mseq_ack, ssk->tx_sa->mseq);
goto out;
return 0;
err:
- sdp_warn(sk, "Error getting pages. done_pages: %d page_cnt: %d\n",
+ sdp_dbg(sk, "Error getting pages. done_pages: %d page_cnt: %d\n",
done_pages, page_cnt);
for (; done_pages > 0; done_pages--)
page_cache_release(pages[done_pages - 1]);
sge += sge_cnt;
if (unlikely(ssk->srcavail_cancel_mseq > rx_sa->mseq)) {
- sdp_warn(sk, "got SrcAvailCancel - Aborting RDMA\n");
+ sdp_dbg_data(sk, "got SrcAvailCancel - Aborting RDMA\n");
rc = -EAGAIN;
}
} while (!rc && sge_left > 0);
if (rc && ssk->qp_active) {
/* post rdma, wait_for_compl or post rdma_rd_comp failed -
* post sendsm */
- sdp_warn(sk, "post rdma, wait_for_compl "
+ sdp_dbg_data(sk, "post rdma, wait_for_compl "
"or post rdma_rd_comp failed - post sendsm\n");
rx_sa->flags |= RX_SA_ABORTED;
ssk->rx_sa = NULL; /* TODO: change it into SDP_MID_DATA and get
rc = sdp_post_srcavail(sk, tx_sa, 0, offset, len);
if (rc) {
- sdp_warn(sk, "Error posting SrcAvail\n");
+ sdp_dbg(sk, "Error posting SrcAvail\n");
goto err_abort_send;
}
enum tx_sa_flag f = tx_sa->abort_flags;
if (f & TX_SA_SENDSM) {
- sdp_warn(sk, "got SendSM. use SEND verb.\n");
+ sdp_dbg_data(sk, "got SendSM. use SEND verb.\n");
} else if (f & TX_SA_ERROR) {
- sdp_warn(sk, "SrcAvail error completion\n");
+ sdp_dbg_data(sk, "SrcAvail error completion\n");
sdp_reset(sk);
} else if (ssk->qp_active) {
if (f & TX_SA_INTRRUPTED)
sdp_dbg_data(sk, "SrcAvail error completion\n");
else
- sdp_warn(sk, "abort_flag = 0x%x.\n", f);
+ sdp_dbg_data(sk, "abort_flag = 0x%x.\n", f);
sdp_post_srcavail_cancel(sk);
sdp_wait_rdmardcompl(ssk, &timeo, len, 1);
sdp_dbg_data(sk, "finished waiting\n");
} else {
- sdp_warn(sk, "QP was destroyed while waiting\n");
+ sdp_dbg_data(sk, "QP was destroyed while waiting\n");
}
goto err_abort_send;
sdp_dbg_data(sk, "%s\n", __func__);
sdp_prf1(sk, NULL, "sdp_sendmsg_zcopy start");
if (ssk->rx_sa) {
- sdp_warn(sk, "Deadlock prevent: crossing SrcAvail\n");
+ sdp_dbg_data(sk, "Deadlock prevent: crossing SrcAvail\n");
return -EAGAIN;
}
rc = sdp_rdma_adv_single(sk, tx_sa, iov, page_cnt, off, len);
if (rc)
- sdp_warn(sk, "Error sending SrcAvail. rc = %d\n", rc);
+ sdp_dbg_data(sk, "Error sending SrcAvail. rc = %d\n", rc);
if (tx_sa->addrs)