From ae4972ffa272cae3b47120800dc5f01233c710ee Mon Sep 17 00:00:00 2001 From: Eldad Zinger Date: Tue, 4 May 2010 12:05:56 +0300 Subject: [PATCH] sdp: debug message for reference count changed from CM_TW to CMA This debug message marks that the sdp_cma_handler() is expected to be invoked. Signed-off-by: Eldad Zinger --- drivers/infiniband/ulp/sdp/sdp_cma.c | 2 +- drivers/infiniband/ulp/sdp/sdp_dbg.h | 2 +- drivers/infiniband/ulp/sdp/sdp_main.c | 8 ++++---- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/drivers/infiniband/ulp/sdp/sdp_cma.c b/drivers/infiniband/ulp/sdp/sdp_cma.c index fdd98a720ec1..e2816544f771 100644 --- a/drivers/infiniband/ulp/sdp/sdp_cma.c +++ b/drivers/infiniband/ulp/sdp/sdp_cma.c @@ -474,7 +474,7 @@ int sdp_cma_handler(struct rdma_cm_id *id, struct rdma_cm_event *event) sdp_dbg(sk, "IB teardown while in " "TCP_CLOSE_WAIT taking reference to " "let close() finish the work\n"); - sock_hold(sk, SOCK_REF_CM_TW); + sock_hold(sk, SOCK_REF_CMA); } sdp_set_error(sk, EPIPE); rc = sdp_disconnected_handler(sk); diff --git a/drivers/infiniband/ulp/sdp/sdp_dbg.h b/drivers/infiniband/ulp/sdp/sdp_dbg.h index f25e86dc5e5c..2c994a44defb 100644 --- a/drivers/infiniband/ulp/sdp/sdp_dbg.h +++ b/drivers/infiniband/ulp/sdp/sdp_dbg.h @@ -145,7 +145,7 @@ extern int sdp_data_debug_level; #define SOCK_REF_RESET "RESET" #define SOCK_REF_ALIVE "ALIVE" /* sock_alloc -> destruct_sock */ #define SOCK_REF_CLONE "CLONE" -#define SOCK_REF_CM_TW "CM_TW" /* TIMEWAIT_ENTER -> TIMEWAIT_EXIT */ +#define SOCK_REF_CMA "CMA" /* sdp_cma_handler() is expected to be invoked */ #define SOCK_REF_SEQ "SEQ" /* during proc read */ #define SOCK_REF_DREQ_TO "DREQ_TO" /* dreq timeout is pending */ #define SOCK_REF_ZCOPY "ZCOPY" /* zcopy send in process */ diff --git a/drivers/infiniband/ulp/sdp/sdp_main.c b/drivers/infiniband/ulp/sdp/sdp_main.c index bf1752a9ee5e..a4d88c661184 100644 --- a/drivers/infiniband/ulp/sdp/sdp_main.c +++ b/drivers/infiniband/ulp/sdp/sdp_main.c @@ -625,7 +625,7 @@ static void sdp_close(struct sock *sk, long timeout) goto adjudge_to_death; } - sock_hold(sk, SOCK_REF_CM_TW); + sock_hold(sk, SOCK_REF_CMA); /* We need to flush the recv. buffs. We do this only on the * descriptor close, not protocol-sourced closes, because the @@ -1002,7 +1002,7 @@ static void sdp_destroy_work(struct work_struct *work) cancel_delayed_work(&ssk->srcavail_cancel_work); if (sk->sk_state == TCP_TIME_WAIT) - sock_put(sk, SOCK_REF_CM_TW); + sock_put(sk, SOCK_REF_CMA); /* In normal close current state is TCP_TIME_WAIT or TCP_CLOSE but if a CM connection is dropped below our legs state could @@ -1045,7 +1045,7 @@ static void sdp_dreq_wait_timeout_work(struct work_struct *work) sdp_sk(sk)->qp_active = 0; rdma_disconnect(sdp_sk(sk)->id); } else - sock_put(sk, SOCK_REF_CM_TW); + sock_put(sk, SOCK_REF_CMA); out: sock_put(sk, SOCK_REF_DREQ_TO); @@ -2745,7 +2745,7 @@ kill_socks: if ((1 << sk->sk_state) & (TCPF_FIN_WAIT1 | TCPF_CLOSE_WAIT | TCPF_LAST_ACK | TCPF_TIME_WAIT)) { - sock_put(sk, SOCK_REF_CM_TW); + sock_put(sk, SOCK_REF_CMA); } schedule(); -- 2.50.1