static void release_tid(struct c4iw_rdev *rdev, u32 hwtid, struct sk_buff *skb)
 {
-       struct cpl_tid_release *req;
+       u32 len = roundup(sizeof(struct cpl_tid_release), 16);
 
-       skb = get_skb(skb, sizeof *req, GFP_KERNEL);
+       skb = get_skb(skb, len, GFP_KERNEL);
        if (!skb)
                return;
-       req = (struct cpl_tid_release *) skb_put(skb, sizeof(*req));
-       INIT_TP_WR(req, hwtid);
-       OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_TID_RELEASE, hwtid));
-       set_wr_txq(skb, CPL_PRIORITY_SETUP, 0);
+
+       cxgb_mk_tid_release(skb, len, hwtid, 0);
        c4iw_ofld_send(rdev, skb);
        return;
 }
 
                wscale++;
        return wscale;
 }
+
+static inline void
+cxgb_mk_tid_release(struct sk_buff *skb, u32 len, u32 tid, u16 chan)
+{
+       struct cpl_tid_release *req;
+
+       req = (struct cpl_tid_release *)__skb_put(skb, len);
+       memset(req, 0, len);
+
+       INIT_TP_WR(req, tid);
+       OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(CPL_TID_RELEASE, tid));
+       set_wr_txq(skb, CPL_PRIORITY_SETUP, chan);
+}
 #endif
 
 
 static void cxgbit_release_tid(struct cxgbit_device *cdev, u32 tid)
 {
-       struct cpl_tid_release *req;
-       unsigned int len = roundup(sizeof(*req), 16);
+       u32 len = roundup(sizeof(struct cpl_tid_release), 16);
        struct sk_buff *skb;
 
        skb = alloc_skb(len, GFP_ATOMIC);
        if (!skb)
                return;
 
-       req = (struct cpl_tid_release *)__skb_put(skb, len);
-       memset(req, 0, len);
-
-       INIT_TP_WR(req, tid);
-       OPCODE_TID(req) = cpu_to_be32(MK_OPCODE_TID(
-                  CPL_TID_RELEASE, tid));
-       set_wr_txq(skb, CPL_PRIORITY_SETUP, 0);
+       cxgb_mk_tid_release(skb, len, tid, 0);
        cxgbit_ofld_send(cdev, skb);
 }