static bool codel_should_drop(const struct sk_buff *skb,
-                             unsigned int *backlog,
+                             struct Qdisc *sch,
                              struct codel_vars *vars,
                              struct codel_params *params,
                              struct codel_stats *stats,
        }
 
        vars->ldelay = now - codel_get_enqueue_time(skb);
-       *backlog -= qdisc_pkt_len(skb);
+       sch->qstats.backlog -= qdisc_pkt_len(skb);
 
        if (unlikely(qdisc_pkt_len(skb) > stats->maxpacket))
                stats->maxpacket = qdisc_pkt_len(skb);
 
        if (codel_time_before(vars->ldelay, params->target) ||
-           *backlog <= stats->maxpacket) {
+           sch->qstats.backlog <= stats->maxpacket) {
                /* went below - stay below for at least interval */
                vars->first_above_time = 0;
                return false;
                                     struct codel_params *params,
                                     struct codel_vars *vars,
                                     struct codel_stats *stats,
-                                    codel_skb_dequeue_t dequeue_func,
-                                    u32 *backlog)
+                                    codel_skb_dequeue_t dequeue_func)
 {
        struct sk_buff *skb = dequeue_func(vars, sch);
        codel_time_t now;
                return skb;
        }
        now = codel_get_time();
-       drop = codel_should_drop(skb, backlog, vars, params, stats, now);
+       drop = codel_should_drop(skb, sch, vars, params, stats, now);
        if (vars->dropping) {
                if (!drop) {
                        /* sojourn time below target - leave dropping state */
                                qdisc_drop(skb, sch);
                                stats->drop_count++;
                                skb = dequeue_func(vars, sch);
-                               if (!codel_should_drop(skb, backlog,
+                               if (!codel_should_drop(skb, sch,
                                                       vars, params, stats, now)) {
                                        /* leave dropping state */
                                        vars->dropping = false;
                        stats->drop_count++;
 
                        skb = dequeue_func(vars, sch);
-                       drop = codel_should_drop(skb, backlog, vars, params,
+                       drop = codel_should_drop(skb, sch, vars, params,
                                                 stats, now);
                }
                vars->dropping = true;
 
        struct codel_sched_data *q = qdisc_priv(sch);
        struct sk_buff *skb;
 
-       skb = codel_dequeue(sch, &q->params, &q->vars, &q->stats,
-                           dequeue, &sch->qstats.backlog);
+       skb = codel_dequeue(sch, &q->params, &q->vars, &q->stats, dequeue);
+
        /* We cant call qdisc_tree_decrease_qlen() if our qlen is 0,
         * or HTB crashes. Defer it for next round.
         */
 
  */
 static struct sk_buff *dequeue(struct codel_vars *vars, struct Qdisc *sch)
 {
+       struct fq_codel_sched_data *q = qdisc_priv(sch);
        struct fq_codel_flow *flow;
        struct sk_buff *skb = NULL;
 
        flow = container_of(vars, struct fq_codel_flow, cvars);
        if (flow->head) {
                skb = dequeue_head(flow);
-               sch->qstats.backlog -= qdisc_pkt_len(skb);
+               q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb);
                sch->q.qlen--;
        }
        return skb;
        prev_ecn_mark = q->cstats.ecn_mark;
 
        skb = codel_dequeue(sch, &q->cparams, &flow->cvars, &q->cstats,
-                           dequeue, &q->backlogs[flow - q->flows]);
+                           dequeue);
 
        flow->dropped += q->cstats.drop_count - prev_drop_count;
        flow->dropped += q->cstats.ecn_mark - prev_ecn_mark;