}
 
 struct rsvd_count {
-       int ndelonly;
+       int ndelayed;
        bool first_do_lblk_found;
        ext4_lblk_t first_do_lblk;
        ext4_lblk_t last_do_lblk;
        struct ext4_sb_info *sbi = EXT4_SB(inode->i_sb);
        struct rb_node *node;
 
-       rc->ndelonly = 0;
+       rc->ndelayed = 0;
 
        /*
-        * for bigalloc, note the first delonly block in the range has not
+        * for bigalloc, note the first delayed block in the range has not
         * been found, record the extent containing the block to the left of
         * the region to be removed, if any, and note that there's no partial
         * cluster to track
 }
 
 /*
- * count_rsvd - count the clusters containing delayed and not unwritten
- *             (delonly) blocks in a range within an extent and add to
- *             the running tally in rsvd_count
+ * count_rsvd - count the clusters containing delayed blocks in a range
+ *             within an extent and add to the running tally in rsvd_count
  *
  * @inode - file containing extent
  * @lblk - first block in range
        WARN_ON(len <= 0);
 
        if (sbi->s_cluster_ratio == 1) {
-               rc->ndelonly += (int) len;
+               rc->ndelayed += (int) len;
                return;
        }
 
        end = lblk + (ext4_lblk_t) len - 1;
        end = (end > ext4_es_end(es)) ? ext4_es_end(es) : end;
 
-       /* record the first block of the first delonly extent seen */
+       /* record the first block of the first delayed extent seen */
        if (!rc->first_do_lblk_found) {
                rc->first_do_lblk = i;
                rc->first_do_lblk_found = true;
         * doesn't start with it, count it and stop tracking
         */
        if (rc->partial && (rc->lclu != EXT4_B2C(sbi, i))) {
-               rc->ndelonly++;
+               rc->ndelayed++;
                rc->partial = false;
        }
 
         */
        if (EXT4_LBLK_COFF(sbi, i) != 0) {
                if (end >= EXT4_LBLK_CFILL(sbi, i)) {
-                       rc->ndelonly++;
+                       rc->ndelayed++;
                        rc->partial = false;
                        i = EXT4_LBLK_CFILL(sbi, i) + 1;
                }
 
        /*
         * if the current cluster starts on a cluster boundary, count the
-        * number of whole delonly clusters in the extent
+        * number of whole delayed clusters in the extent
         */
        if ((i + sbi->s_cluster_ratio - 1) <= end) {
                nclu = (end - i + 1) >> sbi->s_cluster_bits;
-               rc->ndelonly += nclu;
+               rc->ndelayed += nclu;
                i += nclu << sbi->s_cluster_bits;
        }
 
  * @rc - pointer to reserved count data
  *
  * The number of reservations to be released is equal to the number of
- * clusters containing delayed and not unwritten (delonly) blocks within
- * the range, minus the number of clusters still containing delonly blocks
- * at the ends of the range, and minus the number of pending reservations
- * within the range.
+ * clusters containing delayed blocks within the range, minus the number of
+ * clusters still containing delayed blocks at the ends of the range, and
+ * minus the number of pending reservations within the range.
  */
 static unsigned int get_rsvd(struct inode *inode, ext4_lblk_t end,
                             struct extent_status *right_es,
        struct ext4_pending_tree *tree = &EXT4_I(inode)->i_pending_tree;
        struct rb_node *node;
        ext4_lblk_t first_lclu, last_lclu;
-       bool left_delonly, right_delonly, count_pending;
+       bool left_delayed, right_delayed, count_pending;
        struct extent_status *es;
 
        if (sbi->s_cluster_ratio > 1) {
                /* count any remaining partial cluster */
                if (rc->partial)
-                       rc->ndelonly++;
+                       rc->ndelayed++;
 
-               if (rc->ndelonly == 0)
+               if (rc->ndelayed == 0)
                        return 0;
 
                first_lclu = EXT4_B2C(sbi, rc->first_do_lblk);
                last_lclu = EXT4_B2C(sbi, rc->last_do_lblk);
 
                /*
-                * decrease the delonly count by the number of clusters at the
-                * ends of the range that still contain delonly blocks -
+                * decrease the delayed count by the number of clusters at the
+                * ends of the range that still contain delayed blocks -
                 * these clusters still need to be reserved
                 */
-               left_delonly = right_delonly = false;
+               left_delayed = right_delayed = false;
 
                es = rc->left_es;
                while (es && ext4_es_end(es) >=
                       EXT4_LBLK_CMASK(sbi, rc->first_do_lblk)) {
                        if (ext4_es_is_delayed(es)) {
-                               rc->ndelonly--;
-                               left_delonly = true;
+                               rc->ndelayed--;
+                               left_delayed = true;
                                break;
                        }
                        node = rb_prev(&es->rb_node);
                                break;
                        es = rb_entry(node, struct extent_status, rb_node);
                }
-               if (right_es && (!left_delonly || first_lclu != last_lclu)) {
+               if (right_es && (!left_delayed || first_lclu != last_lclu)) {
                        if (end < ext4_es_end(right_es)) {
                                es = right_es;
                        } else {
                        while (es && es->es_lblk <=
                               EXT4_LBLK_CFILL(sbi, rc->last_do_lblk)) {
                                if (ext4_es_is_delayed(es)) {
-                                       rc->ndelonly--;
-                                       right_delonly = true;
+                                       rc->ndelayed--;
+                                       right_delayed = true;
                                        break;
                                }
                                node = rb_next(&es->rb_node);
                /*
                 * Determine the block range that should be searched for
                 * pending reservations, if any.  Clusters on the ends of the
-                * original removed range containing delonly blocks are
+                * original removed range containing delayed blocks are
                 * excluded.  They've already been accounted for and it's not
                 * possible to determine if an associated pending reservation
                 * should be released with the information available in the
                 * extents status tree.
                 */
                if (first_lclu == last_lclu) {
-                       if (left_delonly | right_delonly)
+                       if (left_delayed | right_delayed)
                                count_pending = false;
                        else
                                count_pending = true;
                } else {
-                       if (left_delonly)
+                       if (left_delayed)
                                first_lclu++;
-                       if (right_delonly)
+                       if (right_delayed)
                                last_lclu--;
                        if (first_lclu <= last_lclu)
                                count_pending = true;
                /*
                 * a pending reservation found between first_lclu and last_lclu
                 * represents an allocated cluster that contained at least one
-                * delonly block, so the delonly total must be reduced by one
+                * delayed block, so the delayed total must be reduced by one
                 * for each pending reservation found and released
                 */
                if (count_pending) {
                        pr = __pr_tree_search(&tree->root, first_lclu);
                        while (pr && pr->lclu <= last_lclu) {
-                               rc->ndelonly--;
+                               rc->ndelayed--;
                                node = rb_next(&pr->rb_node);
                                rb_erase(&pr->rb_node, &tree->root);
                                __free_pending(pr);
                        }
                }
        }
-       return rc->ndelonly;
+       return rc->ndelayed;
 }