u64 ino = btrfs_ino(inode);
        struct btrfs_delayed_node *node;
 
-       node = ACCESS_ONCE(btrfs_inode->delayed_node);
+       node = READ_ONCE(btrfs_inode->delayed_node);
        if (node) {
                atomic_inc(&node->refs);
                return node;
 {
        struct btrfs_delayed_node *delayed_node;
 
-       delayed_node = ACCESS_ONCE(BTRFS_I(inode)->delayed_node);
+       delayed_node = READ_ONCE(BTRFS_I(inode)->delayed_node);
        if (!delayed_node)
                return;
 
 
                           function, line, errstr);
                return;
        }
-       ACCESS_ONCE(trans->transaction->aborted) = errno;
+       WRITE_ONCE(trans->transaction->aborted, errno);
        /* Wake up anybody who may be waiting on this transaction */
        wake_up(&fs_info->transaction_wait);
        wake_up(&fs_info->transaction_blocked_wait);
 
 
        if (lock && !atomic_read(&info->open_ioctl_trans) &&
            should_end_transaction(trans) &&
-           ACCESS_ONCE(cur_trans->state) == TRANS_STATE_RUNNING) {
+           READ_ONCE(cur_trans->state) == TRANS_STATE_RUNNING) {
                spin_lock(&info->trans_lock);
                if (cur_trans->state == TRANS_STATE_RUNNING)
                        cur_trans->state = TRANS_STATE_BLOCKED;
                spin_unlock(&info->trans_lock);
        }
 
-       if (lock && ACCESS_ONCE(cur_trans->state) == TRANS_STATE_BLOCKED) {
+       if (lock && READ_ONCE(cur_trans->state) == TRANS_STATE_BLOCKED) {
                if (throttle)
                        return btrfs_commit_transaction(trans);
                else
        int ret;
 
        /* Stop the commit early if ->aborted is set */
-       if (unlikely(ACCESS_ONCE(cur_trans->aborted))) {
+       if (unlikely(READ_ONCE(cur_trans->aborted))) {
                ret = cur_trans->aborted;
                btrfs_end_transaction(trans);
                return ret;
                   atomic_read(&cur_trans->num_writers) == 1);
 
        /* ->aborted might be set after the previous check, so check it */
-       if (unlikely(ACCESS_ONCE(cur_trans->aborted))) {
+       if (unlikely(READ_ONCE(cur_trans->aborted))) {
                ret = cur_trans->aborted;
                goto scrub_continue;
        }
         * The tasks which save the space cache and inode cache may also
         * update ->aborted, check it.
         */
-       if (unlikely(ACCESS_ONCE(cur_trans->aborted))) {
+       if (unlikely(READ_ONCE(cur_trans->aborted))) {
                ret = cur_trans->aborted;
                mutex_unlock(&fs_info->tree_log_mutex);
                mutex_unlock(&fs_info->reloc_mutex);
 
 static inline void btrfs_set_log_full_commit(struct btrfs_fs_info *fs_info,
                                             struct btrfs_trans_handle *trans)
 {
-       ACCESS_ONCE(fs_info->last_trans_log_full_commit) = trans->transid;
+       WRITE_ONCE(fs_info->last_trans_log_full_commit, trans->transid);
 }
 
 static inline int btrfs_need_log_full_commit(struct btrfs_fs_info *fs_info,
                                             struct btrfs_trans_handle *trans)
 {
-       return ACCESS_ONCE(fs_info->last_trans_log_full_commit) ==
+       return READ_ONCE(fs_info->last_trans_log_full_commit) ==
                trans->transid;
 }