do {
                v = atomic_read(&req->poll_refs);
 
-               /* tw handler should be the owner, and so have some references */
-               if (WARN_ON_ONCE(!(v & IO_POLL_REF_MASK)))
-                       return IOU_POLL_DONE;
-               if (v & IO_POLL_CANCEL_FLAG)
-                       return -ECANCELED;
-               /*
-                * cqe.res contains only events of the first wake up
-                * and all others are be lost. Redo vfs_poll() to get
-                * up to date state.
-                */
-               if ((v & IO_POLL_REF_MASK) != 1)
-                       req->cqe.res = 0;
-               if (v & IO_POLL_RETRY_FLAG) {
-                       req->cqe.res = 0;
+               if (unlikely(v != 1)) {
+                       /* tw should be the owner and so have some refs */
+                       if (WARN_ON_ONCE(!(v & IO_POLL_REF_MASK)))
+                               return IOU_POLL_DONE;
+                       if (v & IO_POLL_CANCEL_FLAG)
+                               return -ECANCELED;
                        /*
-                        * We won't find new events that came in between
-                        * vfs_poll and the ref put unless we clear the flag
-                        * in advance.
+                        * cqe.res contains only events of the first wake up
+                        * and all others are to be lost. Redo vfs_poll() to get
+                        * up to date state.
                         */
-                       atomic_andnot(IO_POLL_RETRY_FLAG, &req->poll_refs);
-                       v &= ~IO_POLL_RETRY_FLAG;
+                       if ((v & IO_POLL_REF_MASK) != 1)
+                               req->cqe.res = 0;
+
+                       if (v & IO_POLL_RETRY_FLAG) {
+                               req->cqe.res = 0;
+                               /*
+                                * We won't find new events that came in between
+                                * vfs_poll and the ref put unless we clear the
+                                * flag in advance.
+                                */
+                               atomic_andnot(IO_POLL_RETRY_FLAG, &req->poll_refs);
+                               v &= ~IO_POLL_RETRY_FLAG;
+                       }
                }
 
                /* the mask was stashed in __io_poll_execute */