if (req->flags & REQ_F_IMPORT_BUFFER) {
unsigned uvec_segs = kmsg->msg.msg_iter.nr_segs;
- unsigned iovec_off = kmsg->vec.nr - uvec_segs;
int ret;
ret = io_import_reg_vec(ITER_SOURCE, &kmsg->msg.msg_iter, req,
- &kmsg->vec, uvec_segs, iovec_off,
- issue_flags);
+ &kmsg->vec, uvec_segs, issue_flags);
if (unlikely(ret))
return ret;
kmsg->msg.sg_from_iter = io_sg_from_iter;
int io_import_reg_vec(int ddir, struct iov_iter *iter,
struct io_kiocb *req, struct iou_vec *vec,
- unsigned nr_iovs, unsigned iovec_off,
- unsigned issue_flags)
+ unsigned nr_iovs, unsigned issue_flags)
{
struct io_rsrc_node *node;
struct io_mapped_ubuf *imu;
+ unsigned iovec_off;
struct iovec *iov;
unsigned nr_segs;
if (!(imu->dir & (1 << ddir)))
return -EFAULT;
+ iovec_off = vec->nr - nr_iovs;
iov = vec->iovec + iovec_off;
nr_segs = io_estimate_bvec_size(iov, nr_iovs, imu);
nr_segs += nr_iovs;
}
- if (WARN_ON_ONCE(iovec_off + nr_iovs != vec->nr) ||
- nr_segs > vec->nr) {
+ if (nr_segs > vec->nr) {
struct iou_vec tmp_vec = {};
int ret;
unsigned issue_flags);
int io_import_reg_vec(int ddir, struct iov_iter *iter,
struct io_kiocb *req, struct iou_vec *vec,
- unsigned nr_iovs, unsigned iovec_off,
- unsigned issue_flags);
+ unsigned nr_iovs, unsigned issue_flags);
int io_prep_reg_iovec(struct io_kiocb *req, struct iou_vec *iv,
const struct iovec __user *uvec, size_t uvec_segs);
{
struct io_rw *rw = io_kiocb_to_cmd(req, struct io_rw);
unsigned uvec_segs = rw->len;
- unsigned iovec_off = io->vec.nr - uvec_segs;
int ret;
ret = io_import_reg_vec(ddir, &io->iter, req, &io->vec,
- uvec_segs, iovec_off, issue_flags);
+ uvec_segs, issue_flags);
if (unlikely(ret))
return ret;
iov_iter_save_state(&io->iter, &io->iter_state);