Searched refs:nr_segs (Results 1 - 25 of 55) sorted by relevance

123

/linux-master/net/core/
H A Dnet_test.c51 unsigned int nr_segs; member in struct:gso_test_case
60 .nr_segs = 1,
67 .nr_segs = 3,
76 .nr_segs = 3,
84 .nr_segs = 3,
93 .nr_segs = 2,
103 .nr_segs = 3,
111 .nr_segs = 2,
121 .nr_segs = 4,
135 .nr_segs
[all...]
/linux-master/include/linux/
H A Duio.h38 unsigned long nr_segs; member in struct:iov_iter_state
76 unsigned long nr_segs; member in union:iov_iter::__anon2634
101 state->nr_segs = iter->nr_segs;
151 static inline size_t iov_length(const struct iovec *iov, unsigned long nr_segs) argument
156 for (seg = 0; seg < nr_segs; seg++)
270 unsigned long nr_segs, size_t count);
272 unsigned long nr_segs, size_t count);
274 unsigned long nr_segs, size_t count);
337 unsigned long nr_segs, unsigne
[all...]
H A Diov_iter.h63 iter->nr_segs -= p - iter->__iov;
97 iter->nr_segs -= p - iter->kvec;
136 iter->nr_segs -= p - iter->bvec;
H A Dbio.h17 static inline unsigned int bio_max_segs(unsigned int nr_segs) argument
19 return min(nr_segs, BIO_MAX_VECS);
/linux-master/lib/
H A Diov_iter.c163 const struct iovec *iov, unsigned long nr_segs,
172 .nr_segs = nr_segs,
501 for (bvec = i->bvec, end = bvec + i->nr_segs; bvec < end; bvec++) {
507 i->nr_segs -= bvec - i->bvec;
520 for (iov = iter_iov(i), end = iov + i->nr_segs; iov < end; iov++) {
526 i->nr_segs -= iov - iter_iov(i);
571 i->nr_segs++;
583 i->nr_segs++;
600 if (i->nr_segs >
162 iov_iter_init(struct iov_iter *i, unsigned int direction, const struct iovec *iov, unsigned long nr_segs, size_t count) argument
610 iov_iter_kvec(struct iov_iter *i, unsigned int direction, const struct kvec *kvec, unsigned long nr_segs, size_t count) argument
626 iov_iter_bvec(struct iov_iter *i, unsigned int direction, const struct bio_vec *bvec, unsigned long nr_segs, size_t count) argument
1146 copy_compat_iovec_from_user(struct iovec *iov, const struct iovec __user *uvec, u32 nr_segs) argument
1179 copy_iovec_from_user(struct iovec *iov, const struct iovec __user *uiov, unsigned long nr_segs) argument
1211 iovec_from_user(const struct iovec __user *uvec, unsigned long nr_segs, unsigned long fast_segs, struct iovec *fast_iov, bool compat) argument
1270 __import_iovec(int type, const struct iovec __user *uvec, unsigned nr_segs, unsigned fast_segs, struct iovec **iovp, struct iov_iter *i, bool compat) argument
1342 import_iovec(int type, const struct iovec __user *uvec, unsigned nr_segs, unsigned fast_segs, struct iovec **iovp, struct iov_iter *i) argument
[all...]
H A Dkunit_iov_iter.c126 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0);
176 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0);
285 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0);
339 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0);
/linux-master/fs/smb/server/
H A Dtransport_tcp.c111 * @nr_segs: number of segments in base iov
117 unsigned int nr_segs, size_t bytes)
128 nr_segs--;
133 memcpy(new, iov, sizeof(*iov) * nr_segs);
136 return nr_segs;
142 * @nr_segs: number of segments in iov
146 static struct kvec *get_conn_iovec(struct tcp_transport *t, unsigned int nr_segs) argument
150 if (t->iov && nr_segs <= t->nr_iov)
154 new_iov = kmalloc_array(nr_segs, sizeof(*new_iov), GFP_KERNEL);
158 t->nr_iov = nr_segs;
116 kvec_array_init(struct kvec *new, struct kvec *iov, unsigned int nr_segs, size_t bytes) argument
304 ksmbd_tcp_readv(struct tcp_transport *t, struct kvec *iov_orig, unsigned int nr_segs, unsigned int to_read, int max_retries) argument
[all...]
/linux-master/block/
H A Dblk-mq-sched.h11 unsigned int nr_segs, struct request **merged_request);
13 unsigned int nr_segs);
H A Dblk-merge.c341 * @nr_segs: returns the number of segments in the returned bio
352 unsigned int *nr_segs)
360 split = bio_split_discard(bio, lim, nr_segs, bs);
363 split = bio_split_write_zeroes(bio, lim, nr_segs, bs);
366 split = bio_split_rw(bio, lim, nr_segs, bs,
401 unsigned int nr_segs; local
404 return __bio_split_to_limits(bio, lim, &nr_segs);
633 int ll_back_merge_fn(struct request *req, struct bio *bio, unsigned int nr_segs) argument
648 return ll_new_hw_segment(req, bio, nr_segs);
652 unsigned int nr_segs)
350 __bio_split_to_limits(struct bio *bio, const struct queue_limits *lim, unsigned int *nr_segs) argument
651 ll_front_merge_fn(struct request *req, struct bio *bio, unsigned int nr_segs) argument
978 bio_attempt_back_merge(struct request *req, struct bio *bio, unsigned int nr_segs) argument
1007 bio_attempt_front_merge(struct request *req, struct bio *bio, unsigned int nr_segs) argument
1068 blk_attempt_bio_merge(struct request_queue *q, struct request *rq, struct bio *bio, unsigned int nr_segs, bool sched_allow_merge) argument
1115 blk_attempt_plug_merge(struct request_queue *q, struct bio *bio, unsigned int nr_segs) argument
1146 blk_bio_list_merge(struct request_queue *q, struct list_head *list, struct bio *bio, unsigned int nr_segs) argument
1171 blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio, unsigned int nr_segs, struct request **merged_request) argument
[all...]
H A Dblk-map.c26 if (data->nr_segs > UIO_MAXIOV)
29 bmd = kmalloc(struct_size(bmd, iov, data->nr_segs), gfp_mask);
34 memcpy(bmd->iov, iter_iov(data), sizeof(struct iovec) * data->nr_segs);
541 unsigned int nr_segs = 0; local
544 nr_segs++;
547 blk_rq_bio_prep(rq, bio, nr_segs);
549 if (!ll_back_merge_fn(rq, bio, nr_segs))
566 size_t nr_segs = iter->nr_segs; local
575 if (nr_segs > queue_max_segment
[all...]
H A Dblk.h280 struct bio *bio, unsigned int nr_segs);
282 unsigned int nr_segs);
284 struct bio *bio, unsigned int nr_segs);
345 unsigned int *nr_segs);
347 unsigned int nr_segs);
H A Dblk-mq-sched.c338 unsigned int nr_segs)
347 ret = e->type->ops.bio_merge(q, bio, nr_segs);
365 if (blk_bio_list_merge(q, &ctx->rq_lists[type], bio, nr_segs))
337 blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio, unsigned int nr_segs) argument
H A Dblk-crypto-fallback.c162 unsigned int nr_segs = bio_segments(bio_src); local
167 bio = bio_kmalloc(nr_segs, GFP_NOIO);
170 bio_init(bio, bio_src->bi_bdev, bio->bi_inline_vecs, nr_segs,
H A Dblk-zoned.c852 struct bio *bio, unsigned int nr_segs)
873 bio->__bi_nr_segments = nr_segs;
1039 static bool blk_zone_wplug_handle_write(struct bio *bio, unsigned int nr_segs) argument
1104 blk_zone_wplug_add_bio(zwplug, bio, nr_segs);
1114 * @nr_segs: The number of physical segments of @bio
1123 bool blk_zone_plug_bio(struct bio *bio, unsigned int nr_segs) argument
1172 return blk_zone_wplug_handle_write(bio, nr_segs);
851 blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, struct bio *bio, unsigned int nr_segs) argument
/linux-master/arch/powerpc/mm/
H A Ddma-noncoherent.c65 int nr_segs = 1 + ((size - seg_size) + PAGE_SIZE - 1)/PAGE_SIZE; local
84 } while (seg_nr < nr_segs);
/linux-master/drivers/md/bcache/
H A Ddebug.c110 unsigned int nr_segs = bio_segments(bio); local
115 check = bio_kmalloc(nr_segs, GFP_NOIO);
118 bio_init(check, bio->bi_bdev, check->bi_inline_vecs, nr_segs,
/linux-master/drivers/scsi/
H A Dxen-scsifront.c1132 unsigned int sg_grant, nr_segs; local
1136 nr_segs = min_t(unsigned int, sg_grant, SG_ALL);
1137 nr_segs = max_t(unsigned int, nr_segs, VSCSIIF_SG_TABLESIZE);
1138 nr_segs = min_t(unsigned int, nr_segs,
1143 dev_info(&dev->dev, "using up to %d SG entries\n", nr_segs);
1144 else if (info->pause && nr_segs < host->sg_tablesize)
1147 host->sg_tablesize, nr_segs);
1149 host->sg_tablesize = nr_segs;
[all...]
/linux-master/drivers/hwtracing/intel_th/
H A Dmsu.c63 * @nr_segs: number of segments in this window (<= @nr_blocks)
74 unsigned int nr_segs; member in struct:msc_window
330 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
425 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
662 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) {
994 unsigned int nr_segs)
1000 ret = sg_alloc_table(win->sgt, nr_segs, GFP_KERNEL);
1004 for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) {
1014 return nr_segs;
1039 for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs,
993 __msc_buffer_win_alloc(struct msc_window *win, unsigned int nr_segs) argument
[all...]
/linux-master/io_uring/
H A Dnet.c181 kmsg->free_iov_nr = kmsg->msg.msg_iter.nr_segs;
208 int ret, nr_segs; local
211 nr_segs = iomsg->free_iov_nr;
215 nr_segs = 1;
244 nr_segs, &iov, &iomsg->msg.msg_iter, true);
257 int ret, nr_segs; local
260 nr_segs = iomsg->free_iov_nr;
264 nr_segs = 1;
301 ret = __import_iovec(ddir, msg->msg_iov, msg->msg_iovlen, nr_segs,
317 iomsg->msg.msg_iter.nr_segs
[all...]
H A Drw.c87 int nr_segs, ret; local
106 nr_segs = io->free_iov_nr;
110 nr_segs = 1;
112 ret = __import_iovec(ddir, buf, sqe_len, nr_segs, &iov, &io->iter,
118 io->free_iov_nr = io->iter.nr_segs;
/linux-master/fs/fuse/
H A Ddev.c651 unsigned long nr_segs; member in struct:fuse_copy_state
707 BUG_ON(!cs->nr_segs);
713 cs->nr_segs--;
715 if (cs->nr_segs >= cs->pipe->max_usage)
731 cs->nr_segs++;
807 BUG_ON(!cs->nr_segs);
811 cs->nr_segs--;
900 if (cs->nr_segs >= cs->pipe->max_usage)
918 cs->nr_segs++;
1395 if (pipe_occupancy(pipe->head, pipe->tail) + cs.nr_segs > pip
[all...]
/linux-master/drivers/nvme/target/
H A Dio-cmd-file.c77 unsigned long nr_segs, size_t count, int ki_flags)
94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count);
76 nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, unsigned long nr_segs, size_t count, int ki_flags) argument
/linux-master/fs/netfs/
H A Diterator.c113 unsigned int nbv = iter->nr_segs, ix = 0, nsegs = 0;
/linux-master/drivers/block/xen-blkback/
H A Dcommon.h347 int nr_segs; member in struct:pending_req
H A Dblkback.c710 invcount = xen_blkbk_unmap_prepare(ring, pages, req->nr_segs,
907 pending_req->nr_segs,
923 nseg = pending_req->nr_segs;
1361 pending_req->nr_segs = nseg;
1475 pending_req->nr_segs);

Completed in 615 milliseconds

123