/linux-master/net/core/ |
H A D | net_test.c | 51 unsigned int nr_segs; member in struct:gso_test_case 60 .nr_segs = 1, 67 .nr_segs = 3, 76 .nr_segs = 3, 84 .nr_segs = 3, 93 .nr_segs = 2, 103 .nr_segs = 3, 111 .nr_segs = 2, 121 .nr_segs = 4, 135 .nr_segs [all...] |
/linux-master/include/linux/ |
H A D | uio.h | 38 unsigned long nr_segs; member in struct:iov_iter_state 76 unsigned long nr_segs; member in union:iov_iter::__anon2634 101 state->nr_segs = iter->nr_segs; 151 static inline size_t iov_length(const struct iovec *iov, unsigned long nr_segs) argument 156 for (seg = 0; seg < nr_segs; seg++) 270 unsigned long nr_segs, size_t count); 272 unsigned long nr_segs, size_t count); 274 unsigned long nr_segs, size_t count); 337 unsigned long nr_segs, unsigne [all...] |
H A D | iov_iter.h | 63 iter->nr_segs -= p - iter->__iov; 97 iter->nr_segs -= p - iter->kvec; 136 iter->nr_segs -= p - iter->bvec;
|
H A D | bio.h | 17 static inline unsigned int bio_max_segs(unsigned int nr_segs) argument 19 return min(nr_segs, BIO_MAX_VECS);
|
/linux-master/lib/ |
H A D | iov_iter.c | 163 const struct iovec *iov, unsigned long nr_segs, 172 .nr_segs = nr_segs, 501 for (bvec = i->bvec, end = bvec + i->nr_segs; bvec < end; bvec++) { 507 i->nr_segs -= bvec - i->bvec; 520 for (iov = iter_iov(i), end = iov + i->nr_segs; iov < end; iov++) { 526 i->nr_segs -= iov - iter_iov(i); 571 i->nr_segs++; 583 i->nr_segs++; 600 if (i->nr_segs > 162 iov_iter_init(struct iov_iter *i, unsigned int direction, const struct iovec *iov, unsigned long nr_segs, size_t count) argument 610 iov_iter_kvec(struct iov_iter *i, unsigned int direction, const struct kvec *kvec, unsigned long nr_segs, size_t count) argument 626 iov_iter_bvec(struct iov_iter *i, unsigned int direction, const struct bio_vec *bvec, unsigned long nr_segs, size_t count) argument 1146 copy_compat_iovec_from_user(struct iovec *iov, const struct iovec __user *uvec, u32 nr_segs) argument 1179 copy_iovec_from_user(struct iovec *iov, const struct iovec __user *uiov, unsigned long nr_segs) argument 1211 iovec_from_user(const struct iovec __user *uvec, unsigned long nr_segs, unsigned long fast_segs, struct iovec *fast_iov, bool compat) argument 1270 __import_iovec(int type, const struct iovec __user *uvec, unsigned nr_segs, unsigned fast_segs, struct iovec **iovp, struct iov_iter *i, bool compat) argument 1342 import_iovec(int type, const struct iovec __user *uvec, unsigned nr_segs, unsigned fast_segs, struct iovec **iovp, struct iov_iter *i) argument [all...] |
H A D | kunit_iov_iter.c | 126 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0); 176 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0); 285 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0); 339 KUNIT_EXPECT_EQ(test, iter.nr_segs, 0);
|
/linux-master/fs/smb/server/ |
H A D | transport_tcp.c | 111 * @nr_segs: number of segments in base iov 117 unsigned int nr_segs, size_t bytes) 128 nr_segs--; 133 memcpy(new, iov, sizeof(*iov) * nr_segs); 136 return nr_segs; 142 * @nr_segs: number of segments in iov 146 static struct kvec *get_conn_iovec(struct tcp_transport *t, unsigned int nr_segs) argument 150 if (t->iov && nr_segs <= t->nr_iov) 154 new_iov = kmalloc_array(nr_segs, sizeof(*new_iov), GFP_KERNEL); 158 t->nr_iov = nr_segs; 116 kvec_array_init(struct kvec *new, struct kvec *iov, unsigned int nr_segs, size_t bytes) argument 304 ksmbd_tcp_readv(struct tcp_transport *t, struct kvec *iov_orig, unsigned int nr_segs, unsigned int to_read, int max_retries) argument [all...] |
/linux-master/block/ |
H A D | blk-mq-sched.h | 11 unsigned int nr_segs, struct request **merged_request); 13 unsigned int nr_segs);
|
H A D | blk-merge.c | 341 * @nr_segs: returns the number of segments in the returned bio 352 unsigned int *nr_segs) 360 split = bio_split_discard(bio, lim, nr_segs, bs); 363 split = bio_split_write_zeroes(bio, lim, nr_segs, bs); 366 split = bio_split_rw(bio, lim, nr_segs, bs, 401 unsigned int nr_segs; local 404 return __bio_split_to_limits(bio, lim, &nr_segs); 633 int ll_back_merge_fn(struct request *req, struct bio *bio, unsigned int nr_segs) argument 648 return ll_new_hw_segment(req, bio, nr_segs); 652 unsigned int nr_segs) 350 __bio_split_to_limits(struct bio *bio, const struct queue_limits *lim, unsigned int *nr_segs) argument 651 ll_front_merge_fn(struct request *req, struct bio *bio, unsigned int nr_segs) argument 978 bio_attempt_back_merge(struct request *req, struct bio *bio, unsigned int nr_segs) argument 1007 bio_attempt_front_merge(struct request *req, struct bio *bio, unsigned int nr_segs) argument 1068 blk_attempt_bio_merge(struct request_queue *q, struct request *rq, struct bio *bio, unsigned int nr_segs, bool sched_allow_merge) argument 1115 blk_attempt_plug_merge(struct request_queue *q, struct bio *bio, unsigned int nr_segs) argument 1146 blk_bio_list_merge(struct request_queue *q, struct list_head *list, struct bio *bio, unsigned int nr_segs) argument 1171 blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio, unsigned int nr_segs, struct request **merged_request) argument [all...] |
H A D | blk-map.c | 26 if (data->nr_segs > UIO_MAXIOV) 29 bmd = kmalloc(struct_size(bmd, iov, data->nr_segs), gfp_mask); 34 memcpy(bmd->iov, iter_iov(data), sizeof(struct iovec) * data->nr_segs); 541 unsigned int nr_segs = 0; local 544 nr_segs++; 547 blk_rq_bio_prep(rq, bio, nr_segs); 549 if (!ll_back_merge_fn(rq, bio, nr_segs)) 566 size_t nr_segs = iter->nr_segs; local 575 if (nr_segs > queue_max_segment [all...] |
H A D | blk.h | 280 struct bio *bio, unsigned int nr_segs); 282 unsigned int nr_segs); 284 struct bio *bio, unsigned int nr_segs); 345 unsigned int *nr_segs); 347 unsigned int nr_segs);
|
H A D | blk-mq-sched.c | 338 unsigned int nr_segs) 347 ret = e->type->ops.bio_merge(q, bio, nr_segs); 365 if (blk_bio_list_merge(q, &ctx->rq_lists[type], bio, nr_segs)) 337 blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio, unsigned int nr_segs) argument
|
H A D | blk-crypto-fallback.c | 162 unsigned int nr_segs = bio_segments(bio_src); local 167 bio = bio_kmalloc(nr_segs, GFP_NOIO); 170 bio_init(bio, bio_src->bi_bdev, bio->bi_inline_vecs, nr_segs,
|
H A D | blk-zoned.c | 852 struct bio *bio, unsigned int nr_segs) 873 bio->__bi_nr_segments = nr_segs; 1039 static bool blk_zone_wplug_handle_write(struct bio *bio, unsigned int nr_segs) argument 1104 blk_zone_wplug_add_bio(zwplug, bio, nr_segs); 1114 * @nr_segs: The number of physical segments of @bio 1123 bool blk_zone_plug_bio(struct bio *bio, unsigned int nr_segs) argument 1172 return blk_zone_wplug_handle_write(bio, nr_segs); 851 blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, struct bio *bio, unsigned int nr_segs) argument
|
/linux-master/arch/powerpc/mm/ |
H A D | dma-noncoherent.c | 65 int nr_segs = 1 + ((size - seg_size) + PAGE_SIZE - 1)/PAGE_SIZE; local 84 } while (seg_nr < nr_segs);
|
/linux-master/drivers/md/bcache/ |
H A D | debug.c | 110 unsigned int nr_segs = bio_segments(bio); local 115 check = bio_kmalloc(nr_segs, GFP_NOIO); 118 bio_init(check, bio->bi_bdev, check->bi_inline_vecs, nr_segs,
|
/linux-master/drivers/scsi/ |
H A D | xen-scsifront.c | 1132 unsigned int sg_grant, nr_segs; local 1136 nr_segs = min_t(unsigned int, sg_grant, SG_ALL); 1137 nr_segs = max_t(unsigned int, nr_segs, VSCSIIF_SG_TABLESIZE); 1138 nr_segs = min_t(unsigned int, nr_segs, 1143 dev_info(&dev->dev, "using up to %d SG entries\n", nr_segs); 1144 else if (info->pause && nr_segs < host->sg_tablesize) 1147 host->sg_tablesize, nr_segs); 1149 host->sg_tablesize = nr_segs; [all...] |
/linux-master/drivers/hwtracing/intel_th/ |
H A D | msu.c | 63 * @nr_segs: number of segments in this window (<= @nr_blocks) 74 unsigned int nr_segs; member in struct:msc_window 330 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { 425 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { 662 for_each_sg(win->sgt->sgl, sg, win->nr_segs, blk) { 994 unsigned int nr_segs) 1000 ret = sg_alloc_table(win->sgt, nr_segs, GFP_KERNEL); 1004 for_each_sg(win->sgt->sgl, sg_ptr, nr_segs, i) { 1014 return nr_segs; 1039 for_each_sg(win->sgt->sgl, sg_ptr, win->nr_segs, 993 __msc_buffer_win_alloc(struct msc_window *win, unsigned int nr_segs) argument [all...] |
/linux-master/io_uring/ |
H A D | net.c | 181 kmsg->free_iov_nr = kmsg->msg.msg_iter.nr_segs; 208 int ret, nr_segs; local 211 nr_segs = iomsg->free_iov_nr; 215 nr_segs = 1; 244 nr_segs, &iov, &iomsg->msg.msg_iter, true); 257 int ret, nr_segs; local 260 nr_segs = iomsg->free_iov_nr; 264 nr_segs = 1; 301 ret = __import_iovec(ddir, msg->msg_iov, msg->msg_iovlen, nr_segs, 317 iomsg->msg.msg_iter.nr_segs [all...] |
H A D | rw.c | 87 int nr_segs, ret; local 106 nr_segs = io->free_iov_nr; 110 nr_segs = 1; 112 ret = __import_iovec(ddir, buf, sqe_len, nr_segs, &iov, &io->iter, 118 io->free_iov_nr = io->iter.nr_segs;
|
/linux-master/fs/fuse/ |
H A D | dev.c | 651 unsigned long nr_segs; member in struct:fuse_copy_state 707 BUG_ON(!cs->nr_segs); 713 cs->nr_segs--; 715 if (cs->nr_segs >= cs->pipe->max_usage) 731 cs->nr_segs++; 807 BUG_ON(!cs->nr_segs); 811 cs->nr_segs--; 900 if (cs->nr_segs >= cs->pipe->max_usage) 918 cs->nr_segs++; 1395 if (pipe_occupancy(pipe->head, pipe->tail) + cs.nr_segs > pip [all...] |
/linux-master/drivers/nvme/target/ |
H A D | io-cmd-file.c | 77 unsigned long nr_segs, size_t count, int ki_flags) 94 iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count); 76 nvmet_file_submit_bvec(struct nvmet_req *req, loff_t pos, unsigned long nr_segs, size_t count, int ki_flags) argument
|
/linux-master/fs/netfs/ |
H A D | iterator.c | 113 unsigned int nbv = iter->nr_segs, ix = 0, nsegs = 0;
|
/linux-master/drivers/block/xen-blkback/ |
H A D | common.h | 347 int nr_segs; member in struct:pending_req
|
H A D | blkback.c | 710 invcount = xen_blkbk_unmap_prepare(ring, pages, req->nr_segs, 907 pending_req->nr_segs, 923 nseg = pending_req->nr_segs; 1361 pending_req->nr_segs = nseg; 1475 pending_req->nr_segs);
|