Orange Pi5 kernel

Deprecated Linux kernel 5.10.110 for OrangePi 5/5B/5+ boards

3 Commits   0 Branches   0 Tags
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    1) // SPDX-License-Identifier: GPL-2.0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    3)  * Copyright (C) 2001 Jens Axboe <axboe@kernel.dk>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    4)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    5) #include <linux/mm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    6) #include <linux/swap.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    7) #include <linux/bio.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    8) #include <linux/blkdev.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    9) #include <linux/uio.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   10) #include <linux/iocontext.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   11) #include <linux/slab.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   12) #include <linux/init.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   13) #include <linux/kernel.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   14) #include <linux/export.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   15) #include <linux/mempool.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   16) #include <linux/workqueue.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   17) #include <linux/cgroup.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   18) #include <linux/blk-cgroup.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   19) #include <linux/highmem.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   20) #include <linux/sched/sysctl.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   21) #include <linux/blk-crypto.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   22) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   23) #include <trace/events/block.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   24) #include "blk.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   25) #include "blk-rq-qos.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   26) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   27) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   28)  * Test patch to inline a certain number of bi_io_vec's inside the bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   29)  * itself, to shrink a bio data allocation from two mempool calls to one
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   30)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   31) #define BIO_INLINE_VECS		4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   32) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   33) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   34)  * if you change this list, also change bvec_alloc or things will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   35)  * break badly! cannot be bigger than what you can fit into an
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   36)  * unsigned short
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   37)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   38) #define BV(x, n) { .nr_vecs = x, .name = "biovec-"#n }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   39) static struct biovec_slab bvec_slabs[BVEC_POOL_NR] __read_mostly = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   40) 	BV(1, 1), BV(4, 4), BV(16, 16), BV(64, 64), BV(128, 128), BV(BIO_MAX_PAGES, max),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   41) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   42) #undef BV
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   43) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   44) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   45)  * fs_bio_set is the bio_set containing bio and iovec memory pools used by
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   46)  * IO code that does not need private memory pools.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   47)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   48) struct bio_set fs_bio_set;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   49) EXPORT_SYMBOL(fs_bio_set);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   50) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   51) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   52)  * Our slab pool management
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   53)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   54) struct bio_slab {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   55) 	struct kmem_cache *slab;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   56) 	unsigned int slab_ref;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   57) 	unsigned int slab_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   58) 	char name[8];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   59) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   60) static DEFINE_MUTEX(bio_slab_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   61) static struct bio_slab *bio_slabs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   62) static unsigned int bio_slab_nr, bio_slab_max;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   63) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   64) static struct kmem_cache *bio_find_or_create_slab(unsigned int extra_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   65) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   66) 	unsigned int sz = sizeof(struct bio) + extra_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   67) 	struct kmem_cache *slab = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   68) 	struct bio_slab *bslab, *new_bio_slabs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   69) 	unsigned int new_bio_slab_max;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   70) 	unsigned int i, entry = -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   71) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   72) 	mutex_lock(&bio_slab_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   73) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   74) 	i = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   75) 	while (i < bio_slab_nr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   76) 		bslab = &bio_slabs[i];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   77) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   78) 		if (!bslab->slab && entry == -1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   79) 			entry = i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   80) 		else if (bslab->slab_size == sz) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   81) 			slab = bslab->slab;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   82) 			bslab->slab_ref++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   83) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   84) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   85) 		i++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   86) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   87) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   88) 	if (slab)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   89) 		goto out_unlock;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   90) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   91) 	if (bio_slab_nr == bio_slab_max && entry == -1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   92) 		new_bio_slab_max = bio_slab_max << 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   93) 		new_bio_slabs = krealloc(bio_slabs,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   94) 					 new_bio_slab_max * sizeof(struct bio_slab),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   95) 					 GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   96) 		if (!new_bio_slabs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   97) 			goto out_unlock;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   98) 		bio_slab_max = new_bio_slab_max;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   99) 		bio_slabs = new_bio_slabs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  100) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  101) 	if (entry == -1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  102) 		entry = bio_slab_nr++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  103) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  104) 	bslab = &bio_slabs[entry];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  105) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  106) 	snprintf(bslab->name, sizeof(bslab->name), "bio-%d", entry);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  107) 	slab = kmem_cache_create(bslab->name, sz, ARCH_KMALLOC_MINALIGN,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  108) 				 SLAB_HWCACHE_ALIGN, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  109) 	if (!slab)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  110) 		goto out_unlock;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  111) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  112) 	bslab->slab = slab;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  113) 	bslab->slab_ref = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  114) 	bslab->slab_size = sz;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  115) out_unlock:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  116) 	mutex_unlock(&bio_slab_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  117) 	return slab;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  118) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  119) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  120) static void bio_put_slab(struct bio_set *bs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  121) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  122) 	struct bio_slab *bslab = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  123) 	unsigned int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  124) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  125) 	mutex_lock(&bio_slab_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  126) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  127) 	for (i = 0; i < bio_slab_nr; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  128) 		if (bs->bio_slab == bio_slabs[i].slab) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  129) 			bslab = &bio_slabs[i];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  130) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  131) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  132) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  133) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  134) 	if (WARN(!bslab, KERN_ERR "bio: unable to find slab!\n"))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  135) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  136) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  137) 	WARN_ON(!bslab->slab_ref);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  138) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  139) 	if (--bslab->slab_ref)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  140) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  141) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  142) 	kmem_cache_destroy(bslab->slab);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  143) 	bslab->slab = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  144) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  145) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  146) 	mutex_unlock(&bio_slab_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  147) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  148) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  149) unsigned int bvec_nr_vecs(unsigned short idx)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  150) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  151) 	return bvec_slabs[--idx].nr_vecs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  152) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  153) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  154) void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned int idx)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  155) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  156) 	if (!idx)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  157) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  158) 	idx--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  159) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  160) 	BIO_BUG_ON(idx >= BVEC_POOL_NR);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  161) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  162) 	if (idx == BVEC_POOL_MAX) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  163) 		mempool_free(bv, pool);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  164) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  165) 		struct biovec_slab *bvs = bvec_slabs + idx;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  166) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  167) 		kmem_cache_free(bvs->slab, bv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  168) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  169) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  170) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  171) struct bio_vec *bvec_alloc(gfp_t gfp_mask, int nr, unsigned long *idx,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  172) 			   mempool_t *pool)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  173) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  174) 	struct bio_vec *bvl;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  175) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  176) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  177) 	 * see comment near bvec_array define!
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  178) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  179) 	switch (nr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  180) 	case 1:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  181) 		*idx = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  182) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  183) 	case 2 ... 4:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  184) 		*idx = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  185) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  186) 	case 5 ... 16:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  187) 		*idx = 2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  188) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  189) 	case 17 ... 64:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  190) 		*idx = 3;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  191) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  192) 	case 65 ... 128:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  193) 		*idx = 4;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  194) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  195) 	case 129 ... BIO_MAX_PAGES:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  196) 		*idx = 5;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  197) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  198) 	default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  199) 		return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  200) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  201) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  202) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  203) 	 * idx now points to the pool we want to allocate from. only the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  204) 	 * 1-vec entry pool is mempool backed.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  205) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  206) 	if (*idx == BVEC_POOL_MAX) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  207) fallback:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  208) 		bvl = mempool_alloc(pool, gfp_mask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  209) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  210) 		struct biovec_slab *bvs = bvec_slabs + *idx;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  211) 		gfp_t __gfp_mask = gfp_mask & ~(__GFP_DIRECT_RECLAIM | __GFP_IO);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  212) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  213) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  214) 		 * Make this allocation restricted and don't dump info on
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  215) 		 * allocation failures, since we'll fallback to the mempool
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  216) 		 * in case of failure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  217) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  218) 		__gfp_mask |= __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  219) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  220) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  221) 		 * Try a slab allocation. If this fails and __GFP_DIRECT_RECLAIM
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  222) 		 * is set, retry with the 1-entry mempool
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  223) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  224) 		bvl = kmem_cache_alloc(bvs->slab, __gfp_mask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  225) 		if (unlikely(!bvl && (gfp_mask & __GFP_DIRECT_RECLAIM))) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  226) 			*idx = BVEC_POOL_MAX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  227) 			goto fallback;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  228) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  229) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  230) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  231) 	(*idx)++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  232) 	return bvl;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  233) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  234) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  235) void bio_uninit(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  236) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  237) #ifdef CONFIG_BLK_CGROUP
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  238) 	if (bio->bi_blkg) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  239) 		blkg_put(bio->bi_blkg);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  240) 		bio->bi_blkg = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  241) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  242) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  243) 	if (bio_integrity(bio))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  244) 		bio_integrity_free(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  245) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  246) 	bio_crypt_free_ctx(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  247) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  248) EXPORT_SYMBOL(bio_uninit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  249) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  250) static void bio_free(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  251) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  252) 	struct bio_set *bs = bio->bi_pool;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  253) 	void *p;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  254) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  255) 	bio_uninit(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  256) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  257) 	if (bs) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  258) 		bvec_free(&bs->bvec_pool, bio->bi_io_vec, BVEC_POOL_IDX(bio));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  259) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  260) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  261) 		 * If we have front padding, adjust the bio pointer before freeing
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  262) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  263) 		p = bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  264) 		p -= bs->front_pad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  265) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  266) 		mempool_free(p, &bs->bio_pool);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  267) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  268) 		/* Bio was allocated by bio_kmalloc() */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  269) 		kfree(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  270) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  271) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  272) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  273) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  274)  * Users of this function have their own bio allocation. Subsequently,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  275)  * they must remember to pair any call to bio_init() with bio_uninit()
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  276)  * when IO has completed, or when the bio is released.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  277)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  278) void bio_init(struct bio *bio, struct bio_vec *table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  279) 	      unsigned short max_vecs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  280) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  281) 	memset(bio, 0, sizeof(*bio));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  282) 	atomic_set(&bio->__bi_remaining, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  283) 	atomic_set(&bio->__bi_cnt, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  284) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  285) 	bio->bi_io_vec = table;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  286) 	bio->bi_max_vecs = max_vecs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  287) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  288) EXPORT_SYMBOL(bio_init);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  289) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  290) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  291)  * bio_reset - reinitialize a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  292)  * @bio:	bio to reset
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  293)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  294)  * Description:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  295)  *   After calling bio_reset(), @bio will be in the same state as a freshly
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  296)  *   allocated bio returned bio bio_alloc_bioset() - the only fields that are
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  297)  *   preserved are the ones that are initialized by bio_alloc_bioset(). See
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  298)  *   comment in struct bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  299)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  300) void bio_reset(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  301) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  302) 	unsigned long flags = bio->bi_flags & (~0UL << BIO_RESET_BITS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  303) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  304) 	bio_uninit(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  305) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  306) 	memset(bio, 0, BIO_RESET_BYTES);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  307) 	bio->bi_flags = flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  308) 	atomic_set(&bio->__bi_remaining, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  309) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  310) EXPORT_SYMBOL(bio_reset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  311) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  312) static struct bio *__bio_chain_endio(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  313) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  314) 	struct bio *parent = bio->bi_private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  315) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  316) 	if (bio->bi_status && !parent->bi_status)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  317) 		parent->bi_status = bio->bi_status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  318) 	bio_put(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  319) 	return parent;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  320) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  321) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  322) static void bio_chain_endio(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  323) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  324) 	bio_endio(__bio_chain_endio(bio));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  325) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  326) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  327) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  328)  * bio_chain - chain bio completions
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  329)  * @bio: the target bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  330)  * @parent: the parent bio of @bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  331)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  332)  * The caller won't have a bi_end_io called when @bio completes - instead,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  333)  * @parent's bi_end_io won't be called until both @parent and @bio have
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  334)  * completed; the chained bio will also be freed when it completes.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  335)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  336)  * The caller must not set bi_private or bi_end_io in @bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  337)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  338) void bio_chain(struct bio *bio, struct bio *parent)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  339) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  340) 	BUG_ON(bio->bi_private || bio->bi_end_io);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  341) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  342) 	bio->bi_private = parent;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  343) 	bio->bi_end_io	= bio_chain_endio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  344) 	bio_inc_remaining(parent);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  345) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  346) EXPORT_SYMBOL(bio_chain);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  347) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  348) static void bio_alloc_rescue(struct work_struct *work)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  349) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  350) 	struct bio_set *bs = container_of(work, struct bio_set, rescue_work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  351) 	struct bio *bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  352) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  353) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  354) 		spin_lock(&bs->rescue_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  355) 		bio = bio_list_pop(&bs->rescue_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  356) 		spin_unlock(&bs->rescue_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  357) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  358) 		if (!bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  359) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  360) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  361) 		submit_bio_noacct(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  362) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  363) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  364) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  365) static void punt_bios_to_rescuer(struct bio_set *bs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  366) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  367) 	struct bio_list punt, nopunt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  368) 	struct bio *bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  369) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  370) 	if (WARN_ON_ONCE(!bs->rescue_workqueue))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  371) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  372) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  373) 	 * In order to guarantee forward progress we must punt only bios that
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  374) 	 * were allocated from this bio_set; otherwise, if there was a bio on
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  375) 	 * there for a stacking driver higher up in the stack, processing it
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  376) 	 * could require allocating bios from this bio_set, and doing that from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  377) 	 * our own rescuer would be bad.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  378) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  379) 	 * Since bio lists are singly linked, pop them all instead of trying to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  380) 	 * remove from the middle of the list:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  381) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  382) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  383) 	bio_list_init(&punt);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  384) 	bio_list_init(&nopunt);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  385) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  386) 	while ((bio = bio_list_pop(&current->bio_list[0])))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  387) 		bio_list_add(bio->bi_pool == bs ? &punt : &nopunt, bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  388) 	current->bio_list[0] = nopunt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  389) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  390) 	bio_list_init(&nopunt);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  391) 	while ((bio = bio_list_pop(&current->bio_list[1])))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  392) 		bio_list_add(bio->bi_pool == bs ? &punt : &nopunt, bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  393) 	current->bio_list[1] = nopunt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  394) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  395) 	spin_lock(&bs->rescue_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  396) 	bio_list_merge(&bs->rescue_list, &punt);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  397) 	spin_unlock(&bs->rescue_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  398) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  399) 	queue_work(bs->rescue_workqueue, &bs->rescue_work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  400) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  401) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  402) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  403)  * bio_alloc_bioset - allocate a bio for I/O
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  404)  * @gfp_mask:   the GFP_* mask given to the slab allocator
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  405)  * @nr_iovecs:	number of iovecs to pre-allocate
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  406)  * @bs:		the bio_set to allocate from.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  407)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  408)  * Description:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  409)  *   If @bs is NULL, uses kmalloc() to allocate the bio; else the allocation is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  410)  *   backed by the @bs's mempool.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  411)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  412)  *   When @bs is not NULL, if %__GFP_DIRECT_RECLAIM is set then bio_alloc will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  413)  *   always be able to allocate a bio. This is due to the mempool guarantees.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  414)  *   To make this work, callers must never allocate more than 1 bio at a time
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  415)  *   from this pool. Callers that need to allocate more than 1 bio must always
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  416)  *   submit the previously allocated bio for IO before attempting to allocate
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  417)  *   a new one. Failure to do so can cause deadlocks under memory pressure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  418)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  419)  *   Note that when running under submit_bio_noacct() (i.e. any block
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  420)  *   driver), bios are not submitted until after you return - see the code in
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  421)  *   submit_bio_noacct() that converts recursion into iteration, to prevent
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  422)  *   stack overflows.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  423)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  424)  *   This would normally mean allocating multiple bios under
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  425)  *   submit_bio_noacct() would be susceptible to deadlocks, but we have
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  426)  *   deadlock avoidance code that resubmits any blocked bios from a rescuer
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  427)  *   thread.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  428)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  429)  *   However, we do not guarantee forward progress for allocations from other
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  430)  *   mempools. Doing multiple allocations from the same mempool under
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  431)  *   submit_bio_noacct() should be avoided - instead, use bio_set's front_pad
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  432)  *   for per bio allocations.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  433)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  434)  *   RETURNS:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  435)  *   Pointer to new bio on success, NULL on failure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  436)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  437) struct bio *bio_alloc_bioset(gfp_t gfp_mask, unsigned int nr_iovecs,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  438) 			     struct bio_set *bs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  439) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  440) 	gfp_t saved_gfp = gfp_mask;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  441) 	unsigned front_pad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  442) 	unsigned inline_vecs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  443) 	struct bio_vec *bvl = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  444) 	struct bio *bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  445) 	void *p;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  446) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  447) 	if (!bs) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  448) 		if (nr_iovecs > UIO_MAXIOV)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  449) 			return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  450) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  451) 		p = kmalloc(struct_size(bio, bi_inline_vecs, nr_iovecs), gfp_mask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  452) 		front_pad = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  453) 		inline_vecs = nr_iovecs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  454) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  455) 		/* should not use nobvec bioset for nr_iovecs > 0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  456) 		if (WARN_ON_ONCE(!mempool_initialized(&bs->bvec_pool) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  457) 				 nr_iovecs > 0))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  458) 			return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  459) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  460) 		 * submit_bio_noacct() converts recursion to iteration; this
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  461) 		 * means if we're running beneath it, any bios we allocate and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  462) 		 * submit will not be submitted (and thus freed) until after we
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  463) 		 * return.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  464) 		 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  465) 		 * This exposes us to a potential deadlock if we allocate
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  466) 		 * multiple bios from the same bio_set() while running
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  467) 		 * underneath submit_bio_noacct(). If we were to allocate
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  468) 		 * multiple bios (say a stacking block driver that was splitting
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  469) 		 * bios), we would deadlock if we exhausted the mempool's
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  470) 		 * reserve.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  471) 		 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  472) 		 * We solve this, and guarantee forward progress, with a rescuer
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  473) 		 * workqueue per bio_set. If we go to allocate and there are
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  474) 		 * bios on current->bio_list, we first try the allocation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  475) 		 * without __GFP_DIRECT_RECLAIM; if that fails, we punt those
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  476) 		 * bios we would be blocking to the rescuer workqueue before
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  477) 		 * we retry with the original gfp_flags.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  478) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  479) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  480) 		if (current->bio_list &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  481) 		    (!bio_list_empty(&current->bio_list[0]) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  482) 		     !bio_list_empty(&current->bio_list[1])) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  483) 		    bs->rescue_workqueue)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  484) 			gfp_mask &= ~__GFP_DIRECT_RECLAIM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  485) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  486) 		p = mempool_alloc(&bs->bio_pool, gfp_mask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  487) 		if (!p && gfp_mask != saved_gfp) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  488) 			punt_bios_to_rescuer(bs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  489) 			gfp_mask = saved_gfp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  490) 			p = mempool_alloc(&bs->bio_pool, gfp_mask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  491) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  492) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  493) 		front_pad = bs->front_pad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  494) 		inline_vecs = BIO_INLINE_VECS;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  495) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  496) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  497) 	if (unlikely(!p))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  498) 		return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  499) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  500) 	bio = p + front_pad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  501) 	bio_init(bio, NULL, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  502) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  503) 	if (nr_iovecs > inline_vecs) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  504) 		unsigned long idx = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  505) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  506) 		bvl = bvec_alloc(gfp_mask, nr_iovecs, &idx, &bs->bvec_pool);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  507) 		if (!bvl && gfp_mask != saved_gfp) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  508) 			punt_bios_to_rescuer(bs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  509) 			gfp_mask = saved_gfp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  510) 			bvl = bvec_alloc(gfp_mask, nr_iovecs, &idx, &bs->bvec_pool);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  511) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  512) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  513) 		if (unlikely(!bvl))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  514) 			goto err_free;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  515) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  516) 		bio->bi_flags |= idx << BVEC_POOL_OFFSET;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  517) 	} else if (nr_iovecs) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  518) 		bvl = bio->bi_inline_vecs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  519) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  520) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  521) 	bio->bi_pool = bs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  522) 	bio->bi_max_vecs = nr_iovecs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  523) 	bio->bi_io_vec = bvl;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  524) 	return bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  525) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  526) err_free:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  527) 	mempool_free(p, &bs->bio_pool);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  528) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  529) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  530) EXPORT_SYMBOL(bio_alloc_bioset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  531) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  532) void zero_fill_bio_iter(struct bio *bio, struct bvec_iter start)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  533) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  534) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  535) 	struct bio_vec bv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  536) 	struct bvec_iter iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  537) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  538) 	__bio_for_each_segment(bv, bio, iter, start) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  539) 		char *data = bvec_kmap_irq(&bv, &flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  540) 		memset(data, 0, bv.bv_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  541) 		flush_dcache_page(bv.bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  542) 		bvec_kunmap_irq(data, &flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  543) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  544) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  545) EXPORT_SYMBOL(zero_fill_bio_iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  546) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  547) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  548)  * bio_truncate - truncate the bio to small size of @new_size
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  549)  * @bio:	the bio to be truncated
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  550)  * @new_size:	new size for truncating the bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  551)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  552)  * Description:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  553)  *   Truncate the bio to new size of @new_size. If bio_op(bio) is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  554)  *   REQ_OP_READ, zero the truncated part. This function should only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  555)  *   be used for handling corner cases, such as bio eod.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  556)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  557) void bio_truncate(struct bio *bio, unsigned new_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  558) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  559) 	struct bio_vec bv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  560) 	struct bvec_iter iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  561) 	unsigned int done = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  562) 	bool truncated = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  563) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  564) 	if (new_size >= bio->bi_iter.bi_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  565) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  566) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  567) 	if (bio_op(bio) != REQ_OP_READ)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  568) 		goto exit;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  569) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  570) 	bio_for_each_segment(bv, bio, iter) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  571) 		if (done + bv.bv_len > new_size) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  572) 			unsigned offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  573) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  574) 			if (!truncated)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  575) 				offset = new_size - done;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  576) 			else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  577) 				offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  578) 			zero_user(bv.bv_page, bv.bv_offset + offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  579) 				  bv.bv_len - offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  580) 			truncated = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  581) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  582) 		done += bv.bv_len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  583) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  584) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  585)  exit:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  586) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  587) 	 * Don't touch bvec table here and make it really immutable, since
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  588) 	 * fs bio user has to retrieve all pages via bio_for_each_segment_all
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  589) 	 * in its .end_bio() callback.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  590) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  591) 	 * It is enough to truncate bio by updating .bi_size since we can make
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  592) 	 * correct bvec with the updated .bi_size for drivers.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  593) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  594) 	bio->bi_iter.bi_size = new_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  595) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  596) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  597) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  598)  * guard_bio_eod - truncate a BIO to fit the block device
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  599)  * @bio:	bio to truncate
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  600)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  601)  * This allows us to do IO even on the odd last sectors of a device, even if the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  602)  * block size is some multiple of the physical sector size.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  603)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  604)  * We'll just truncate the bio to the size of the device, and clear the end of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  605)  * the buffer head manually.  Truly out-of-range accesses will turn into actual
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  606)  * I/O errors, this only handles the "we need to be able to do I/O at the final
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  607)  * sector" case.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  608)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  609) void guard_bio_eod(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  610) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  611) 	sector_t maxsector;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  612) 	struct hd_struct *part;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  613) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  614) 	rcu_read_lock();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  615) 	part = __disk_get_part(bio->bi_disk, bio->bi_partno);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  616) 	if (part)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  617) 		maxsector = part_nr_sects_read(part);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  618) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  619) 		maxsector = get_capacity(bio->bi_disk);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  620) 	rcu_read_unlock();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  621) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  622) 	if (!maxsector)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  623) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  624) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  625) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  626) 	 * If the *whole* IO is past the end of the device,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  627) 	 * let it through, and the IO layer will turn it into
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  628) 	 * an EIO.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  629) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  630) 	if (unlikely(bio->bi_iter.bi_sector >= maxsector))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  631) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  632) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  633) 	maxsector -= bio->bi_iter.bi_sector;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  634) 	if (likely((bio->bi_iter.bi_size >> 9) <= maxsector))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  635) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  636) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  637) 	bio_truncate(bio, maxsector << 9);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  638) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  639) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  640) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  641)  * bio_put - release a reference to a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  642)  * @bio:   bio to release reference to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  643)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  644)  * Description:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  645)  *   Put a reference to a &struct bio, either one you have gotten with
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  646)  *   bio_alloc, bio_get or bio_clone_*. The last put of a bio will free it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  647)  **/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  648) void bio_put(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  649) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  650) 	if (!bio_flagged(bio, BIO_REFFED))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  651) 		bio_free(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  652) 	else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  653) 		BIO_BUG_ON(!atomic_read(&bio->__bi_cnt));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  654) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  655) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  656) 		 * last put frees it
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  657) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  658) 		if (atomic_dec_and_test(&bio->__bi_cnt))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  659) 			bio_free(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  660) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  661) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  662) EXPORT_SYMBOL(bio_put);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  663) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  664) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  665)  * 	__bio_clone_fast - clone a bio that shares the original bio's biovec
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  666)  * 	@bio: destination bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  667)  * 	@bio_src: bio to clone
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  668)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  669)  *	Clone a &bio. Caller will own the returned bio, but not
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  670)  *	the actual data it points to. Reference count of returned
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  671)  * 	bio will be one.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  672)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  673)  * 	Caller must ensure that @bio_src is not freed before @bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  674)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  675) void __bio_clone_fast(struct bio *bio, struct bio *bio_src)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  676) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  677) 	BUG_ON(bio->bi_pool && BVEC_POOL_IDX(bio));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  678) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  679) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  680) 	 * most users will be overriding ->bi_disk with a new target,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  681) 	 * so we don't set nor calculate new physical/hw segment counts here
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  682) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  683) 	bio->bi_disk = bio_src->bi_disk;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  684) 	bio->bi_partno = bio_src->bi_partno;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  685) 	bio_set_flag(bio, BIO_CLONED);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  686) 	if (bio_flagged(bio_src, BIO_THROTTLED))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  687) 		bio_set_flag(bio, BIO_THROTTLED);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  688) 	bio->bi_opf = bio_src->bi_opf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  689) 	bio->bi_ioprio = bio_src->bi_ioprio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  690) 	bio->bi_write_hint = bio_src->bi_write_hint;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  691) 	bio->bi_iter = bio_src->bi_iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  692) 	bio->bi_io_vec = bio_src->bi_io_vec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  693) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  694) 	bio_clone_blkg_association(bio, bio_src);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  695) 	blkcg_bio_issue_init(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  696) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  697) EXPORT_SYMBOL(__bio_clone_fast);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  698) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  699) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  700)  *	bio_clone_fast - clone a bio that shares the original bio's biovec
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  701)  *	@bio: bio to clone
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  702)  *	@gfp_mask: allocation priority
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  703)  *	@bs: bio_set to allocate from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  704)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  705)  * 	Like __bio_clone_fast, only also allocates the returned bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  706)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  707) struct bio *bio_clone_fast(struct bio *bio, gfp_t gfp_mask, struct bio_set *bs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  708) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  709) 	struct bio *b;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  710) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  711) 	b = bio_alloc_bioset(gfp_mask, 0, bs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  712) 	if (!b)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  713) 		return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  714) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  715) 	__bio_clone_fast(b, bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  716) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  717) 	if (bio_crypt_clone(b, bio, gfp_mask) < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  718) 		goto err_put;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  719) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  720) 	if (bio_integrity(bio) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  721) 	    bio_integrity_clone(b, bio, gfp_mask) < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  722) 		goto err_put;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  723) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  724) 	return b;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  725) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  726) err_put:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  727) 	bio_put(b);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  728) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  729) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  730) EXPORT_SYMBOL(bio_clone_fast);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  731) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  732) const char *bio_devname(struct bio *bio, char *buf)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  733) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  734) 	return disk_name(bio->bi_disk, bio->bi_partno, buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  735) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  736) EXPORT_SYMBOL(bio_devname);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  737) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  738) static inline bool page_is_mergeable(const struct bio_vec *bv,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  739) 		struct page *page, unsigned int len, unsigned int off,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  740) 		bool *same_page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  741) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  742) 	size_t bv_end = bv->bv_offset + bv->bv_len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  743) 	phys_addr_t vec_end_addr = page_to_phys(bv->bv_page) + bv_end - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  744) 	phys_addr_t page_addr = page_to_phys(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  745) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  746) 	if (vec_end_addr + 1 != page_addr + off)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  747) 		return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  748) 	if (xen_domain() && !xen_biovec_phys_mergeable(bv, page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  749) 		return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  750) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  751) 	*same_page = ((vec_end_addr & PAGE_MASK) == page_addr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  752) 	if (*same_page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  753) 		return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  754) 	return (bv->bv_page + bv_end / PAGE_SIZE) == (page + off / PAGE_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  755) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  756) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  757) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  758)  * Try to merge a page into a segment, while obeying the hardware segment
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  759)  * size limit.  This is not for normal read/write bios, but for passthrough
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  760)  * or Zone Append operations that we can't split.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  761)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  762) static bool bio_try_merge_hw_seg(struct request_queue *q, struct bio *bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  763) 				 struct page *page, unsigned len,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  764) 				 unsigned offset, bool *same_page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  765) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  766) 	struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt - 1];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  767) 	unsigned long mask = queue_segment_boundary(q);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  768) 	phys_addr_t addr1 = page_to_phys(bv->bv_page) + bv->bv_offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  769) 	phys_addr_t addr2 = page_to_phys(page) + offset + len - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  770) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  771) 	if ((addr1 | mask) != (addr2 | mask))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  772) 		return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  773) 	if (bv->bv_len + len > queue_max_segment_size(q))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  774) 		return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  775) 	return __bio_try_merge_page(bio, page, len, offset, same_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  776) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  777) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  778) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  779)  * bio_add_hw_page - attempt to add a page to a bio with hw constraints
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  780)  * @q: the target queue
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  781)  * @bio: destination bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  782)  * @page: page to add
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  783)  * @len: vec entry length
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  784)  * @offset: vec entry offset
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  785)  * @max_sectors: maximum number of sectors that can be added
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  786)  * @same_page: return if the segment has been merged inside the same page
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  787)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  788)  * Add a page to a bio while respecting the hardware max_sectors, max_segment
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  789)  * and gap limitations.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  790)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  791) int bio_add_hw_page(struct request_queue *q, struct bio *bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  792) 		struct page *page, unsigned int len, unsigned int offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  793) 		unsigned int max_sectors, bool *same_page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  794) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  795) 	struct bio_vec *bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  796) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  797) 	if (WARN_ON_ONCE(bio_flagged(bio, BIO_CLONED)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  798) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  799) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  800) 	if (((bio->bi_iter.bi_size + len) >> 9) > max_sectors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  801) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  802) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  803) 	if (bio->bi_vcnt > 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  804) 		if (bio_try_merge_hw_seg(q, bio, page, len, offset, same_page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  805) 			return len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  806) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  807) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  808) 		 * If the queue doesn't support SG gaps and adding this segment
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  809) 		 * would create a gap, disallow it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  810) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  811) 		bvec = &bio->bi_io_vec[bio->bi_vcnt - 1];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  812) 		if (bvec_gap_to_prev(q, bvec, offset))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  813) 			return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  814) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  815) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  816) 	if (bio_full(bio, len))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  817) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  818) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  819) 	if (bio->bi_vcnt >= queue_max_segments(q))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  820) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  821) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  822) 	bvec = &bio->bi_io_vec[bio->bi_vcnt];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  823) 	bvec->bv_page = page;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  824) 	bvec->bv_len = len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  825) 	bvec->bv_offset = offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  826) 	bio->bi_vcnt++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  827) 	bio->bi_iter.bi_size += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  828) 	return len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  829) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  830) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  831) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  832)  * bio_add_pc_page	- attempt to add page to passthrough bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  833)  * @q: the target queue
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  834)  * @bio: destination bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  835)  * @page: page to add
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  836)  * @len: vec entry length
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  837)  * @offset: vec entry offset
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  838)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  839)  * Attempt to add a page to the bio_vec maplist. This can fail for a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  840)  * number of reasons, such as the bio being full or target block device
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  841)  * limitations. The target block device must allow bio's up to PAGE_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  842)  * so it is always possible to add a single page to an empty bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  843)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  844)  * This should only be used by passthrough bios.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  845)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  846) int bio_add_pc_page(struct request_queue *q, struct bio *bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  847) 		struct page *page, unsigned int len, unsigned int offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  848) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  849) 	bool same_page = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  850) 	return bio_add_hw_page(q, bio, page, len, offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  851) 			queue_max_hw_sectors(q), &same_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  852) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  853) EXPORT_SYMBOL(bio_add_pc_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  854) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  855) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  856)  * __bio_try_merge_page - try appending data to an existing bvec.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  857)  * @bio: destination bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  858)  * @page: start page to add
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  859)  * @len: length of the data to add
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  860)  * @off: offset of the data relative to @page
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  861)  * @same_page: return if the segment has been merged inside the same page
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  862)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  863)  * Try to add the data at @page + @off to the last bvec of @bio.  This is a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  864)  * useful optimisation for file systems with a block size smaller than the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  865)  * page size.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  866)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  867)  * Warn if (@len, @off) crosses pages in case that @same_page is true.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  868)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  869)  * Return %true on success or %false on failure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  870)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  871) bool __bio_try_merge_page(struct bio *bio, struct page *page,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  872) 		unsigned int len, unsigned int off, bool *same_page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  873) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  874) 	if (WARN_ON_ONCE(bio_flagged(bio, BIO_CLONED)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  875) 		return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  876) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  877) 	if (bio->bi_vcnt > 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  878) 		struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt - 1];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  879) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  880) 		if (page_is_mergeable(bv, page, len, off, same_page)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  881) 			if (bio->bi_iter.bi_size > UINT_MAX - len) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  882) 				*same_page = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  883) 				return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  884) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  885) 			bv->bv_len += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  886) 			bio->bi_iter.bi_size += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  887) 			return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  888) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  889) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  890) 	return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  891) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  892) EXPORT_SYMBOL_GPL(__bio_try_merge_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  893) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  894) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  895)  * __bio_add_page - add page(s) to a bio in a new segment
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  896)  * @bio: destination bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  897)  * @page: start page to add
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  898)  * @len: length of the data to add, may cross pages
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  899)  * @off: offset of the data relative to @page, may cross pages
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  900)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  901)  * Add the data at @page + @off to @bio as a new bvec.  The caller must ensure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  902)  * that @bio has space for another bvec.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  903)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  904) void __bio_add_page(struct bio *bio, struct page *page,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  905) 		unsigned int len, unsigned int off)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  906) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  907) 	struct bio_vec *bv = &bio->bi_io_vec[bio->bi_vcnt];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  908) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  909) 	WARN_ON_ONCE(bio_flagged(bio, BIO_CLONED));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  910) 	WARN_ON_ONCE(bio_full(bio, len));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  911) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  912) 	bv->bv_page = page;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  913) 	bv->bv_offset = off;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  914) 	bv->bv_len = len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  915) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  916) 	bio->bi_iter.bi_size += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  917) 	bio->bi_vcnt++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  918) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  919) 	if (!bio_flagged(bio, BIO_WORKINGSET) && unlikely(PageWorkingset(page)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  920) 		bio_set_flag(bio, BIO_WORKINGSET);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  921) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  922) EXPORT_SYMBOL_GPL(__bio_add_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  923) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  924) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  925)  *	bio_add_page	-	attempt to add page(s) to bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  926)  *	@bio: destination bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  927)  *	@page: start page to add
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  928)  *	@len: vec entry length, may cross pages
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  929)  *	@offset: vec entry offset relative to @page, may cross pages
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  930)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  931)  *	Attempt to add page(s) to the bio_vec maplist. This will only fail
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  932)  *	if either bio->bi_vcnt == bio->bi_max_vecs or it's a cloned bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  933)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  934) int bio_add_page(struct bio *bio, struct page *page,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  935) 		 unsigned int len, unsigned int offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  936) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  937) 	bool same_page = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  938) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  939) 	if (!__bio_try_merge_page(bio, page, len, offset, &same_page)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  940) 		if (bio_full(bio, len))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  941) 			return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  942) 		__bio_add_page(bio, page, len, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  943) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  944) 	return len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  945) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  946) EXPORT_SYMBOL(bio_add_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  947) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  948) void bio_release_pages(struct bio *bio, bool mark_dirty)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  949) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  950) 	struct bvec_iter_all iter_all;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  951) 	struct bio_vec *bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  952) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  953) 	if (bio_flagged(bio, BIO_NO_PAGE_REF))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  954) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  955) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  956) 	bio_for_each_segment_all(bvec, bio, iter_all) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  957) 		if (mark_dirty && !PageCompound(bvec->bv_page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  958) 			set_page_dirty_lock(bvec->bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  959) 		put_page(bvec->bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  960) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  961) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  962) EXPORT_SYMBOL_GPL(bio_release_pages);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  963) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  964) static int __bio_iov_bvec_add_pages(struct bio *bio, struct iov_iter *iter)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  965) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  966) 	const struct bio_vec *bv = iter->bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  967) 	unsigned int len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  968) 	size_t size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  969) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  970) 	if (WARN_ON_ONCE(iter->iov_offset > bv->bv_len))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  971) 		return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  972) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  973) 	len = min_t(size_t, bv->bv_len - iter->iov_offset, iter->count);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  974) 	size = bio_add_page(bio, bv->bv_page, len,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  975) 				bv->bv_offset + iter->iov_offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  976) 	if (unlikely(size != len))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  977) 		return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  978) 	iov_iter_advance(iter, size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  979) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  980) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  981) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  982) static void bio_put_pages(struct page **pages, size_t size, size_t off)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  983) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  984) 	size_t i, nr = DIV_ROUND_UP(size + (off & ~PAGE_MASK), PAGE_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  985) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  986) 	for (i = 0; i < nr; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  987) 		put_page(pages[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  988) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  989) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  990) #define PAGE_PTRS_PER_BVEC     (sizeof(struct bio_vec) / sizeof(struct page *))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  991) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  992) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  993)  * __bio_iov_iter_get_pages - pin user or kernel pages and add them to a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  994)  * @bio: bio to add pages to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  995)  * @iter: iov iterator describing the region to be mapped
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  996)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  997)  * Pins pages from *iter and appends them to @bio's bvec array. The
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  998)  * pages will have to be released using put_page() when done.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  999)  * For multi-segment *iter, this function only adds pages from the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1000)  * next non-empty segment of the iov iterator.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1001)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1002) static int __bio_iov_iter_get_pages(struct bio *bio, struct iov_iter *iter)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1003) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1004) 	unsigned short nr_pages = bio->bi_max_vecs - bio->bi_vcnt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1005) 	unsigned short entries_left = bio->bi_max_vecs - bio->bi_vcnt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1006) 	struct bio_vec *bv = bio->bi_io_vec + bio->bi_vcnt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1007) 	struct page **pages = (struct page **)bv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1008) 	bool same_page = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1009) 	ssize_t size, left;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1010) 	unsigned len, i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1011) 	size_t offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1012) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1013) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1014) 	 * Move page array up in the allocated memory for the bio vecs as far as
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1015) 	 * possible so that we can start filling biovecs from the beginning
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1016) 	 * without overwriting the temporary page array.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1017) 	*/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1018) 	BUILD_BUG_ON(PAGE_PTRS_PER_BVEC < 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1019) 	pages += entries_left * (PAGE_PTRS_PER_BVEC - 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1020) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1021) 	size = iov_iter_get_pages(iter, pages, LONG_MAX, nr_pages, &offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1022) 	if (unlikely(size <= 0))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1023) 		return size ? size : -EFAULT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1024) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1025) 	for (left = size, i = 0; left > 0; left -= len, i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1026) 		struct page *page = pages[i];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1027) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1028) 		len = min_t(size_t, PAGE_SIZE - offset, left);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1029) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1030) 		if (__bio_try_merge_page(bio, page, len, offset, &same_page)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1031) 			if (same_page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1032) 				put_page(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1033) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1034) 			if (WARN_ON_ONCE(bio_full(bio, len))) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1035) 				bio_put_pages(pages + i, left, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1036) 				return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1037) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1038) 			__bio_add_page(bio, page, len, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1039) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1040) 		offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1041) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1042) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1043) 	iov_iter_advance(iter, size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1044) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1045) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1046) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1047) static int __bio_iov_append_get_pages(struct bio *bio, struct iov_iter *iter)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1048) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1049) 	unsigned short nr_pages = bio->bi_max_vecs - bio->bi_vcnt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1050) 	unsigned short entries_left = bio->bi_max_vecs - bio->bi_vcnt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1051) 	struct request_queue *q = bio->bi_disk->queue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1052) 	unsigned int max_append_sectors = queue_max_zone_append_sectors(q);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1053) 	struct bio_vec *bv = bio->bi_io_vec + bio->bi_vcnt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1054) 	struct page **pages = (struct page **)bv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1055) 	ssize_t size, left;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1056) 	unsigned len, i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1057) 	size_t offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1058) 	int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1059) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1060) 	if (WARN_ON_ONCE(!max_append_sectors))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1061) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1062) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1063) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1064) 	 * Move page array up in the allocated memory for the bio vecs as far as
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1065) 	 * possible so that we can start filling biovecs from the beginning
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1066) 	 * without overwriting the temporary page array.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1067) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1068) 	BUILD_BUG_ON(PAGE_PTRS_PER_BVEC < 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1069) 	pages += entries_left * (PAGE_PTRS_PER_BVEC - 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1070) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1071) 	size = iov_iter_get_pages(iter, pages, LONG_MAX, nr_pages, &offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1072) 	if (unlikely(size <= 0))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1073) 		return size ? size : -EFAULT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1074) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1075) 	for (left = size, i = 0; left > 0; left -= len, i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1076) 		struct page *page = pages[i];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1077) 		bool same_page = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1078) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1079) 		len = min_t(size_t, PAGE_SIZE - offset, left);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1080) 		if (bio_add_hw_page(q, bio, page, len, offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1081) 				max_append_sectors, &same_page) != len) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1082) 			bio_put_pages(pages + i, left, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1083) 			ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1084) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1085) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1086) 		if (same_page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1087) 			put_page(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1088) 		offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1089) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1090) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1091) 	iov_iter_advance(iter, size - left);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1092) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1093) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1094) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1095) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1096)  * bio_iov_iter_get_pages - add user or kernel pages to a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1097)  * @bio: bio to add pages to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1098)  * @iter: iov iterator describing the region to be added
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1099)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1100)  * This takes either an iterator pointing to user memory, or one pointing to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1101)  * kernel pages (BVEC iterator). If we're adding user pages, we pin them and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1102)  * map them into the kernel. On IO completion, the caller should put those
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1103)  * pages. If we're adding kernel pages, and the caller told us it's safe to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1104)  * do so, we just have to add the pages to the bio directly. We don't grab an
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1105)  * extra reference to those pages (the user should already have that), and we
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1106)  * don't put the page on IO completion. The caller needs to check if the bio is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1107)  * flagged BIO_NO_PAGE_REF on IO completion. If it isn't, then pages should be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1108)  * released.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1109)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1110)  * The function tries, but does not guarantee, to pin as many pages as
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1111)  * fit into the bio, or are requested in @iter, whatever is smaller. If
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1112)  * MM encounters an error pinning the requested pages, it stops. Error
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1113)  * is returned only if 0 pages could be pinned.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1114)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1115) int bio_iov_iter_get_pages(struct bio *bio, struct iov_iter *iter)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1116) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1117) 	const bool is_bvec = iov_iter_is_bvec(iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1118) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1119) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1120) 	if (WARN_ON_ONCE(bio->bi_vcnt))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1121) 		return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1122) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1123) 	do {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1124) 		if (bio_op(bio) == REQ_OP_ZONE_APPEND) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1125) 			if (WARN_ON_ONCE(is_bvec))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1126) 				return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1127) 			ret = __bio_iov_append_get_pages(bio, iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1128) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1129) 			if (is_bvec)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1130) 				ret = __bio_iov_bvec_add_pages(bio, iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1131) 			else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1132) 				ret = __bio_iov_iter_get_pages(bio, iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1133) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1134) 	} while (!ret && iov_iter_count(iter) && !bio_full(bio, 0));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1135) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1136) 	if (is_bvec)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1137) 		bio_set_flag(bio, BIO_NO_PAGE_REF);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1138) 	return bio->bi_vcnt ? 0 : ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1139) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1140) EXPORT_SYMBOL_GPL(bio_iov_iter_get_pages);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1141) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1142) static void submit_bio_wait_endio(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1143) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1144) 	complete(bio->bi_private);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1145) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1146) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1147) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1148)  * submit_bio_wait - submit a bio, and wait until it completes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1149)  * @bio: The &struct bio which describes the I/O
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1150)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1151)  * Simple wrapper around submit_bio(). Returns 0 on success, or the error from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1152)  * bio_endio() on failure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1153)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1154)  * WARNING: Unlike to how submit_bio() is usually used, this function does not
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1155)  * result in bio reference to be consumed. The caller must drop the reference
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1156)  * on his own.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1157)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1158) int submit_bio_wait(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1159) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1160) 	DECLARE_COMPLETION_ONSTACK_MAP(done, bio->bi_disk->lockdep_map);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1161) 	unsigned long hang_check;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1162) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1163) 	bio->bi_private = &done;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1164) 	bio->bi_end_io = submit_bio_wait_endio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1165) 	bio->bi_opf |= REQ_SYNC;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1166) 	submit_bio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1167) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1168) 	/* Prevent hang_check timer from firing at us during very long I/O */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1169) 	hang_check = sysctl_hung_task_timeout_secs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1170) 	if (hang_check)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1171) 		while (!wait_for_completion_io_timeout(&done,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1172) 					hang_check * (HZ/2)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1173) 			;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1174) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1175) 		wait_for_completion_io(&done);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1176) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1177) 	return blk_status_to_errno(bio->bi_status);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1178) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1179) EXPORT_SYMBOL(submit_bio_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1180) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1181) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1182)  * bio_advance - increment/complete a bio by some number of bytes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1183)  * @bio:	bio to advance
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1184)  * @bytes:	number of bytes to complete
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1185)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1186)  * This updates bi_sector, bi_size and bi_idx; if the number of bytes to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1187)  * complete doesn't align with a bvec boundary, then bv_len and bv_offset will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1188)  * be updated on the last bvec as well.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1189)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1190)  * @bio will then represent the remaining, uncompleted portion of the io.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1191)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1192) void bio_advance(struct bio *bio, unsigned bytes)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1193) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1194) 	if (bio_integrity(bio))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1195) 		bio_integrity_advance(bio, bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1196) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1197) 	bio_crypt_advance(bio, bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1198) 	bio_advance_iter(bio, &bio->bi_iter, bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1199) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1200) EXPORT_SYMBOL(bio_advance);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1201) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1202) void bio_copy_data_iter(struct bio *dst, struct bvec_iter *dst_iter,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1203) 			struct bio *src, struct bvec_iter *src_iter)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1204) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1205) 	struct bio_vec src_bv, dst_bv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1206) 	void *src_p, *dst_p;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1207) 	unsigned bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1208) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1209) 	while (src_iter->bi_size && dst_iter->bi_size) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1210) 		src_bv = bio_iter_iovec(src, *src_iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1211) 		dst_bv = bio_iter_iovec(dst, *dst_iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1212) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1213) 		bytes = min(src_bv.bv_len, dst_bv.bv_len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1214) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1215) 		src_p = kmap_atomic(src_bv.bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1216) 		dst_p = kmap_atomic(dst_bv.bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1217) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1218) 		memcpy(dst_p + dst_bv.bv_offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1219) 		       src_p + src_bv.bv_offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1220) 		       bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1221) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1222) 		kunmap_atomic(dst_p);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1223) 		kunmap_atomic(src_p);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1224) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1225) 		flush_dcache_page(dst_bv.bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1226) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1227) 		bio_advance_iter(src, src_iter, bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1228) 		bio_advance_iter(dst, dst_iter, bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1229) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1230) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1231) EXPORT_SYMBOL(bio_copy_data_iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1232) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1233) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1234)  * bio_copy_data - copy contents of data buffers from one bio to another
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1235)  * @src: source bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1236)  * @dst: destination bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1237)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1238)  * Stops when it reaches the end of either @src or @dst - that is, copies
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1239)  * min(src->bi_size, dst->bi_size) bytes (or the equivalent for lists of bios).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1240)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1241) void bio_copy_data(struct bio *dst, struct bio *src)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1242) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1243) 	struct bvec_iter src_iter = src->bi_iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1244) 	struct bvec_iter dst_iter = dst->bi_iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1245) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1246) 	bio_copy_data_iter(dst, &dst_iter, src, &src_iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1247) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1248) EXPORT_SYMBOL(bio_copy_data);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1249) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1250) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1251)  * bio_list_copy_data - copy contents of data buffers from one chain of bios to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1252)  * another
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1253)  * @src: source bio list
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1254)  * @dst: destination bio list
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1255)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1256)  * Stops when it reaches the end of either the @src list or @dst list - that is,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1257)  * copies min(src->bi_size, dst->bi_size) bytes (or the equivalent for lists of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1258)  * bios).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1259)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1260) void bio_list_copy_data(struct bio *dst, struct bio *src)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1261) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1262) 	struct bvec_iter src_iter = src->bi_iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1263) 	struct bvec_iter dst_iter = dst->bi_iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1264) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1265) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1266) 		if (!src_iter.bi_size) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1267) 			src = src->bi_next;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1268) 			if (!src)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1269) 				break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1270) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1271) 			src_iter = src->bi_iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1272) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1273) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1274) 		if (!dst_iter.bi_size) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1275) 			dst = dst->bi_next;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1276) 			if (!dst)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1277) 				break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1278) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1279) 			dst_iter = dst->bi_iter;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1280) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1281) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1282) 		bio_copy_data_iter(dst, &dst_iter, src, &src_iter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1283) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1284) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1285) EXPORT_SYMBOL(bio_list_copy_data);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1286) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1287) void bio_free_pages(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1288) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1289) 	struct bio_vec *bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1290) 	struct bvec_iter_all iter_all;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1291) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1292) 	bio_for_each_segment_all(bvec, bio, iter_all)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1293) 		__free_page(bvec->bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1294) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1295) EXPORT_SYMBOL(bio_free_pages);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1296) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1297) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1298)  * bio_set_pages_dirty() and bio_check_pages_dirty() are support functions
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1299)  * for performing direct-IO in BIOs.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1300)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1301)  * The problem is that we cannot run set_page_dirty() from interrupt context
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1302)  * because the required locks are not interrupt-safe.  So what we can do is to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1303)  * mark the pages dirty _before_ performing IO.  And in interrupt context,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1304)  * check that the pages are still dirty.   If so, fine.  If not, redirty them
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1305)  * in process context.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1306)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1307)  * We special-case compound pages here: normally this means reads into hugetlb
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1308)  * pages.  The logic in here doesn't really work right for compound pages
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1309)  * because the VM does not uniformly chase down the head page in all cases.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1310)  * But dirtiness of compound pages is pretty meaningless anyway: the VM doesn't
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1311)  * handle them at all.  So we skip compound pages here at an early stage.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1312)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1313)  * Note that this code is very hard to test under normal circumstances because
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1314)  * direct-io pins the pages with get_user_pages().  This makes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1315)  * is_page_cache_freeable return false, and the VM will not clean the pages.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1316)  * But other code (eg, flusher threads) could clean the pages if they are mapped
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1317)  * pagecache.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1318)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1319)  * Simply disabling the call to bio_set_pages_dirty() is a good way to test the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1320)  * deferred bio dirtying paths.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1321)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1322) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1323) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1324)  * bio_set_pages_dirty() will mark all the bio's pages as dirty.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1325)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1326) void bio_set_pages_dirty(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1327) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1328) 	struct bio_vec *bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1329) 	struct bvec_iter_all iter_all;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1330) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1331) 	bio_for_each_segment_all(bvec, bio, iter_all) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1332) 		if (!PageCompound(bvec->bv_page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1333) 			set_page_dirty_lock(bvec->bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1334) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1335) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1336) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1337) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1338)  * bio_check_pages_dirty() will check that all the BIO's pages are still dirty.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1339)  * If they are, then fine.  If, however, some pages are clean then they must
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1340)  * have been written out during the direct-IO read.  So we take another ref on
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1341)  * the BIO and re-dirty the pages in process context.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1342)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1343)  * It is expected that bio_check_pages_dirty() will wholly own the BIO from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1344)  * here on.  It will run one put_page() against each page and will run one
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1345)  * bio_put() against the BIO.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1346)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1347) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1348) static void bio_dirty_fn(struct work_struct *work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1349) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1350) static DECLARE_WORK(bio_dirty_work, bio_dirty_fn);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1351) static DEFINE_SPINLOCK(bio_dirty_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1352) static struct bio *bio_dirty_list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1353) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1354) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1355)  * This runs in process context
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1356)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1357) static void bio_dirty_fn(struct work_struct *work)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1358) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1359) 	struct bio *bio, *next;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1360) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1361) 	spin_lock_irq(&bio_dirty_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1362) 	next = bio_dirty_list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1363) 	bio_dirty_list = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1364) 	spin_unlock_irq(&bio_dirty_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1365) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1366) 	while ((bio = next) != NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1367) 		next = bio->bi_private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1368) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1369) 		bio_release_pages(bio, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1370) 		bio_put(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1371) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1372) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1373) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1374) void bio_check_pages_dirty(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1375) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1376) 	struct bio_vec *bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1377) 	unsigned long flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1378) 	struct bvec_iter_all iter_all;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1379) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1380) 	bio_for_each_segment_all(bvec, bio, iter_all) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1381) 		if (!PageDirty(bvec->bv_page) && !PageCompound(bvec->bv_page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1382) 			goto defer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1383) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1384) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1385) 	bio_release_pages(bio, false);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1386) 	bio_put(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1387) 	return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1388) defer:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1389) 	spin_lock_irqsave(&bio_dirty_lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1390) 	bio->bi_private = bio_dirty_list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1391) 	bio_dirty_list = bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1392) 	spin_unlock_irqrestore(&bio_dirty_lock, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1393) 	schedule_work(&bio_dirty_work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1394) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1395) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1396) static inline bool bio_remaining_done(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1397) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1398) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1399) 	 * If we're not chaining, then ->__bi_remaining is always 1 and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1400) 	 * we always end io on the first invocation.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1401) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1402) 	if (!bio_flagged(bio, BIO_CHAIN))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1403) 		return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1404) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1405) 	BUG_ON(atomic_read(&bio->__bi_remaining) <= 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1406) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1407) 	if (atomic_dec_and_test(&bio->__bi_remaining)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1408) 		bio_clear_flag(bio, BIO_CHAIN);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1409) 		return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1410) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1411) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1412) 	return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1413) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1414) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1415) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1416)  * bio_endio - end I/O on a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1417)  * @bio:	bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1418)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1419)  * Description:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1420)  *   bio_endio() will end I/O on the whole bio. bio_endio() is the preferred
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1421)  *   way to end I/O on a bio. No one should call bi_end_io() directly on a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1422)  *   bio unless they own it and thus know that it has an end_io function.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1423)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1424)  *   bio_endio() can be called several times on a bio that has been chained
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1425)  *   using bio_chain().  The ->bi_end_io() function will only be called the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1426)  *   last time.  At this point the BLK_TA_COMPLETE tracing event will be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1427)  *   generated if BIO_TRACE_COMPLETION is set.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1428)  **/
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1429) void bio_endio(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1430) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1431) again:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1432) 	if (!bio_remaining_done(bio))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1433) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1434) 	if (!bio_integrity_endio(bio))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1435) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1436) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1437) 	if (bio->bi_disk)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1438) 		rq_qos_done_bio(bio->bi_disk->queue, bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1439) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1440) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1441) 	 * Need to have a real endio function for chained bios, otherwise
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1442) 	 * various corner cases will break (like stacking block devices that
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1443) 	 * save/restore bi_end_io) - however, we want to avoid unbounded
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1444) 	 * recursion and blowing the stack. Tail call optimization would
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1445) 	 * handle this, but compiling with frame pointers also disables
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1446) 	 * gcc's sibling call optimization.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1447) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1448) 	if (bio->bi_end_io == bio_chain_endio) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1449) 		bio = __bio_chain_endio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1450) 		goto again;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1451) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1452) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1453) 	if (bio->bi_disk && bio_flagged(bio, BIO_TRACE_COMPLETION)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1454) 		trace_block_bio_complete(bio->bi_disk->queue, bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1455) 		bio_clear_flag(bio, BIO_TRACE_COMPLETION);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1456) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1457) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1458) 	blk_throtl_bio_endio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1459) 	/* release cgroup info */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1460) 	bio_uninit(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1461) 	if (bio->bi_end_io)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1462) 		bio->bi_end_io(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1463) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1464) EXPORT_SYMBOL(bio_endio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1465) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1466) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1467)  * bio_split - split a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1468)  * @bio:	bio to split
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1469)  * @sectors:	number of sectors to split from the front of @bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1470)  * @gfp:	gfp mask
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1471)  * @bs:		bio set to allocate from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1472)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1473)  * Allocates and returns a new bio which represents @sectors from the start of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1474)  * @bio, and updates @bio to represent the remaining sectors.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1475)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1476)  * Unless this is a discard request the newly allocated bio will point
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1477)  * to @bio's bi_io_vec. It is the caller's responsibility to ensure that
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1478)  * neither @bio nor @bs are freed before the split bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1479)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1480) struct bio *bio_split(struct bio *bio, int sectors,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1481) 		      gfp_t gfp, struct bio_set *bs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1482) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1483) 	struct bio *split;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1484) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1485) 	BUG_ON(sectors <= 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1486) 	BUG_ON(sectors >= bio_sectors(bio));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1487) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1488) 	/* Zone append commands cannot be split */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1489) 	if (WARN_ON_ONCE(bio_op(bio) == REQ_OP_ZONE_APPEND))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1490) 		return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1491) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1492) 	split = bio_clone_fast(bio, gfp, bs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1493) 	if (!split)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1494) 		return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1495) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1496) 	split->bi_iter.bi_size = sectors << 9;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1497) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1498) 	if (bio_integrity(split))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1499) 		bio_integrity_trim(split);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1500) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1501) 	bio_advance(bio, split->bi_iter.bi_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1502) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1503) 	if (bio_flagged(bio, BIO_TRACE_COMPLETION))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1504) 		bio_set_flag(split, BIO_TRACE_COMPLETION);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1505) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1506) 	return split;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1507) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1508) EXPORT_SYMBOL(bio_split);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1509) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1510) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1511)  * bio_trim - trim a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1512)  * @bio:	bio to trim
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1513)  * @offset:	number of sectors to trim from the front of @bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1514)  * @size:	size we want to trim @bio to, in sectors
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1515)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1516) void bio_trim(struct bio *bio, int offset, int size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1517) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1518) 	/* 'bio' is a cloned bio which we need to trim to match
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1519) 	 * the given offset and size.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1520) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1521) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1522) 	size <<= 9;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1523) 	if (offset == 0 && size == bio->bi_iter.bi_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1524) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1525) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1526) 	bio_advance(bio, offset << 9);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1527) 	bio->bi_iter.bi_size = size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1528) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1529) 	if (bio_integrity(bio))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1530) 		bio_integrity_trim(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1531) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1532) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1533) EXPORT_SYMBOL_GPL(bio_trim);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1534) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1535) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1536)  * create memory pools for biovec's in a bio_set.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1537)  * use the global biovec slabs created for general use.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1538)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1539) int biovec_init_pool(mempool_t *pool, int pool_entries)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1540) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1541) 	struct biovec_slab *bp = bvec_slabs + BVEC_POOL_MAX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1542) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1543) 	return mempool_init_slab_pool(pool, pool_entries, bp->slab);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1544) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1545) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1546) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1547)  * bioset_exit - exit a bioset initialized with bioset_init()
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1548)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1549)  * May be called on a zeroed but uninitialized bioset (i.e. allocated with
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1550)  * kzalloc()).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1551)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1552) void bioset_exit(struct bio_set *bs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1553) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1554) 	if (bs->rescue_workqueue)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1555) 		destroy_workqueue(bs->rescue_workqueue);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1556) 	bs->rescue_workqueue = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1557) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1558) 	mempool_exit(&bs->bio_pool);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1559) 	mempool_exit(&bs->bvec_pool);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1560) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1561) 	bioset_integrity_free(bs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1562) 	if (bs->bio_slab)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1563) 		bio_put_slab(bs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1564) 	bs->bio_slab = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1565) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1566) EXPORT_SYMBOL(bioset_exit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1567) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1568) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1569)  * bioset_init - Initialize a bio_set
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1570)  * @bs:		pool to initialize
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1571)  * @pool_size:	Number of bio and bio_vecs to cache in the mempool
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1572)  * @front_pad:	Number of bytes to allocate in front of the returned bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1573)  * @flags:	Flags to modify behavior, currently %BIOSET_NEED_BVECS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1574)  *              and %BIOSET_NEED_RESCUER
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1575)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1576)  * Description:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1577)  *    Set up a bio_set to be used with @bio_alloc_bioset. Allows the caller
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1578)  *    to ask for a number of bytes to be allocated in front of the bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1579)  *    Front pad allocation is useful for embedding the bio inside
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1580)  *    another structure, to avoid allocating extra data to go with the bio.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1581)  *    Note that the bio must be embedded at the END of that structure always,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1582)  *    or things will break badly.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1583)  *    If %BIOSET_NEED_BVECS is set in @flags, a separate pool will be allocated
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1584)  *    for allocating iovecs.  This pool is not needed e.g. for bio_clone_fast().
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1585)  *    If %BIOSET_NEED_RESCUER is set, a workqueue is created which can be used to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1586)  *    dispatch queued requests when the mempool runs out of space.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1587)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1588)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1589) int bioset_init(struct bio_set *bs,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1590) 		unsigned int pool_size,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1591) 		unsigned int front_pad,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1592) 		int flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1593) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1594) 	unsigned int back_pad = BIO_INLINE_VECS * sizeof(struct bio_vec);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1595) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1596) 	bs->front_pad = front_pad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1597) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1598) 	spin_lock_init(&bs->rescue_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1599) 	bio_list_init(&bs->rescue_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1600) 	INIT_WORK(&bs->rescue_work, bio_alloc_rescue);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1601) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1602) 	bs->bio_slab = bio_find_or_create_slab(front_pad + back_pad);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1603) 	if (!bs->bio_slab)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1604) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1605) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1606) 	if (mempool_init_slab_pool(&bs->bio_pool, pool_size, bs->bio_slab))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1607) 		goto bad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1608) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1609) 	if ((flags & BIOSET_NEED_BVECS) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1610) 	    biovec_init_pool(&bs->bvec_pool, pool_size))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1611) 		goto bad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1612) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1613) 	if (!(flags & BIOSET_NEED_RESCUER))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1614) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1615) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1616) 	bs->rescue_workqueue = alloc_workqueue("bioset", WQ_MEM_RECLAIM, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1617) 	if (!bs->rescue_workqueue)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1618) 		goto bad;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1619) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1620) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1621) bad:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1622) 	bioset_exit(bs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1623) 	return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1624) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1625) EXPORT_SYMBOL(bioset_init);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1626) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1627) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1628)  * Initialize and setup a new bio_set, based on the settings from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1629)  * another bio_set.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1630)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1631) int bioset_init_from_src(struct bio_set *bs, struct bio_set *src)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1632) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1633) 	int flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1634) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1635) 	flags = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1636) 	if (src->bvec_pool.min_nr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1637) 		flags |= BIOSET_NEED_BVECS;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1638) 	if (src->rescue_workqueue)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1639) 		flags |= BIOSET_NEED_RESCUER;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1640) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1641) 	return bioset_init(bs, src->bio_pool.min_nr, src->front_pad, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1642) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1643) EXPORT_SYMBOL(bioset_init_from_src);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1644) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1645) static void __init biovec_init_slabs(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1646) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1647) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1648) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1649) 	for (i = 0; i < BVEC_POOL_NR; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1650) 		int size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1651) 		struct biovec_slab *bvs = bvec_slabs + i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1652) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1653) 		if (bvs->nr_vecs <= BIO_INLINE_VECS) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1654) 			bvs->slab = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1655) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1656) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1657) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1658) 		size = bvs->nr_vecs * sizeof(struct bio_vec);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1659) 		bvs->slab = kmem_cache_create(bvs->name, size, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1660)                                 SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1661) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1662) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1663) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1664) static int __init init_bio(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1665) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1666) 	bio_slab_max = 2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1667) 	bio_slab_nr = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1668) 	bio_slabs = kcalloc(bio_slab_max, sizeof(struct bio_slab),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1669) 			    GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1670) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1671) 	BUILD_BUG_ON(BIO_FLAG_LAST > BVEC_POOL_OFFSET);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1672) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1673) 	if (!bio_slabs)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1674) 		panic("bio: can't allocate bios\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1675) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1676) 	bio_integrity_init();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1677) 	biovec_init_slabs();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1678) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1679) 	if (bioset_init(&fs_bio_set, BIO_POOL_SIZE, 0, BIOSET_NEED_BVECS))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1680) 		panic("bio: can't allocate bios\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1681) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1682) 	if (bioset_integrity_create(&fs_bio_set, BIO_POOL_SIZE))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1683) 		panic("bio: can't create integrity pool\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1684) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1685) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1686) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1687) subsys_initcall(init_bio);