Orange Pi5 kernel

Deprecated Linux kernel 5.10.110 for OrangePi 5/5B/5+ boards

3 Commits   0 Branches   0 Tags
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    1) // SPDX-License-Identifier: GPL-2.0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    3)  * Copyright (C) 2007 Oracle.  All rights reserved.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    4)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    5) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    6) #include <linux/fs.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    7) #include <linux/blkdev.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    8) #include <linux/radix-tree.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    9) #include <linux/writeback.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   10) #include <linux/workqueue.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   11) #include <linux/kthread.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   12) #include <linux/slab.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   13) #include <linux/migrate.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   14) #include <linux/ratelimit.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   15) #include <linux/uuid.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   16) #include <linux/semaphore.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   17) #include <linux/error-injection.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   18) #include <linux/crc32c.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   19) #include <linux/sched/mm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   20) #include <asm/unaligned.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   21) #include <crypto/hash.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   22) #include "ctree.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   23) #include "disk-io.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   24) #include "transaction.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   25) #include "btrfs_inode.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   26) #include "volumes.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   27) #include "print-tree.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   28) #include "locking.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   29) #include "tree-log.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   30) #include "free-space-cache.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   31) #include "free-space-tree.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   32) #include "inode-map.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   33) #include "check-integrity.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   34) #include "rcu-string.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   35) #include "dev-replace.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   36) #include "raid56.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   37) #include "sysfs.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   38) #include "qgroup.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   39) #include "compression.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   40) #include "tree-checker.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   41) #include "ref-verify.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   42) #include "block-group.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   43) #include "discard.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   44) #include "space-info.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   45) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   46) #define BTRFS_SUPER_FLAG_SUPP	(BTRFS_HEADER_FLAG_WRITTEN |\
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   47) 				 BTRFS_HEADER_FLAG_RELOC |\
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   48) 				 BTRFS_SUPER_FLAG_ERROR |\
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   49) 				 BTRFS_SUPER_FLAG_SEEDING |\
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   50) 				 BTRFS_SUPER_FLAG_METADUMP |\
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   51) 				 BTRFS_SUPER_FLAG_METADUMP_V2)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   52) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   53) static void end_workqueue_fn(struct btrfs_work *work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   54) static void btrfs_destroy_ordered_extents(struct btrfs_root *root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   55) static int btrfs_destroy_delayed_refs(struct btrfs_transaction *trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   56) 				      struct btrfs_fs_info *fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   57) static void btrfs_destroy_delalloc_inodes(struct btrfs_root *root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   58) static int btrfs_destroy_marked_extents(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   59) 					struct extent_io_tree *dirty_pages,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   60) 					int mark);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   61) static int btrfs_destroy_pinned_extent(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   62) 				       struct extent_io_tree *pinned_extents);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   63) static int btrfs_cleanup_transaction(struct btrfs_fs_info *fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   64) static void btrfs_error_commit_super(struct btrfs_fs_info *fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   65) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   66) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   67)  * btrfs_end_io_wq structs are used to do processing in task context when an IO
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   68)  * is complete.  This is used during reads to verify checksums, and it is used
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   69)  * by writes to insert metadata for new file extents after IO is complete.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   70)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   71) struct btrfs_end_io_wq {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   72) 	struct bio *bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   73) 	bio_end_io_t *end_io;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   74) 	void *private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   75) 	struct btrfs_fs_info *info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   76) 	blk_status_t status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   77) 	enum btrfs_wq_endio_type metadata;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   78) 	struct btrfs_work work;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   79) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   80) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   81) static struct kmem_cache *btrfs_end_io_wq_cache;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   82) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   83) int __init btrfs_end_io_wq_init(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   84) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   85) 	btrfs_end_io_wq_cache = kmem_cache_create("btrfs_end_io_wq",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   86) 					sizeof(struct btrfs_end_io_wq),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   87) 					0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   88) 					SLAB_MEM_SPREAD,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   89) 					NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   90) 	if (!btrfs_end_io_wq_cache)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   91) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   92) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   93) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   94) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   95) void __cold btrfs_end_io_wq_exit(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   96) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   97) 	kmem_cache_destroy(btrfs_end_io_wq_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   98) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   99) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  100) static void btrfs_free_csum_hash(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  101) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  102) 	if (fs_info->csum_shash)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  103) 		crypto_free_shash(fs_info->csum_shash);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  104) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  105) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  106) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  107)  * async submit bios are used to offload expensive checksumming
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  108)  * onto the worker threads.  They checksum file and metadata bios
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  109)  * just before they are sent down the IO stack.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  110)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  111) struct async_submit_bio {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  112) 	void *private_data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  113) 	struct bio *bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  114) 	extent_submit_bio_start_t *submit_bio_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  115) 	int mirror_num;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  116) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  117) 	 * bio_offset is optional, can be used if the pages in the bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  118) 	 * can't tell us where in the file the bio should go
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  119) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  120) 	u64 bio_offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  121) 	struct btrfs_work work;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  122) 	blk_status_t status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  123) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  124) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  125) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  126)  * Lockdep class keys for extent_buffer->lock's in this root.  For a given
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  127)  * eb, the lockdep key is determined by the btrfs_root it belongs to and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  128)  * the level the eb occupies in the tree.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  129)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  130)  * Different roots are used for different purposes and may nest inside each
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  131)  * other and they require separate keysets.  As lockdep keys should be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  132)  * static, assign keysets according to the purpose of the root as indicated
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  133)  * by btrfs_root->root_key.objectid.  This ensures that all special purpose
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  134)  * roots have separate keysets.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  135)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  136)  * Lock-nesting across peer nodes is always done with the immediate parent
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  137)  * node locked thus preventing deadlock.  As lockdep doesn't know this, use
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  138)  * subclass to avoid triggering lockdep warning in such cases.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  139)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  140)  * The key is set by the readpage_end_io_hook after the buffer has passed
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  141)  * csum validation but before the pages are unlocked.  It is also set by
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  142)  * btrfs_init_new_buffer on freshly allocated blocks.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  143)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  144)  * We also add a check to make sure the highest level of the tree is the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  145)  * same as our lockdep setup here.  If BTRFS_MAX_LEVEL changes, this code
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  146)  * needs update as well.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  147)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  148) #ifdef CONFIG_DEBUG_LOCK_ALLOC
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  149) # if BTRFS_MAX_LEVEL != 8
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  150) #  error
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  151) # endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  152) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  153) static struct btrfs_lockdep_keyset {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  154) 	u64			id;		/* root objectid */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  155) 	const char		*name_stem;	/* lock name stem */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  156) 	char			names[BTRFS_MAX_LEVEL + 1][20];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  157) 	struct lock_class_key	keys[BTRFS_MAX_LEVEL + 1];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  158) } btrfs_lockdep_keysets[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  159) 	{ .id = BTRFS_ROOT_TREE_OBJECTID,	.name_stem = "root"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  160) 	{ .id = BTRFS_EXTENT_TREE_OBJECTID,	.name_stem = "extent"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  161) 	{ .id = BTRFS_CHUNK_TREE_OBJECTID,	.name_stem = "chunk"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  162) 	{ .id = BTRFS_DEV_TREE_OBJECTID,	.name_stem = "dev"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  163) 	{ .id = BTRFS_FS_TREE_OBJECTID,		.name_stem = "fs"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  164) 	{ .id = BTRFS_CSUM_TREE_OBJECTID,	.name_stem = "csum"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  165) 	{ .id = BTRFS_QUOTA_TREE_OBJECTID,	.name_stem = "quota"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  166) 	{ .id = BTRFS_TREE_LOG_OBJECTID,	.name_stem = "log"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  167) 	{ .id = BTRFS_TREE_RELOC_OBJECTID,	.name_stem = "treloc"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  168) 	{ .id = BTRFS_DATA_RELOC_TREE_OBJECTID,	.name_stem = "dreloc"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  169) 	{ .id = BTRFS_UUID_TREE_OBJECTID,	.name_stem = "uuid"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  170) 	{ .id = BTRFS_FREE_SPACE_TREE_OBJECTID,	.name_stem = "free-space" },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  171) 	{ .id = 0,				.name_stem = "tree"	},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  172) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  173) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  174) void __init btrfs_init_lockdep(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  175) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  176) 	int i, j;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  177) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  178) 	/* initialize lockdep class names */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  179) 	for (i = 0; i < ARRAY_SIZE(btrfs_lockdep_keysets); i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  180) 		struct btrfs_lockdep_keyset *ks = &btrfs_lockdep_keysets[i];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  181) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  182) 		for (j = 0; j < ARRAY_SIZE(ks->names); j++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  183) 			snprintf(ks->names[j], sizeof(ks->names[j]),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  184) 				 "btrfs-%s-%02d", ks->name_stem, j);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  185) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  186) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  187) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  188) void btrfs_set_buffer_lockdep_class(u64 objectid, struct extent_buffer *eb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  189) 				    int level)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  190) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  191) 	struct btrfs_lockdep_keyset *ks;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  192) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  193) 	BUG_ON(level >= ARRAY_SIZE(ks->keys));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  194) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  195) 	/* find the matching keyset, id 0 is the default entry */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  196) 	for (ks = btrfs_lockdep_keysets; ks->id; ks++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  197) 		if (ks->id == objectid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  198) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  199) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  200) 	lockdep_set_class_and_name(&eb->lock,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  201) 				   &ks->keys[level], ks->names[level]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  202) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  203) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  204) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  205) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  206) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  207)  * Compute the csum of a btree block and store the result to provided buffer.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  208)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  209) static void csum_tree_block(struct extent_buffer *buf, u8 *result)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  210) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  211) 	struct btrfs_fs_info *fs_info = buf->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  212) 	const int num_pages = fs_info->nodesize >> PAGE_SHIFT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  213) 	SHASH_DESC_ON_STACK(shash, fs_info->csum_shash);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  214) 	char *kaddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  215) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  216) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  217) 	shash->tfm = fs_info->csum_shash;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  218) 	crypto_shash_init(shash);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  219) 	kaddr = page_address(buf->pages[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  220) 	crypto_shash_update(shash, kaddr + BTRFS_CSUM_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  221) 			    PAGE_SIZE - BTRFS_CSUM_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  222) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  223) 	for (i = 1; i < num_pages; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  224) 		kaddr = page_address(buf->pages[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  225) 		crypto_shash_update(shash, kaddr, PAGE_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  226) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  227) 	memset(result, 0, BTRFS_CSUM_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  228) 	crypto_shash_final(shash, result);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  229) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  230) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  231) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  232)  * we can't consider a given block up to date unless the transid of the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  233)  * block matches the transid in the parent node's pointer.  This is how we
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  234)  * detect blocks that either didn't get written at all or got written
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  235)  * in the wrong place.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  236)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  237) static int verify_parent_transid(struct extent_io_tree *io_tree,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  238) 				 struct extent_buffer *eb, u64 parent_transid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  239) 				 int atomic)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  240) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  241) 	struct extent_state *cached_state = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  242) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  243) 	bool need_lock = (current->journal_info == BTRFS_SEND_TRANS_STUB);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  244) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  245) 	if (!parent_transid || btrfs_header_generation(eb) == parent_transid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  246) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  247) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  248) 	if (atomic)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  249) 		return -EAGAIN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  250) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  251) 	if (need_lock) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  252) 		btrfs_tree_read_lock(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  253) 		btrfs_set_lock_blocking_read(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  254) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  255) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  256) 	lock_extent_bits(io_tree, eb->start, eb->start + eb->len - 1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  257) 			 &cached_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  258) 	if (extent_buffer_uptodate(eb) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  259) 	    btrfs_header_generation(eb) == parent_transid) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  260) 		ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  261) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  262) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  263) 	btrfs_err_rl(eb->fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  264) 		"parent transid verify failed on %llu wanted %llu found %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  265) 			eb->start,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  266) 			parent_transid, btrfs_header_generation(eb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  267) 	ret = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  268) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  269) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  270) 	 * Things reading via commit roots that don't have normal protection,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  271) 	 * like send, can have a really old block in cache that may point at a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  272) 	 * block that has been freed and re-allocated.  So don't clear uptodate
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  273) 	 * if we find an eb that is under IO (dirty/writeback) because we could
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  274) 	 * end up reading in the stale data and then writing it back out and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  275) 	 * making everybody very sad.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  276) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  277) 	if (!extent_buffer_under_io(eb))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  278) 		clear_extent_buffer_uptodate(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  279) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  280) 	unlock_extent_cached(io_tree, eb->start, eb->start + eb->len - 1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  281) 			     &cached_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  282) 	if (need_lock)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  283) 		btrfs_tree_read_unlock_blocking(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  284) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  285) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  286) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  287) static bool btrfs_supported_super_csum(u16 csum_type)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  288) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  289) 	switch (csum_type) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  290) 	case BTRFS_CSUM_TYPE_CRC32:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  291) 	case BTRFS_CSUM_TYPE_XXHASH:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  292) 	case BTRFS_CSUM_TYPE_SHA256:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  293) 	case BTRFS_CSUM_TYPE_BLAKE2:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  294) 		return true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  295) 	default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  296) 		return false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  297) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  298) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  299) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  300) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  301)  * Return 0 if the superblock checksum type matches the checksum value of that
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  302)  * algorithm. Pass the raw disk superblock data.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  303)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  304) static int btrfs_check_super_csum(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  305) 				  char *raw_disk_sb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  306) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  307) 	struct btrfs_super_block *disk_sb =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  308) 		(struct btrfs_super_block *)raw_disk_sb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  309) 	char result[BTRFS_CSUM_SIZE];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  310) 	SHASH_DESC_ON_STACK(shash, fs_info->csum_shash);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  311) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  312) 	shash->tfm = fs_info->csum_shash;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  313) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  314) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  315) 	 * The super_block structure does not span the whole
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  316) 	 * BTRFS_SUPER_INFO_SIZE range, we expect that the unused space is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  317) 	 * filled with zeros and is included in the checksum.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  318) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  319) 	crypto_shash_digest(shash, raw_disk_sb + BTRFS_CSUM_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  320) 			    BTRFS_SUPER_INFO_SIZE - BTRFS_CSUM_SIZE, result);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  321) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  322) 	if (memcmp(disk_sb->csum, result, btrfs_super_csum_size(disk_sb)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  323) 		return 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  324) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  325) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  326) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  327) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  328) int btrfs_verify_level_key(struct extent_buffer *eb, int level,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  329) 			   struct btrfs_key *first_key, u64 parent_transid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  330) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  331) 	struct btrfs_fs_info *fs_info = eb->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  332) 	int found_level;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  333) 	struct btrfs_key found_key;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  334) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  335) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  336) 	found_level = btrfs_header_level(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  337) 	if (found_level != level) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  338) 		WARN(IS_ENABLED(CONFIG_BTRFS_DEBUG),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  339) 		     KERN_ERR "BTRFS: tree level check failed\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  340) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  341) "tree level mismatch detected, bytenr=%llu level expected=%u has=%u",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  342) 			  eb->start, level, found_level);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  343) 		return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  344) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  345) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  346) 	if (!first_key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  347) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  348) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  349) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  350) 	 * For live tree block (new tree blocks in current transaction),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  351) 	 * we need proper lock context to avoid race, which is impossible here.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  352) 	 * So we only checks tree blocks which is read from disk, whose
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  353) 	 * generation <= fs_info->last_trans_committed.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  354) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  355) 	if (btrfs_header_generation(eb) > fs_info->last_trans_committed)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  356) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  357) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  358) 	/* We have @first_key, so this @eb must have at least one item */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  359) 	if (btrfs_header_nritems(eb) == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  360) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  361) 		"invalid tree nritems, bytenr=%llu nritems=0 expect >0",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  362) 			  eb->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  363) 		WARN_ON(IS_ENABLED(CONFIG_BTRFS_DEBUG));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  364) 		return -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  365) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  366) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  367) 	if (found_level)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  368) 		btrfs_node_key_to_cpu(eb, &found_key, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  369) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  370) 		btrfs_item_key_to_cpu(eb, &found_key, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  371) 	ret = btrfs_comp_cpu_keys(first_key, &found_key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  372) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  373) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  374) 		WARN(IS_ENABLED(CONFIG_BTRFS_DEBUG),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  375) 		     KERN_ERR "BTRFS: tree first key check failed\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  376) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  377) "tree first key mismatch detected, bytenr=%llu parent_transid=%llu key expected=(%llu,%u,%llu) has=(%llu,%u,%llu)",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  378) 			  eb->start, parent_transid, first_key->objectid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  379) 			  first_key->type, first_key->offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  380) 			  found_key.objectid, found_key.type,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  381) 			  found_key.offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  382) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  383) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  384) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  385) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  386) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  387)  * helper to read a given tree block, doing retries as required when
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  388)  * the checksums don't match and we have alternate mirrors to try.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  389)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  390)  * @parent_transid:	expected transid, skip check if 0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  391)  * @level:		expected level, mandatory check
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  392)  * @first_key:		expected key of first slot, skip check if NULL
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  393)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  394) static int btree_read_extent_buffer_pages(struct extent_buffer *eb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  395) 					  u64 parent_transid, int level,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  396) 					  struct btrfs_key *first_key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  397) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  398) 	struct btrfs_fs_info *fs_info = eb->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  399) 	struct extent_io_tree *io_tree;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  400) 	int failed = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  401) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  402) 	int num_copies = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  403) 	int mirror_num = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  404) 	int failed_mirror = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  405) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  406) 	io_tree = &BTRFS_I(fs_info->btree_inode)->io_tree;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  407) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  408) 		clear_bit(EXTENT_BUFFER_CORRUPT, &eb->bflags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  409) 		ret = read_extent_buffer_pages(eb, WAIT_COMPLETE, mirror_num);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  410) 		if (!ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  411) 			if (verify_parent_transid(io_tree, eb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  412) 						   parent_transid, 0))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  413) 				ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  414) 			else if (btrfs_verify_level_key(eb, level,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  415) 						first_key, parent_transid))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  416) 				ret = -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  417) 			else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  418) 				break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  419) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  420) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  421) 		num_copies = btrfs_num_copies(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  422) 					      eb->start, eb->len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  423) 		if (num_copies == 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  424) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  425) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  426) 		if (!failed_mirror) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  427) 			failed = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  428) 			failed_mirror = eb->read_mirror;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  429) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  430) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  431) 		mirror_num++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  432) 		if (mirror_num == failed_mirror)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  433) 			mirror_num++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  434) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  435) 		if (mirror_num > num_copies)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  436) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  437) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  438) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  439) 	if (failed && !ret && failed_mirror)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  440) 		btrfs_repair_eb_io_failure(eb, failed_mirror);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  441) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  442) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  443) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  444) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  445) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  446)  * checksum a dirty tree block before IO.  This has extra checks to make sure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  447)  * we only fill in the checksum field in the first page of a multi-page block
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  448)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  449) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  450) static int csum_dirty_buffer(struct btrfs_fs_info *fs_info, struct page *page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  451) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  452) 	u64 start = page_offset(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  453) 	u64 found_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  454) 	u8 result[BTRFS_CSUM_SIZE];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  455) 	u16 csum_size = btrfs_super_csum_size(fs_info->super_copy);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  456) 	struct extent_buffer *eb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  457) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  458) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  459) 	eb = (struct extent_buffer *)page->private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  460) 	if (page != eb->pages[0])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  461) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  462) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  463) 	found_start = btrfs_header_bytenr(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  464) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  465) 	 * Please do not consolidate these warnings into a single if.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  466) 	 * It is useful to know what went wrong.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  467) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  468) 	if (WARN_ON(found_start != start))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  469) 		return -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  470) 	if (WARN_ON(!PageUptodate(page)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  471) 		return -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  472) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  473) 	ASSERT(memcmp_extent_buffer(eb, fs_info->fs_devices->metadata_uuid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  474) 				    offsetof(struct btrfs_header, fsid),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  475) 				    BTRFS_FSID_SIZE) == 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  476) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  477) 	csum_tree_block(eb, result);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  478) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  479) 	if (btrfs_header_level(eb))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  480) 		ret = btrfs_check_node(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  481) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  482) 		ret = btrfs_check_leaf_full(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  483) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  484) 	if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  485) 		btrfs_print_tree(eb, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  486) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  487) 		"block=%llu write time tree block corruption detected",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  488) 			  eb->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  489) 		WARN_ON(IS_ENABLED(CONFIG_BTRFS_DEBUG));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  490) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  491) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  492) 	write_extent_buffer(eb, result, 0, csum_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  493) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  494) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  495) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  496) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  497) static int check_tree_block_fsid(struct extent_buffer *eb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  498) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  499) 	struct btrfs_fs_info *fs_info = eb->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  500) 	struct btrfs_fs_devices *fs_devices = fs_info->fs_devices, *seed_devs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  501) 	u8 fsid[BTRFS_FSID_SIZE];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  502) 	u8 *metadata_uuid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  503) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  504) 	read_extent_buffer(eb, fsid, offsetof(struct btrfs_header, fsid),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  505) 			   BTRFS_FSID_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  506) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  507) 	 * Checking the incompat flag is only valid for the current fs. For
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  508) 	 * seed devices it's forbidden to have their uuid changed so reading
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  509) 	 * ->fsid in this case is fine
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  510) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  511) 	if (btrfs_fs_incompat(fs_info, METADATA_UUID))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  512) 		metadata_uuid = fs_devices->metadata_uuid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  513) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  514) 		metadata_uuid = fs_devices->fsid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  515) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  516) 	if (!memcmp(fsid, metadata_uuid, BTRFS_FSID_SIZE))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  517) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  518) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  519) 	list_for_each_entry(seed_devs, &fs_devices->seed_list, seed_list)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  520) 		if (!memcmp(fsid, seed_devs->fsid, BTRFS_FSID_SIZE))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  521) 			return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  522) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  523) 	return 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  524) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  525) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  526) int btrfs_validate_metadata_buffer(struct btrfs_io_bio *io_bio, u64 phy_offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  527) 				   struct page *page, u64 start, u64 end,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  528) 				   int mirror)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  529) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  530) 	u64 found_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  531) 	int found_level;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  532) 	struct extent_buffer *eb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  533) 	struct btrfs_fs_info *fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  534) 	u16 csum_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  535) 	int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  536) 	u8 result[BTRFS_CSUM_SIZE];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  537) 	int reads_done;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  538) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  539) 	if (!page->private)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  540) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  541) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  542) 	eb = (struct extent_buffer *)page->private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  543) 	fs_info = eb->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  544) 	csum_size = btrfs_super_csum_size(fs_info->super_copy);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  545) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  546) 	/* the pending IO might have been the only thing that kept this buffer
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  547) 	 * in memory.  Make sure we have a ref for all this other checks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  548) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  549) 	atomic_inc(&eb->refs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  550) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  551) 	reads_done = atomic_dec_and_test(&eb->io_pages);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  552) 	if (!reads_done)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  553) 		goto err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  554) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  555) 	eb->read_mirror = mirror;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  556) 	if (test_bit(EXTENT_BUFFER_READ_ERR, &eb->bflags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  557) 		ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  558) 		goto err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  559) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  560) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  561) 	found_start = btrfs_header_bytenr(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  562) 	if (found_start != eb->start) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  563) 		btrfs_err_rl(fs_info, "bad tree block start, want %llu have %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  564) 			     eb->start, found_start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  565) 		ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  566) 		goto err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  567) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  568) 	if (check_tree_block_fsid(eb)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  569) 		btrfs_err_rl(fs_info, "bad fsid on block %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  570) 			     eb->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  571) 		ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  572) 		goto err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  573) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  574) 	found_level = btrfs_header_level(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  575) 	if (found_level >= BTRFS_MAX_LEVEL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  576) 		btrfs_err(fs_info, "bad tree block level %d on %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  577) 			  (int)btrfs_header_level(eb), eb->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  578) 		ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  579) 		goto err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  580) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  581) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  582) 	btrfs_set_buffer_lockdep_class(btrfs_header_owner(eb),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  583) 				       eb, found_level);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  584) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  585) 	csum_tree_block(eb, result);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  586) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  587) 	if (memcmp_extent_buffer(eb, result, 0, csum_size)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  588) 		u8 val[BTRFS_CSUM_SIZE] = { 0 };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  589) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  590) 		read_extent_buffer(eb, &val, 0, csum_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  591) 		btrfs_warn_rl(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  592) 	"%s checksum verify failed on %llu wanted " CSUM_FMT " found " CSUM_FMT " level %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  593) 			      fs_info->sb->s_id, eb->start,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  594) 			      CSUM_FMT_VALUE(csum_size, val),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  595) 			      CSUM_FMT_VALUE(csum_size, result),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  596) 			      btrfs_header_level(eb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  597) 		ret = -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  598) 		goto err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  599) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  600) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  601) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  602) 	 * If this is a leaf block and it is corrupt, set the corrupt bit so
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  603) 	 * that we don't try and read the other copies of this block, just
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  604) 	 * return -EIO.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  605) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  606) 	if (found_level == 0 && btrfs_check_leaf_full(eb)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  607) 		set_bit(EXTENT_BUFFER_CORRUPT, &eb->bflags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  608) 		ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  609) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  610) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  611) 	if (found_level > 0 && btrfs_check_node(eb))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  612) 		ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  613) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  614) 	if (!ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  615) 		set_extent_buffer_uptodate(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  616) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  617) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  618) 			  "block=%llu read time tree block corruption detected",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  619) 			  eb->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  620) err:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  621) 	if (reads_done &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  622) 	    test_and_clear_bit(EXTENT_BUFFER_READAHEAD, &eb->bflags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  623) 		btree_readahead_hook(eb, ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  624) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  625) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  626) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  627) 		 * our io error hook is going to dec the io pages
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  628) 		 * again, we have to make sure it has something
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  629) 		 * to decrement
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  630) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  631) 		atomic_inc(&eb->io_pages);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  632) 		clear_extent_buffer_uptodate(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  633) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  634) 	free_extent_buffer(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  635) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  636) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  637) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  638) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  639) static void end_workqueue_bio(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  640) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  641) 	struct btrfs_end_io_wq *end_io_wq = bio->bi_private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  642) 	struct btrfs_fs_info *fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  643) 	struct btrfs_workqueue *wq;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  644) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  645) 	fs_info = end_io_wq->info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  646) 	end_io_wq->status = bio->bi_status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  647) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  648) 	if (bio_op(bio) == REQ_OP_WRITE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  649) 		if (end_io_wq->metadata == BTRFS_WQ_ENDIO_METADATA)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  650) 			wq = fs_info->endio_meta_write_workers;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  651) 		else if (end_io_wq->metadata == BTRFS_WQ_ENDIO_FREE_SPACE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  652) 			wq = fs_info->endio_freespace_worker;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  653) 		else if (end_io_wq->metadata == BTRFS_WQ_ENDIO_RAID56)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  654) 			wq = fs_info->endio_raid56_workers;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  655) 		else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  656) 			wq = fs_info->endio_write_workers;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  657) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  658) 		if (end_io_wq->metadata == BTRFS_WQ_ENDIO_RAID56)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  659) 			wq = fs_info->endio_raid56_workers;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  660) 		else if (end_io_wq->metadata)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  661) 			wq = fs_info->endio_meta_workers;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  662) 		else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  663) 			wq = fs_info->endio_workers;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  664) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  665) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  666) 	btrfs_init_work(&end_io_wq->work, end_workqueue_fn, NULL, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  667) 	btrfs_queue_work(wq, &end_io_wq->work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  668) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  669) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  670) blk_status_t btrfs_bio_wq_end_io(struct btrfs_fs_info *info, struct bio *bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  671) 			enum btrfs_wq_endio_type metadata)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  672) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  673) 	struct btrfs_end_io_wq *end_io_wq;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  674) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  675) 	end_io_wq = kmem_cache_alloc(btrfs_end_io_wq_cache, GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  676) 	if (!end_io_wq)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  677) 		return BLK_STS_RESOURCE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  678) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  679) 	end_io_wq->private = bio->bi_private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  680) 	end_io_wq->end_io = bio->bi_end_io;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  681) 	end_io_wq->info = info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  682) 	end_io_wq->status = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  683) 	end_io_wq->bio = bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  684) 	end_io_wq->metadata = metadata;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  685) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  686) 	bio->bi_private = end_io_wq;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  687) 	bio->bi_end_io = end_workqueue_bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  688) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  689) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  690) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  691) static void run_one_async_start(struct btrfs_work *work)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  692) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  693) 	struct async_submit_bio *async;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  694) 	blk_status_t ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  695) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  696) 	async = container_of(work, struct  async_submit_bio, work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  697) 	ret = async->submit_bio_start(async->private_data, async->bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  698) 				      async->bio_offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  699) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  700) 		async->status = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  701) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  702) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  703) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  704)  * In order to insert checksums into the metadata in large chunks, we wait
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  705)  * until bio submission time.   All the pages in the bio are checksummed and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  706)  * sums are attached onto the ordered extent record.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  707)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  708)  * At IO completion time the csums attached on the ordered extent record are
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  709)  * inserted into the tree.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  710)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  711) static void run_one_async_done(struct btrfs_work *work)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  712) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  713) 	struct async_submit_bio *async;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  714) 	struct inode *inode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  715) 	blk_status_t ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  716) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  717) 	async = container_of(work, struct  async_submit_bio, work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  718) 	inode = async->private_data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  719) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  720) 	/* If an error occurred we just want to clean up the bio and move on */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  721) 	if (async->status) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  722) 		async->bio->bi_status = async->status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  723) 		bio_endio(async->bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  724) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  725) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  726) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  727) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  728) 	 * All of the bios that pass through here are from async helpers.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  729) 	 * Use REQ_CGROUP_PUNT to issue them from the owning cgroup's context.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  730) 	 * This changes nothing when cgroups aren't in use.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  731) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  732) 	async->bio->bi_opf |= REQ_CGROUP_PUNT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  733) 	ret = btrfs_map_bio(btrfs_sb(inode->i_sb), async->bio, async->mirror_num);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  734) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  735) 		async->bio->bi_status = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  736) 		bio_endio(async->bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  737) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  738) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  739) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  740) static void run_one_async_free(struct btrfs_work *work)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  741) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  742) 	struct async_submit_bio *async;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  743) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  744) 	async = container_of(work, struct  async_submit_bio, work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  745) 	kfree(async);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  746) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  747) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  748) blk_status_t btrfs_wq_submit_bio(struct btrfs_fs_info *fs_info, struct bio *bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  749) 				 int mirror_num, unsigned long bio_flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  750) 				 u64 bio_offset, void *private_data,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  751) 				 extent_submit_bio_start_t *submit_bio_start)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  752) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  753) 	struct async_submit_bio *async;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  754) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  755) 	async = kmalloc(sizeof(*async), GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  756) 	if (!async)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  757) 		return BLK_STS_RESOURCE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  758) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  759) 	async->private_data = private_data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  760) 	async->bio = bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  761) 	async->mirror_num = mirror_num;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  762) 	async->submit_bio_start = submit_bio_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  763) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  764) 	btrfs_init_work(&async->work, run_one_async_start, run_one_async_done,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  765) 			run_one_async_free);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  766) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  767) 	async->bio_offset = bio_offset;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  768) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  769) 	async->status = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  770) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  771) 	if (op_is_sync(bio->bi_opf))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  772) 		btrfs_set_work_high_priority(&async->work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  773) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  774) 	btrfs_queue_work(fs_info->workers, &async->work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  775) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  776) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  777) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  778) static blk_status_t btree_csum_one_bio(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  779) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  780) 	struct bio_vec *bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  781) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  782) 	int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  783) 	struct bvec_iter_all iter_all;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  784) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  785) 	ASSERT(!bio_flagged(bio, BIO_CLONED));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  786) 	bio_for_each_segment_all(bvec, bio, iter_all) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  787) 		root = BTRFS_I(bvec->bv_page->mapping->host)->root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  788) 		ret = csum_dirty_buffer(root->fs_info, bvec->bv_page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  789) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  790) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  791) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  792) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  793) 	return errno_to_blk_status(ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  794) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  795) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  796) static blk_status_t btree_submit_bio_start(void *private_data, struct bio *bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  797) 					     u64 bio_offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  798) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  799) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  800) 	 * when we're called for a write, we're already in the async
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  801) 	 * submission context.  Just jump into btrfs_map_bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  802) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  803) 	return btree_csum_one_bio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  804) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  805) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  806) static int check_async_write(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  807) 			     struct btrfs_inode *bi)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  808) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  809) 	if (atomic_read(&bi->sync_writers))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  810) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  811) 	if (test_bit(BTRFS_FS_CSUM_IMPL_FAST, &fs_info->flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  812) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  813) 	return 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  814) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  815) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  816) blk_status_t btrfs_submit_metadata_bio(struct inode *inode, struct bio *bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  817) 				       int mirror_num, unsigned long bio_flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  818) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  819) 	struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  820) 	int async = check_async_write(fs_info, BTRFS_I(inode));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  821) 	blk_status_t ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  822) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  823) 	if (bio_op(bio) != REQ_OP_WRITE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  824) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  825) 		 * called for a read, do the setup so that checksum validation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  826) 		 * can happen in the async kernel threads
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  827) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  828) 		ret = btrfs_bio_wq_end_io(fs_info, bio,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  829) 					  BTRFS_WQ_ENDIO_METADATA);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  830) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  831) 			goto out_w_error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  832) 		ret = btrfs_map_bio(fs_info, bio, mirror_num);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  833) 	} else if (!async) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  834) 		ret = btree_csum_one_bio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  835) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  836) 			goto out_w_error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  837) 		ret = btrfs_map_bio(fs_info, bio, mirror_num);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  838) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  839) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  840) 		 * kthread helpers are used to submit writes so that
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  841) 		 * checksumming can happen in parallel across all CPUs
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  842) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  843) 		ret = btrfs_wq_submit_bio(fs_info, bio, mirror_num, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  844) 					  0, inode, btree_submit_bio_start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  845) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  846) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  847) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  848) 		goto out_w_error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  849) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  850) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  851) out_w_error:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  852) 	bio->bi_status = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  853) 	bio_endio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  854) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  855) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  856) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  857) #ifdef CONFIG_MIGRATION
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  858) static int btree_migratepage(struct address_space *mapping,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  859) 			struct page *newpage, struct page *page,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  860) 			enum migrate_mode mode)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  861) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  862) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  863) 	 * we can't safely write a btree page from here,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  864) 	 * we haven't done the locking hook
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  865) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  866) 	if (PageDirty(page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  867) 		return -EAGAIN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  868) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  869) 	 * Buffers may be managed in a filesystem specific way.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  870) 	 * We must have no buffers or drop them.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  871) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  872) 	if (page_has_private(page) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  873) 	    !try_to_release_page(page, GFP_KERNEL))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  874) 		return -EAGAIN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  875) 	return migrate_page(mapping, newpage, page, mode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  876) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  877) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  878) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  879) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  880) static int btree_writepages(struct address_space *mapping,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  881) 			    struct writeback_control *wbc)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  882) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  883) 	struct btrfs_fs_info *fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  884) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  885) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  886) 	if (wbc->sync_mode == WB_SYNC_NONE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  887) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  888) 		if (wbc->for_kupdate)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  889) 			return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  890) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  891) 		fs_info = BTRFS_I(mapping->host)->root->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  892) 		/* this is a bit racy, but that's ok */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  893) 		ret = __percpu_counter_compare(&fs_info->dirty_metadata_bytes,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  894) 					     BTRFS_DIRTY_METADATA_THRESH,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  895) 					     fs_info->dirty_metadata_batch);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  896) 		if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  897) 			return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  898) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  899) 	return btree_write_cache_pages(mapping, wbc);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  900) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  901) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  902) static int btree_releasepage(struct page *page, gfp_t gfp_flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  903) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  904) 	if (PageWriteback(page) || PageDirty(page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  905) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  906) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  907) 	return try_release_extent_buffer(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  908) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  909) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  910) static void btree_invalidatepage(struct page *page, unsigned int offset,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  911) 				 unsigned int length)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  912) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  913) 	struct extent_io_tree *tree;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  914) 	tree = &BTRFS_I(page->mapping->host)->io_tree;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  915) 	extent_invalidatepage(tree, page, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  916) 	btree_releasepage(page, GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  917) 	if (PagePrivate(page)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  918) 		btrfs_warn(BTRFS_I(page->mapping->host)->root->fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  919) 			   "page private not zero on page %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  920) 			   (unsigned long long)page_offset(page));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  921) 		detach_page_private(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  922) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  923) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  924) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  925) static int btree_set_page_dirty(struct page *page)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  926) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  927) #ifdef DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  928) 	struct extent_buffer *eb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  929) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  930) 	BUG_ON(!PagePrivate(page));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  931) 	eb = (struct extent_buffer *)page->private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  932) 	BUG_ON(!eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  933) 	BUG_ON(!test_bit(EXTENT_BUFFER_DIRTY, &eb->bflags));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  934) 	BUG_ON(!atomic_read(&eb->refs));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  935) 	btrfs_assert_tree_locked(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  936) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  937) 	return __set_page_dirty_nobuffers(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  938) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  939) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  940) static const struct address_space_operations btree_aops = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  941) 	.writepages	= btree_writepages,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  942) 	.releasepage	= btree_releasepage,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  943) 	.invalidatepage = btree_invalidatepage,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  944) #ifdef CONFIG_MIGRATION
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  945) 	.migratepage	= btree_migratepage,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  946) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  947) 	.set_page_dirty = btree_set_page_dirty,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  948) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  949) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  950) void readahead_tree_block(struct btrfs_fs_info *fs_info, u64 bytenr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  951) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  952) 	struct extent_buffer *buf = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  953) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  954) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  955) 	buf = btrfs_find_create_tree_block(fs_info, bytenr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  956) 	if (IS_ERR(buf))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  957) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  958) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  959) 	ret = read_extent_buffer_pages(buf, WAIT_NONE, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  960) 	if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  961) 		free_extent_buffer_stale(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  962) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  963) 		free_extent_buffer(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  964) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  965) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  966) struct extent_buffer *btrfs_find_create_tree_block(
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  967) 						struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  968) 						u64 bytenr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  969) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  970) 	if (btrfs_is_testing(fs_info))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  971) 		return alloc_test_extent_buffer(fs_info, bytenr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  972) 	return alloc_extent_buffer(fs_info, bytenr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  973) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  974) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  975) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  976)  * Read tree block at logical address @bytenr and do variant basic but critical
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  977)  * verification.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  978)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  979)  * @parent_transid:	expected transid of this tree block, skip check if 0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  980)  * @level:		expected level, mandatory check
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  981)  * @first_key:		expected key in slot 0, skip check if NULL
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  982)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  983) struct extent_buffer *read_tree_block(struct btrfs_fs_info *fs_info, u64 bytenr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  984) 				      u64 parent_transid, int level,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  985) 				      struct btrfs_key *first_key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  986) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  987) 	struct extent_buffer *buf = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  988) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  989) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  990) 	buf = btrfs_find_create_tree_block(fs_info, bytenr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  991) 	if (IS_ERR(buf))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  992) 		return buf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  993) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  994) 	ret = btree_read_extent_buffer_pages(buf, parent_transid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  995) 					     level, first_key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  996) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  997) 		free_extent_buffer_stale(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  998) 		return ERR_PTR(ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  999) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1000) 	return buf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1001) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1002) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1003) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1004) void btrfs_clean_tree_block(struct extent_buffer *buf)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1005) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1006) 	struct btrfs_fs_info *fs_info = buf->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1007) 	if (btrfs_header_generation(buf) ==
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1008) 	    fs_info->running_transaction->transid) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1009) 		btrfs_assert_tree_locked(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1010) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1011) 		if (test_and_clear_bit(EXTENT_BUFFER_DIRTY, &buf->bflags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1012) 			percpu_counter_add_batch(&fs_info->dirty_metadata_bytes,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1013) 						 -buf->len,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1014) 						 fs_info->dirty_metadata_batch);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1015) 			/* ugh, clear_extent_buffer_dirty needs to lock the page */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1016) 			btrfs_set_lock_blocking_write(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1017) 			clear_extent_buffer_dirty(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1018) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1019) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1020) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1021) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1022) static void __setup_root(struct btrfs_root *root, struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1023) 			 u64 objectid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1024) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1025) 	bool dummy = test_bit(BTRFS_FS_STATE_DUMMY_FS_INFO, &fs_info->fs_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1026) 	root->fs_info = fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1027) 	root->node = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1028) 	root->commit_root = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1029) 	root->state = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1030) 	root->orphan_cleanup_state = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1031) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1032) 	root->last_trans = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1033) 	root->highest_objectid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1034) 	root->nr_delalloc_inodes = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1035) 	root->nr_ordered_extents = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1036) 	root->inode_tree = RB_ROOT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1037) 	INIT_RADIX_TREE(&root->delayed_nodes_tree, GFP_ATOMIC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1038) 	root->block_rsv = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1039) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1040) 	INIT_LIST_HEAD(&root->dirty_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1041) 	INIT_LIST_HEAD(&root->root_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1042) 	INIT_LIST_HEAD(&root->delalloc_inodes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1043) 	INIT_LIST_HEAD(&root->delalloc_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1044) 	INIT_LIST_HEAD(&root->ordered_extents);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1045) 	INIT_LIST_HEAD(&root->ordered_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1046) 	INIT_LIST_HEAD(&root->reloc_dirty_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1047) 	INIT_LIST_HEAD(&root->logged_list[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1048) 	INIT_LIST_HEAD(&root->logged_list[1]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1049) 	spin_lock_init(&root->inode_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1050) 	spin_lock_init(&root->delalloc_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1051) 	spin_lock_init(&root->ordered_extent_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1052) 	spin_lock_init(&root->accounting_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1053) 	spin_lock_init(&root->log_extents_lock[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1054) 	spin_lock_init(&root->log_extents_lock[1]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1055) 	spin_lock_init(&root->qgroup_meta_rsv_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1056) 	mutex_init(&root->objectid_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1057) 	mutex_init(&root->log_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1058) 	mutex_init(&root->ordered_extent_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1059) 	mutex_init(&root->delalloc_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1060) 	init_waitqueue_head(&root->qgroup_flush_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1061) 	init_waitqueue_head(&root->log_writer_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1062) 	init_waitqueue_head(&root->log_commit_wait[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1063) 	init_waitqueue_head(&root->log_commit_wait[1]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1064) 	INIT_LIST_HEAD(&root->log_ctxs[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1065) 	INIT_LIST_HEAD(&root->log_ctxs[1]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1066) 	atomic_set(&root->log_commit[0], 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1067) 	atomic_set(&root->log_commit[1], 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1068) 	atomic_set(&root->log_writers, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1069) 	atomic_set(&root->log_batch, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1070) 	refcount_set(&root->refs, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1071) 	atomic_set(&root->snapshot_force_cow, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1072) 	atomic_set(&root->nr_swapfiles, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1073) 	root->log_transid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1074) 	root->log_transid_committed = -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1075) 	root->last_log_commit = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1076) 	if (!dummy) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1077) 		extent_io_tree_init(fs_info, &root->dirty_log_pages,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1078) 				    IO_TREE_ROOT_DIRTY_LOG_PAGES, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1079) 		extent_io_tree_init(fs_info, &root->log_csum_range,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1080) 				    IO_TREE_LOG_CSUM_RANGE, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1081) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1082) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1083) 	memset(&root->root_key, 0, sizeof(root->root_key));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1084) 	memset(&root->root_item, 0, sizeof(root->root_item));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1085) 	memset(&root->defrag_progress, 0, sizeof(root->defrag_progress));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1086) 	root->root_key.objectid = objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1087) 	root->anon_dev = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1088) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1089) 	spin_lock_init(&root->root_item_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1090) 	btrfs_qgroup_init_swapped_blocks(&root->swapped_blocks);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1091) #ifdef CONFIG_BTRFS_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1092) 	INIT_LIST_HEAD(&root->leak_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1093) 	spin_lock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1094) 	list_add_tail(&root->leak_list, &fs_info->allocated_roots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1095) 	spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1096) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1097) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1098) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1099) static struct btrfs_root *btrfs_alloc_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1100) 					   u64 objectid, gfp_t flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1101) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1102) 	struct btrfs_root *root = kzalloc(sizeof(*root), flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1103) 	if (root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1104) 		__setup_root(root, fs_info, objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1105) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1106) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1107) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1108) #ifdef CONFIG_BTRFS_FS_RUN_SANITY_TESTS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1109) /* Should only be used by the testing infrastructure */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1110) struct btrfs_root *btrfs_alloc_dummy_root(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1111) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1112) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1113) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1114) 	if (!fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1115) 		return ERR_PTR(-EINVAL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1116) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1117) 	root = btrfs_alloc_root(fs_info, BTRFS_ROOT_TREE_OBJECTID, GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1118) 	if (!root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1119) 		return ERR_PTR(-ENOMEM);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1120) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1121) 	/* We don't use the stripesize in selftest, set it as sectorsize */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1122) 	root->alloc_bytenr = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1123) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1124) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1125) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1126) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1127) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1128) struct btrfs_root *btrfs_create_tree(struct btrfs_trans_handle *trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1129) 				     u64 objectid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1130) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1131) 	struct btrfs_fs_info *fs_info = trans->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1132) 	struct extent_buffer *leaf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1133) 	struct btrfs_root *tree_root = fs_info->tree_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1134) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1135) 	struct btrfs_key key;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1136) 	unsigned int nofs_flag;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1137) 	int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1138) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1139) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1140) 	 * We're holding a transaction handle, so use a NOFS memory allocation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1141) 	 * context to avoid deadlock if reclaim happens.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1142) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1143) 	nofs_flag = memalloc_nofs_save();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1144) 	root = btrfs_alloc_root(fs_info, objectid, GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1145) 	memalloc_nofs_restore(nofs_flag);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1146) 	if (!root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1147) 		return ERR_PTR(-ENOMEM);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1148) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1149) 	root->root_key.objectid = objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1150) 	root->root_key.type = BTRFS_ROOT_ITEM_KEY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1151) 	root->root_key.offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1152) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1153) 	leaf = btrfs_alloc_tree_block(trans, root, 0, objectid, NULL, 0, 0, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1154) 				      BTRFS_NESTING_NORMAL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1155) 	if (IS_ERR(leaf)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1156) 		ret = PTR_ERR(leaf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1157) 		leaf = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1158) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1159) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1160) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1161) 	root->node = leaf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1162) 	btrfs_mark_buffer_dirty(leaf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1163) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1164) 	root->commit_root = btrfs_root_node(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1165) 	set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1166) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1167) 	root->root_item.flags = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1168) 	root->root_item.byte_limit = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1169) 	btrfs_set_root_bytenr(&root->root_item, leaf->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1170) 	btrfs_set_root_generation(&root->root_item, trans->transid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1171) 	btrfs_set_root_level(&root->root_item, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1172) 	btrfs_set_root_refs(&root->root_item, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1173) 	btrfs_set_root_used(&root->root_item, leaf->len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1174) 	btrfs_set_root_last_snapshot(&root->root_item, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1175) 	btrfs_set_root_dirid(&root->root_item, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1176) 	if (is_fstree(objectid))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1177) 		generate_random_guid(root->root_item.uuid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1178) 	else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1179) 		export_guid(root->root_item.uuid, &guid_null);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1180) 	root->root_item.drop_level = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1181) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1182) 	key.objectid = objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1183) 	key.type = BTRFS_ROOT_ITEM_KEY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1184) 	key.offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1185) 	ret = btrfs_insert_root(trans, tree_root, &key, &root->root_item);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1186) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1187) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1188) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1189) 	btrfs_tree_unlock(leaf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1190) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1191) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1192) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1193) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1194) 	if (leaf)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1195) 		btrfs_tree_unlock(leaf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1196) 	btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1197) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1198) 	return ERR_PTR(ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1199) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1200) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1201) static struct btrfs_root *alloc_log_tree(struct btrfs_trans_handle *trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1202) 					 struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1203) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1204) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1205) 	struct extent_buffer *leaf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1206) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1207) 	root = btrfs_alloc_root(fs_info, BTRFS_TREE_LOG_OBJECTID, GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1208) 	if (!root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1209) 		return ERR_PTR(-ENOMEM);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1210) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1211) 	root->root_key.objectid = BTRFS_TREE_LOG_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1212) 	root->root_key.type = BTRFS_ROOT_ITEM_KEY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1213) 	root->root_key.offset = BTRFS_TREE_LOG_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1214) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1215) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1216) 	 * DON'T set SHAREABLE bit for log trees.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1217) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1218) 	 * Log trees are not exposed to user space thus can't be snapshotted,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1219) 	 * and they go away before a real commit is actually done.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1220) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1221) 	 * They do store pointers to file data extents, and those reference
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1222) 	 * counts still get updated (along with back refs to the log tree).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1223) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1224) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1225) 	leaf = btrfs_alloc_tree_block(trans, root, 0, BTRFS_TREE_LOG_OBJECTID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1226) 			NULL, 0, 0, 0, BTRFS_NESTING_NORMAL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1227) 	if (IS_ERR(leaf)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1228) 		btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1229) 		return ERR_CAST(leaf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1230) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1231) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1232) 	root->node = leaf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1233) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1234) 	btrfs_mark_buffer_dirty(root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1235) 	btrfs_tree_unlock(root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1236) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1237) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1238) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1239) int btrfs_init_log_root_tree(struct btrfs_trans_handle *trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1240) 			     struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1241) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1242) 	struct btrfs_root *log_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1243) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1244) 	log_root = alloc_log_tree(trans, fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1245) 	if (IS_ERR(log_root))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1246) 		return PTR_ERR(log_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1247) 	WARN_ON(fs_info->log_root_tree);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1248) 	fs_info->log_root_tree = log_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1249) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1250) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1251) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1252) int btrfs_add_log_tree(struct btrfs_trans_handle *trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1253) 		       struct btrfs_root *root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1254) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1255) 	struct btrfs_fs_info *fs_info = root->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1256) 	struct btrfs_root *log_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1257) 	struct btrfs_inode_item *inode_item;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1258) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1259) 	log_root = alloc_log_tree(trans, fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1260) 	if (IS_ERR(log_root))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1261) 		return PTR_ERR(log_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1262) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1263) 	log_root->last_trans = trans->transid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1264) 	log_root->root_key.offset = root->root_key.objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1265) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1266) 	inode_item = &log_root->root_item.inode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1267) 	btrfs_set_stack_inode_generation(inode_item, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1268) 	btrfs_set_stack_inode_size(inode_item, 3);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1269) 	btrfs_set_stack_inode_nlink(inode_item, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1270) 	btrfs_set_stack_inode_nbytes(inode_item,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1271) 				     fs_info->nodesize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1272) 	btrfs_set_stack_inode_mode(inode_item, S_IFDIR | 0755);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1273) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1274) 	btrfs_set_root_node(&log_root->root_item, log_root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1275) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1276) 	WARN_ON(root->log_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1277) 	root->log_root = log_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1278) 	root->log_transid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1279) 	root->log_transid_committed = -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1280) 	root->last_log_commit = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1281) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1282) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1283) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1284) static struct btrfs_root *read_tree_root_path(struct btrfs_root *tree_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1285) 					      struct btrfs_path *path,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1286) 					      struct btrfs_key *key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1287) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1288) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1289) 	struct btrfs_fs_info *fs_info = tree_root->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1290) 	u64 generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1291) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1292) 	int level;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1293) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1294) 	root = btrfs_alloc_root(fs_info, key->objectid, GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1295) 	if (!root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1296) 		return ERR_PTR(-ENOMEM);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1297) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1298) 	ret = btrfs_find_root(tree_root, key, path,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1299) 			      &root->root_item, &root->root_key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1300) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1301) 		if (ret > 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1302) 			ret = -ENOENT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1303) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1304) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1305) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1306) 	generation = btrfs_root_generation(&root->root_item);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1307) 	level = btrfs_root_level(&root->root_item);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1308) 	root->node = read_tree_block(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1309) 				     btrfs_root_bytenr(&root->root_item),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1310) 				     generation, level, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1311) 	if (IS_ERR(root->node)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1312) 		ret = PTR_ERR(root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1313) 		root->node = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1314) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1315) 	} else if (!btrfs_buffer_uptodate(root->node, generation, 0)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1316) 		ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1317) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1318) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1319) 	root->commit_root = btrfs_root_node(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1320) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1321) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1322) 	btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1323) 	return ERR_PTR(ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1324) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1325) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1326) struct btrfs_root *btrfs_read_tree_root(struct btrfs_root *tree_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1327) 					struct btrfs_key *key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1328) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1329) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1330) 	struct btrfs_path *path;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1331) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1332) 	path = btrfs_alloc_path();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1333) 	if (!path)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1334) 		return ERR_PTR(-ENOMEM);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1335) 	root = read_tree_root_path(tree_root, path, key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1336) 	btrfs_free_path(path);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1337) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1338) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1339) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1340) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1341) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1342)  * Initialize subvolume root in-memory structure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1343)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1344)  * @anon_dev:	anonymous device to attach to the root, if zero, allocate new
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1345)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1346) static int btrfs_init_fs_root(struct btrfs_root *root, dev_t anon_dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1347) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1348) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1349) 	unsigned int nofs_flag;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1350) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1351) 	root->free_ino_ctl = kzalloc(sizeof(*root->free_ino_ctl), GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1352) 	root->free_ino_pinned = kzalloc(sizeof(*root->free_ino_pinned),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1353) 					GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1354) 	if (!root->free_ino_pinned || !root->free_ino_ctl) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1355) 		ret = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1356) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1357) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1358) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1359) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1360) 	 * We might be called under a transaction (e.g. indirect backref
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1361) 	 * resolution) which could deadlock if it triggers memory reclaim
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1362) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1363) 	nofs_flag = memalloc_nofs_save();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1364) 	ret = btrfs_drew_lock_init(&root->snapshot_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1365) 	memalloc_nofs_restore(nofs_flag);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1366) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1367) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1368) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1369) 	if (root->root_key.objectid != BTRFS_TREE_LOG_OBJECTID &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1370) 	    root->root_key.objectid != BTRFS_DATA_RELOC_TREE_OBJECTID) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1371) 		set_bit(BTRFS_ROOT_SHAREABLE, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1372) 		btrfs_check_and_init_root_item(&root->root_item);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1373) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1374) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1375) 	btrfs_init_free_ino_ctl(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1376) 	spin_lock_init(&root->ino_cache_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1377) 	init_waitqueue_head(&root->ino_cache_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1378) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1379) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1380) 	 * Don't assign anonymous block device to roots that are not exposed to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1381) 	 * userspace, the id pool is limited to 1M
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1382) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1383) 	if (is_fstree(root->root_key.objectid) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1384) 	    btrfs_root_refs(&root->root_item) > 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1385) 		if (!anon_dev) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1386) 			ret = get_anon_bdev(&root->anon_dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1387) 			if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1388) 				goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1389) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1390) 			root->anon_dev = anon_dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1391) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1392) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1393) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1394) 	mutex_lock(&root->objectid_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1395) 	ret = btrfs_find_highest_objectid(root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1396) 					&root->highest_objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1397) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1398) 		mutex_unlock(&root->objectid_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1399) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1400) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1401) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1402) 	ASSERT(root->highest_objectid <= BTRFS_LAST_FREE_OBJECTID);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1403) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1404) 	mutex_unlock(&root->objectid_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1405) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1406) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1407) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1408) 	/* The caller is responsible to call btrfs_free_fs_root */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1409) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1410) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1411) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1412) static struct btrfs_root *btrfs_lookup_fs_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1413) 					       u64 root_id)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1414) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1415) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1416) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1417) 	spin_lock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1418) 	root = radix_tree_lookup(&fs_info->fs_roots_radix,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1419) 				 (unsigned long)root_id);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1420) 	if (root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1421) 		root = btrfs_grab_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1422) 	spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1423) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1424) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1425) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1426) static struct btrfs_root *btrfs_get_global_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1427) 						u64 objectid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1428) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1429) 	if (objectid == BTRFS_ROOT_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1430) 		return btrfs_grab_root(fs_info->tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1431) 	if (objectid == BTRFS_EXTENT_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1432) 		return btrfs_grab_root(fs_info->extent_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1433) 	if (objectid == BTRFS_CHUNK_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1434) 		return btrfs_grab_root(fs_info->chunk_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1435) 	if (objectid == BTRFS_DEV_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1436) 		return btrfs_grab_root(fs_info->dev_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1437) 	if (objectid == BTRFS_CSUM_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1438) 		return btrfs_grab_root(fs_info->csum_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1439) 	if (objectid == BTRFS_QUOTA_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1440) 		return btrfs_grab_root(fs_info->quota_root) ?
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1441) 			fs_info->quota_root : ERR_PTR(-ENOENT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1442) 	if (objectid == BTRFS_UUID_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1443) 		return btrfs_grab_root(fs_info->uuid_root) ?
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1444) 			fs_info->uuid_root : ERR_PTR(-ENOENT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1445) 	if (objectid == BTRFS_FREE_SPACE_TREE_OBJECTID)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1446) 		return btrfs_grab_root(fs_info->free_space_root) ?
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1447) 			fs_info->free_space_root : ERR_PTR(-ENOENT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1448) 	return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1449) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1450) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1451) int btrfs_insert_fs_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1452) 			 struct btrfs_root *root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1453) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1454) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1455) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1456) 	ret = radix_tree_preload(GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1457) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1458) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1459) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1460) 	spin_lock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1461) 	ret = radix_tree_insert(&fs_info->fs_roots_radix,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1462) 				(unsigned long)root->root_key.objectid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1463) 				root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1464) 	if (ret == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1465) 		btrfs_grab_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1466) 		set_bit(BTRFS_ROOT_IN_RADIX, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1467) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1468) 	spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1469) 	radix_tree_preload_end();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1470) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1471) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1472) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1473) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1474) void btrfs_check_leaked_roots(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1475) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1476) #ifdef CONFIG_BTRFS_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1477) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1478) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1479) 	while (!list_empty(&fs_info->allocated_roots)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1480) 		char buf[BTRFS_ROOT_NAME_BUF_LEN];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1481) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1482) 		root = list_first_entry(&fs_info->allocated_roots,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1483) 					struct btrfs_root, leak_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1484) 		btrfs_err(fs_info, "leaked root %s refcount %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1485) 			  btrfs_root_name(&root->root_key, buf),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1486) 			  refcount_read(&root->refs));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1487) 		while (refcount_read(&root->refs) > 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1488) 			btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1489) 		btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1490) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1491) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1492) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1493) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1494) void btrfs_free_fs_info(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1495) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1496) 	percpu_counter_destroy(&fs_info->dirty_metadata_bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1497) 	percpu_counter_destroy(&fs_info->delalloc_bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1498) 	percpu_counter_destroy(&fs_info->dio_bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1499) 	percpu_counter_destroy(&fs_info->dev_replace.bio_counter);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1500) 	btrfs_free_csum_hash(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1501) 	btrfs_free_stripe_hash_table(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1502) 	btrfs_free_ref_cache(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1503) 	kfree(fs_info->balance_ctl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1504) 	kfree(fs_info->delayed_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1505) 	btrfs_put_root(fs_info->extent_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1506) 	btrfs_put_root(fs_info->tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1507) 	btrfs_put_root(fs_info->chunk_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1508) 	btrfs_put_root(fs_info->dev_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1509) 	btrfs_put_root(fs_info->csum_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1510) 	btrfs_put_root(fs_info->quota_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1511) 	btrfs_put_root(fs_info->uuid_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1512) 	btrfs_put_root(fs_info->free_space_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1513) 	btrfs_put_root(fs_info->fs_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1514) 	btrfs_put_root(fs_info->data_reloc_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1515) 	btrfs_check_leaked_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1516) 	btrfs_extent_buffer_leak_debug_check(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1517) 	kfree(fs_info->super_copy);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1518) 	kfree(fs_info->super_for_commit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1519) 	kvfree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1520) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1521) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1522) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1523) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1524)  * Get an in-memory reference of a root structure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1525)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1526)  * For essential trees like root/extent tree, we grab it from fs_info directly.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1527)  * For subvolume trees, we check the cached filesystem roots first. If not
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1528)  * found, then read it from disk and add it to cached fs roots.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1529)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1530)  * Caller should release the root by calling btrfs_put_root() after the usage.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1531)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1532)  * NOTE: Reloc and log trees can't be read by this function as they share the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1533)  *	 same root objectid.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1534)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1535)  * @objectid:	root id
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1536)  * @anon_dev:	preallocated anonymous block device number for new roots,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1537)  * 		pass 0 for new allocation.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1538)  * @check_ref:	whether to check root item references, If true, return -ENOENT
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1539)  *		for orphan roots
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1540)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1541) static struct btrfs_root *btrfs_get_root_ref(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1542) 					     u64 objectid, dev_t anon_dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1543) 					     bool check_ref)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1544) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1545) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1546) 	struct btrfs_path *path;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1547) 	struct btrfs_key key;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1548) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1549) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1550) 	root = btrfs_get_global_root(fs_info, objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1551) 	if (root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1552) 		return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1553) again:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1554) 	root = btrfs_lookup_fs_root(fs_info, objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1555) 	if (root) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1556) 		/* Shouldn't get preallocated anon_dev for cached roots */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1557) 		ASSERT(!anon_dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1558) 		if (check_ref && btrfs_root_refs(&root->root_item) == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1559) 			btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1560) 			return ERR_PTR(-ENOENT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1561) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1562) 		return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1563) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1564) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1565) 	key.objectid = objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1566) 	key.type = BTRFS_ROOT_ITEM_KEY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1567) 	key.offset = (u64)-1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1568) 	root = btrfs_read_tree_root(fs_info->tree_root, &key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1569) 	if (IS_ERR(root))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1570) 		return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1571) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1572) 	if (check_ref && btrfs_root_refs(&root->root_item) == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1573) 		ret = -ENOENT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1574) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1575) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1576) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1577) 	ret = btrfs_init_fs_root(root, anon_dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1578) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1579) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1580) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1581) 	path = btrfs_alloc_path();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1582) 	if (!path) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1583) 		ret = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1584) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1585) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1586) 	key.objectid = BTRFS_ORPHAN_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1587) 	key.type = BTRFS_ORPHAN_ITEM_KEY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1588) 	key.offset = objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1589) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1590) 	ret = btrfs_search_slot(NULL, fs_info->tree_root, &key, path, 0, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1591) 	btrfs_free_path(path);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1592) 	if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1593) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1594) 	if (ret == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1595) 		set_bit(BTRFS_ROOT_ORPHAN_ITEM_INSERTED, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1596) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1597) 	ret = btrfs_insert_fs_root(fs_info, root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1598) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1599) 		btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1600) 		if (ret == -EEXIST)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1601) 			goto again;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1602) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1603) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1604) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1605) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1606) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1607) 	 * If our caller provided us an anonymous device, then it's his
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1608) 	 * responsability to free it in case we fail. So we have to set our
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1609) 	 * root's anon_dev to 0 to avoid a double free, once by btrfs_put_root()
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1610) 	 * and once again by our caller.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1611) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1612) 	if (anon_dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1613) 		root->anon_dev = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1614) 	btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1615) 	return ERR_PTR(ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1616) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1617) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1618) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1619)  * Get in-memory reference of a root structure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1620)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1621)  * @objectid:	tree objectid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1622)  * @check_ref:	if set, verify that the tree exists and the item has at least
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1623)  *		one reference
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1624)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1625) struct btrfs_root *btrfs_get_fs_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1626) 				     u64 objectid, bool check_ref)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1627) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1628) 	return btrfs_get_root_ref(fs_info, objectid, 0, check_ref);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1629) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1630) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1631) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1632)  * Get in-memory reference of a root structure, created as new, optionally pass
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1633)  * the anonymous block device id
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1634)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1635)  * @objectid:	tree objectid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1636)  * @anon_dev:	if zero, allocate a new anonymous block device or use the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1637)  *		parameter value
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1638)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1639) struct btrfs_root *btrfs_get_new_fs_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1640) 					 u64 objectid, dev_t anon_dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1641) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1642) 	return btrfs_get_root_ref(fs_info, objectid, anon_dev, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1643) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1644) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1645) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1646)  * btrfs_get_fs_root_commit_root - return a root for the given objectid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1647)  * @fs_info:	the fs_info
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1648)  * @objectid:	the objectid we need to lookup
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1649)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1650)  * This is exclusively used for backref walking, and exists specifically because
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1651)  * of how qgroups does lookups.  Qgroups will do a backref lookup at delayed ref
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1652)  * creation time, which means we may have to read the tree_root in order to look
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1653)  * up a fs root that is not in memory.  If the root is not in memory we will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1654)  * read the tree root commit root and look up the fs root from there.  This is a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1655)  * temporary root, it will not be inserted into the radix tree as it doesn't
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1656)  * have the most uptodate information, it'll simply be discarded once the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1657)  * backref code is finished using the root.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1658)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1659) struct btrfs_root *btrfs_get_fs_root_commit_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1660) 						 struct btrfs_path *path,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1661) 						 u64 objectid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1662) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1663) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1664) 	struct btrfs_key key;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1665) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1666) 	ASSERT(path->search_commit_root && path->skip_locking);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1667) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1668) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1669) 	 * This can return -ENOENT if we ask for a root that doesn't exist, but
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1670) 	 * since this is called via the backref walking code we won't be looking
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1671) 	 * up a root that doesn't exist, unless there's corruption.  So if root
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1672) 	 * != NULL just return it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1673) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1674) 	root = btrfs_get_global_root(fs_info, objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1675) 	if (root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1676) 		return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1677) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1678) 	root = btrfs_lookup_fs_root(fs_info, objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1679) 	if (root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1680) 		return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1681) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1682) 	key.objectid = objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1683) 	key.type = BTRFS_ROOT_ITEM_KEY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1684) 	key.offset = (u64)-1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1685) 	root = read_tree_root_path(fs_info->tree_root, path, &key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1686) 	btrfs_release_path(path);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1687) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1688) 	return root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1689) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1690) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1691) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1692)  * called by the kthread helper functions to finally call the bio end_io
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1693)  * functions.  This is where read checksum verification actually happens
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1694)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1695) static void end_workqueue_fn(struct btrfs_work *work)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1696) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1697) 	struct bio *bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1698) 	struct btrfs_end_io_wq *end_io_wq;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1699) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1700) 	end_io_wq = container_of(work, struct btrfs_end_io_wq, work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1701) 	bio = end_io_wq->bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1702) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1703) 	bio->bi_status = end_io_wq->status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1704) 	bio->bi_private = end_io_wq->private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1705) 	bio->bi_end_io = end_io_wq->end_io;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1706) 	bio_endio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1707) 	kmem_cache_free(btrfs_end_io_wq_cache, end_io_wq);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1708) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1709) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1710) static int cleaner_kthread(void *arg)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1711) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1712) 	struct btrfs_root *root = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1713) 	struct btrfs_fs_info *fs_info = root->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1714) 	int again;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1715) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1716) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1717) 		again = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1718) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1719) 		set_bit(BTRFS_FS_CLEANER_RUNNING, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1720) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1721) 		/* Make the cleaner go to sleep early. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1722) 		if (btrfs_need_cleaner_sleep(fs_info))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1723) 			goto sleep;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1724) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1725) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1726) 		 * Do not do anything if we might cause open_ctree() to block
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1727) 		 * before we have finished mounting the filesystem.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1728) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1729) 		if (!test_bit(BTRFS_FS_OPEN, &fs_info->flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1730) 			goto sleep;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1731) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1732) 		if (!mutex_trylock(&fs_info->cleaner_mutex))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1733) 			goto sleep;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1734) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1735) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1736) 		 * Avoid the problem that we change the status of the fs
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1737) 		 * during the above check and trylock.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1738) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1739) 		if (btrfs_need_cleaner_sleep(fs_info)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1740) 			mutex_unlock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1741) 			goto sleep;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1742) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1743) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1744) 		btrfs_run_delayed_iputs(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1745) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1746) 		again = btrfs_clean_one_deleted_snapshot(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1747) 		mutex_unlock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1748) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1749) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1750) 		 * The defragger has dealt with the R/O remount and umount,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1751) 		 * needn't do anything special here.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1752) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1753) 		btrfs_run_defrag_inodes(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1754) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1755) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1756) 		 * Acquires fs_info->delete_unused_bgs_mutex to avoid racing
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1757) 		 * with relocation (btrfs_relocate_chunk) and relocation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1758) 		 * acquires fs_info->cleaner_mutex (btrfs_relocate_block_group)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1759) 		 * after acquiring fs_info->delete_unused_bgs_mutex. So we
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1760) 		 * can't hold, nor need to, fs_info->cleaner_mutex when deleting
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1761) 		 * unused block groups.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1762) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1763) 		btrfs_delete_unused_bgs(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1764) sleep:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1765) 		clear_bit(BTRFS_FS_CLEANER_RUNNING, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1766) 		if (kthread_should_park())
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1767) 			kthread_parkme();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1768) 		if (kthread_should_stop())
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1769) 			return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1770) 		if (!again) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1771) 			set_current_state(TASK_INTERRUPTIBLE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1772) 			schedule();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1773) 			__set_current_state(TASK_RUNNING);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1774) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1775) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1776) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1777) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1778) static int transaction_kthread(void *arg)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1779) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1780) 	struct btrfs_root *root = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1781) 	struct btrfs_fs_info *fs_info = root->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1782) 	struct btrfs_trans_handle *trans;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1783) 	struct btrfs_transaction *cur;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1784) 	u64 transid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1785) 	time64_t now;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1786) 	unsigned long delay;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1787) 	bool cannot_commit;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1788) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1789) 	do {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1790) 		cannot_commit = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1791) 		delay = HZ * fs_info->commit_interval;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1792) 		mutex_lock(&fs_info->transaction_kthread_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1793) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1794) 		spin_lock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1795) 		cur = fs_info->running_transaction;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1796) 		if (!cur) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1797) 			spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1798) 			goto sleep;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1799) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1800) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1801) 		now = ktime_get_seconds();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1802) 		if (cur->state < TRANS_STATE_COMMIT_START &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1803) 		    (now < cur->start_time ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1804) 		     now - cur->start_time < fs_info->commit_interval)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1805) 			spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1806) 			delay = HZ * 5;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1807) 			goto sleep;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1808) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1809) 		transid = cur->transid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1810) 		spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1811) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1812) 		/* If the file system is aborted, this will always fail. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1813) 		trans = btrfs_attach_transaction(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1814) 		if (IS_ERR(trans)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1815) 			if (PTR_ERR(trans) != -ENOENT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1816) 				cannot_commit = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1817) 			goto sleep;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1818) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1819) 		if (transid == trans->transid) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1820) 			btrfs_commit_transaction(trans);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1821) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1822) 			btrfs_end_transaction(trans);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1823) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1824) sleep:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1825) 		wake_up_process(fs_info->cleaner_kthread);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1826) 		mutex_unlock(&fs_info->transaction_kthread_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1827) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1828) 		if (unlikely(test_bit(BTRFS_FS_STATE_ERROR,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1829) 				      &fs_info->fs_state)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1830) 			btrfs_cleanup_transaction(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1831) 		if (!kthread_should_stop() &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1832) 				(!btrfs_transaction_blocked(fs_info) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1833) 				 cannot_commit))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1834) 			schedule_timeout_interruptible(delay);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1835) 	} while (!kthread_should_stop());
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1836) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1837) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1838) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1839) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1840)  * This will find the highest generation in the array of root backups.  The
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1841)  * index of the highest array is returned, or -EINVAL if we can't find
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1842)  * anything.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1843)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1844)  * We check to make sure the array is valid by comparing the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1845)  * generation of the latest  root in the array with the generation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1846)  * in the super block.  If they don't match we pitch it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1847)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1848) static int find_newest_super_backup(struct btrfs_fs_info *info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1849) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1850) 	const u64 newest_gen = btrfs_super_generation(info->super_copy);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1851) 	u64 cur;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1852) 	struct btrfs_root_backup *root_backup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1853) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1854) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1855) 	for (i = 0; i < BTRFS_NUM_BACKUP_ROOTS; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1856) 		root_backup = info->super_copy->super_roots + i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1857) 		cur = btrfs_backup_tree_root_gen(root_backup);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1858) 		if (cur == newest_gen)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1859) 			return i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1860) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1861) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1862) 	return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1863) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1864) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1865) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1866)  * copy all the root pointers into the super backup array.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1867)  * this will bump the backup pointer by one when it is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1868)  * done
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1869)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1870) static void backup_super_roots(struct btrfs_fs_info *info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1871) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1872) 	const int next_backup = info->backup_root_index;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1873) 	struct btrfs_root_backup *root_backup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1874) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1875) 	root_backup = info->super_for_commit->super_roots + next_backup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1876) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1877) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1878) 	 * make sure all of our padding and empty slots get zero filled
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1879) 	 * regardless of which ones we use today
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1880) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1881) 	memset(root_backup, 0, sizeof(*root_backup));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1882) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1883) 	info->backup_root_index = (next_backup + 1) % BTRFS_NUM_BACKUP_ROOTS;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1884) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1885) 	btrfs_set_backup_tree_root(root_backup, info->tree_root->node->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1886) 	btrfs_set_backup_tree_root_gen(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1887) 			       btrfs_header_generation(info->tree_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1888) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1889) 	btrfs_set_backup_tree_root_level(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1890) 			       btrfs_header_level(info->tree_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1891) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1892) 	btrfs_set_backup_chunk_root(root_backup, info->chunk_root->node->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1893) 	btrfs_set_backup_chunk_root_gen(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1894) 			       btrfs_header_generation(info->chunk_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1895) 	btrfs_set_backup_chunk_root_level(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1896) 			       btrfs_header_level(info->chunk_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1897) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1898) 	btrfs_set_backup_extent_root(root_backup, info->extent_root->node->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1899) 	btrfs_set_backup_extent_root_gen(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1900) 			       btrfs_header_generation(info->extent_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1901) 	btrfs_set_backup_extent_root_level(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1902) 			       btrfs_header_level(info->extent_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1903) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1904) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1905) 	 * we might commit during log recovery, which happens before we set
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1906) 	 * the fs_root.  Make sure it is valid before we fill it in.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1907) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1908) 	if (info->fs_root && info->fs_root->node) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1909) 		btrfs_set_backup_fs_root(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1910) 					 info->fs_root->node->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1911) 		btrfs_set_backup_fs_root_gen(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1912) 			       btrfs_header_generation(info->fs_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1913) 		btrfs_set_backup_fs_root_level(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1914) 			       btrfs_header_level(info->fs_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1915) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1916) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1917) 	btrfs_set_backup_dev_root(root_backup, info->dev_root->node->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1918) 	btrfs_set_backup_dev_root_gen(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1919) 			       btrfs_header_generation(info->dev_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1920) 	btrfs_set_backup_dev_root_level(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1921) 				       btrfs_header_level(info->dev_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1922) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1923) 	btrfs_set_backup_csum_root(root_backup, info->csum_root->node->start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1924) 	btrfs_set_backup_csum_root_gen(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1925) 			       btrfs_header_generation(info->csum_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1926) 	btrfs_set_backup_csum_root_level(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1927) 			       btrfs_header_level(info->csum_root->node));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1928) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1929) 	btrfs_set_backup_total_bytes(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1930) 			     btrfs_super_total_bytes(info->super_copy));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1931) 	btrfs_set_backup_bytes_used(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1932) 			     btrfs_super_bytes_used(info->super_copy));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1933) 	btrfs_set_backup_num_devices(root_backup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1934) 			     btrfs_super_num_devices(info->super_copy));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1935) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1936) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1937) 	 * if we don't copy this out to the super_copy, it won't get remembered
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1938) 	 * for the next commit
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1939) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1940) 	memcpy(&info->super_copy->super_roots,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1941) 	       &info->super_for_commit->super_roots,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1942) 	       sizeof(*root_backup) * BTRFS_NUM_BACKUP_ROOTS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1943) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1944) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1945) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1946)  * read_backup_root - Reads a backup root based on the passed priority. Prio 0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1947)  * is the newest, prio 1/2/3 are 2nd newest/3rd newest/4th (oldest) backup roots
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1948)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1949)  * fs_info - filesystem whose backup roots need to be read
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1950)  * priority - priority of backup root required
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1951)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1952)  * Returns backup root index on success and -EINVAL otherwise.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1953)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1954) static int read_backup_root(struct btrfs_fs_info *fs_info, u8 priority)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1955) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1956) 	int backup_index = find_newest_super_backup(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1957) 	struct btrfs_super_block *super = fs_info->super_copy;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1958) 	struct btrfs_root_backup *root_backup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1959) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1960) 	if (priority < BTRFS_NUM_BACKUP_ROOTS && backup_index >= 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1961) 		if (priority == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1962) 			return backup_index;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1963) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1964) 		backup_index = backup_index + BTRFS_NUM_BACKUP_ROOTS - priority;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1965) 		backup_index %= BTRFS_NUM_BACKUP_ROOTS;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1966) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1967) 		return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1968) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1969) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1970) 	root_backup = super->super_roots + backup_index;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1971) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1972) 	btrfs_set_super_generation(super,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1973) 				   btrfs_backup_tree_root_gen(root_backup));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1974) 	btrfs_set_super_root(super, btrfs_backup_tree_root(root_backup));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1975) 	btrfs_set_super_root_level(super,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1976) 				   btrfs_backup_tree_root_level(root_backup));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1977) 	btrfs_set_super_bytes_used(super, btrfs_backup_bytes_used(root_backup));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1978) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1979) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1980) 	 * Fixme: the total bytes and num_devices need to match or we should
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1981) 	 * need a fsck
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1982) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1983) 	btrfs_set_super_total_bytes(super, btrfs_backup_total_bytes(root_backup));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1984) 	btrfs_set_super_num_devices(super, btrfs_backup_num_devices(root_backup));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1985) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1986) 	return backup_index;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1987) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1988) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1989) /* helper to cleanup workers */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1990) static void btrfs_stop_all_workers(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1991) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1992) 	btrfs_destroy_workqueue(fs_info->fixup_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1993) 	btrfs_destroy_workqueue(fs_info->delalloc_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1994) 	btrfs_destroy_workqueue(fs_info->workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1995) 	btrfs_destroy_workqueue(fs_info->endio_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1996) 	btrfs_destroy_workqueue(fs_info->endio_raid56_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1997) 	btrfs_destroy_workqueue(fs_info->rmw_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1998) 	btrfs_destroy_workqueue(fs_info->endio_write_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1999) 	btrfs_destroy_workqueue(fs_info->endio_freespace_worker);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2000) 	btrfs_destroy_workqueue(fs_info->delayed_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2001) 	btrfs_destroy_workqueue(fs_info->caching_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2002) 	btrfs_destroy_workqueue(fs_info->readahead_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2003) 	btrfs_destroy_workqueue(fs_info->flush_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2004) 	btrfs_destroy_workqueue(fs_info->qgroup_rescan_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2005) 	if (fs_info->discard_ctl.discard_workers)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2006) 		destroy_workqueue(fs_info->discard_ctl.discard_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2007) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2008) 	 * Now that all other work queues are destroyed, we can safely destroy
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2009) 	 * the queues used for metadata I/O, since tasks from those other work
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2010) 	 * queues can do metadata I/O operations.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2011) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2012) 	btrfs_destroy_workqueue(fs_info->endio_meta_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2013) 	btrfs_destroy_workqueue(fs_info->endio_meta_write_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2014) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2015) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2016) static void free_root_extent_buffers(struct btrfs_root *root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2017) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2018) 	if (root) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2019) 		free_extent_buffer(root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2020) 		free_extent_buffer(root->commit_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2021) 		root->node = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2022) 		root->commit_root = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2023) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2024) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2025) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2026) /* helper to cleanup tree roots */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2027) static void free_root_pointers(struct btrfs_fs_info *info, bool free_chunk_root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2028) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2029) 	free_root_extent_buffers(info->tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2030) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2031) 	free_root_extent_buffers(info->dev_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2032) 	free_root_extent_buffers(info->extent_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2033) 	free_root_extent_buffers(info->csum_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2034) 	free_root_extent_buffers(info->quota_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2035) 	free_root_extent_buffers(info->uuid_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2036) 	free_root_extent_buffers(info->fs_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2037) 	free_root_extent_buffers(info->data_reloc_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2038) 	if (free_chunk_root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2039) 		free_root_extent_buffers(info->chunk_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2040) 	free_root_extent_buffers(info->free_space_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2041) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2042) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2043) void btrfs_put_root(struct btrfs_root *root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2044) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2045) 	if (!root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2046) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2047) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2048) 	if (refcount_dec_and_test(&root->refs)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2049) 		WARN_ON(!RB_EMPTY_ROOT(&root->inode_tree));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2050) 		WARN_ON(test_bit(BTRFS_ROOT_DEAD_RELOC_TREE, &root->state));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2051) 		if (root->anon_dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2052) 			free_anon_bdev(root->anon_dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2053) 		btrfs_drew_lock_destroy(&root->snapshot_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2054) 		free_root_extent_buffers(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2055) 		kfree(root->free_ino_ctl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2056) 		kfree(root->free_ino_pinned);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2057) #ifdef CONFIG_BTRFS_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2058) 		spin_lock(&root->fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2059) 		list_del_init(&root->leak_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2060) 		spin_unlock(&root->fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2061) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2062) 		kfree(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2063) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2064) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2065) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2066) void btrfs_free_fs_roots(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2067) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2068) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2069) 	struct btrfs_root *gang[8];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2070) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2071) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2072) 	while (!list_empty(&fs_info->dead_roots)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2073) 		gang[0] = list_entry(fs_info->dead_roots.next,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2074) 				     struct btrfs_root, root_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2075) 		list_del(&gang[0]->root_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2076) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2077) 		if (test_bit(BTRFS_ROOT_IN_RADIX, &gang[0]->state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2078) 			btrfs_drop_and_free_fs_root(fs_info, gang[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2079) 		btrfs_put_root(gang[0]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2080) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2081) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2082) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2083) 		ret = radix_tree_gang_lookup(&fs_info->fs_roots_radix,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2084) 					     (void **)gang, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2085) 					     ARRAY_SIZE(gang));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2086) 		if (!ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2087) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2088) 		for (i = 0; i < ret; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2089) 			btrfs_drop_and_free_fs_root(fs_info, gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2090) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2091) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2092) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2093) static void btrfs_init_scrub(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2094) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2095) 	mutex_init(&fs_info->scrub_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2096) 	atomic_set(&fs_info->scrubs_running, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2097) 	atomic_set(&fs_info->scrub_pause_req, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2098) 	atomic_set(&fs_info->scrubs_paused, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2099) 	atomic_set(&fs_info->scrub_cancel_req, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2100) 	init_waitqueue_head(&fs_info->scrub_pause_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2101) 	refcount_set(&fs_info->scrub_workers_refcnt, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2102) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2103) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2104) static void btrfs_init_balance(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2105) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2106) 	spin_lock_init(&fs_info->balance_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2107) 	mutex_init(&fs_info->balance_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2108) 	atomic_set(&fs_info->balance_pause_req, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2109) 	atomic_set(&fs_info->balance_cancel_req, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2110) 	fs_info->balance_ctl = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2111) 	init_waitqueue_head(&fs_info->balance_wait_q);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2112) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2113) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2114) static void btrfs_init_btree_inode(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2115) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2116) 	struct inode *inode = fs_info->btree_inode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2117) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2118) 	inode->i_ino = BTRFS_BTREE_INODE_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2119) 	set_nlink(inode, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2120) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2121) 	 * we set the i_size on the btree inode to the max possible int.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2122) 	 * the real end of the address space is determined by all of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2123) 	 * the devices in the system
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2124) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2125) 	inode->i_size = OFFSET_MAX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2126) 	inode->i_mapping->a_ops = &btree_aops;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2127) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2128) 	RB_CLEAR_NODE(&BTRFS_I(inode)->rb_node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2129) 	extent_io_tree_init(fs_info, &BTRFS_I(inode)->io_tree,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2130) 			    IO_TREE_BTREE_INODE_IO, inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2131) 	BTRFS_I(inode)->io_tree.track_uptodate = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2132) 	extent_map_tree_init(&BTRFS_I(inode)->extent_tree);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2133) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2134) 	BTRFS_I(inode)->root = btrfs_grab_root(fs_info->tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2135) 	memset(&BTRFS_I(inode)->location, 0, sizeof(struct btrfs_key));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2136) 	set_bit(BTRFS_INODE_DUMMY, &BTRFS_I(inode)->runtime_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2137) 	btrfs_insert_inode_hash(inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2138) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2139) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2140) static void btrfs_init_dev_replace_locks(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2141) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2142) 	mutex_init(&fs_info->dev_replace.lock_finishing_cancel_unmount);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2143) 	init_rwsem(&fs_info->dev_replace.rwsem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2144) 	init_waitqueue_head(&fs_info->dev_replace.replace_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2145) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2146) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2147) static void btrfs_init_qgroup(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2148) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2149) 	spin_lock_init(&fs_info->qgroup_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2150) 	mutex_init(&fs_info->qgroup_ioctl_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2151) 	fs_info->qgroup_tree = RB_ROOT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2152) 	INIT_LIST_HEAD(&fs_info->dirty_qgroups);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2153) 	fs_info->qgroup_seq = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2154) 	fs_info->qgroup_ulist = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2155) 	fs_info->qgroup_rescan_running = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2156) 	mutex_init(&fs_info->qgroup_rescan_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2157) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2158) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2159) static int btrfs_init_workqueues(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2160) 		struct btrfs_fs_devices *fs_devices)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2161) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2162) 	u32 max_active = fs_info->thread_pool_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2163) 	unsigned int flags = WQ_MEM_RECLAIM | WQ_FREEZABLE | WQ_UNBOUND;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2164) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2165) 	fs_info->workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2166) 		btrfs_alloc_workqueue(fs_info, "worker",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2167) 				      flags | WQ_HIGHPRI, max_active, 16);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2168) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2169) 	fs_info->delalloc_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2170) 		btrfs_alloc_workqueue(fs_info, "delalloc",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2171) 				      flags, max_active, 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2172) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2173) 	fs_info->flush_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2174) 		btrfs_alloc_workqueue(fs_info, "flush_delalloc",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2175) 				      flags, max_active, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2176) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2177) 	fs_info->caching_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2178) 		btrfs_alloc_workqueue(fs_info, "cache", flags, max_active, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2179) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2180) 	fs_info->fixup_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2181) 		btrfs_alloc_workqueue(fs_info, "fixup", flags, 1, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2182) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2183) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2184) 	 * endios are largely parallel and should have a very
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2185) 	 * low idle thresh
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2186) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2187) 	fs_info->endio_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2188) 		btrfs_alloc_workqueue(fs_info, "endio", flags, max_active, 4);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2189) 	fs_info->endio_meta_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2190) 		btrfs_alloc_workqueue(fs_info, "endio-meta", flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2191) 				      max_active, 4);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2192) 	fs_info->endio_meta_write_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2193) 		btrfs_alloc_workqueue(fs_info, "endio-meta-write", flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2194) 				      max_active, 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2195) 	fs_info->endio_raid56_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2196) 		btrfs_alloc_workqueue(fs_info, "endio-raid56", flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2197) 				      max_active, 4);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2198) 	fs_info->rmw_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2199) 		btrfs_alloc_workqueue(fs_info, "rmw", flags, max_active, 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2200) 	fs_info->endio_write_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2201) 		btrfs_alloc_workqueue(fs_info, "endio-write", flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2202) 				      max_active, 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2203) 	fs_info->endio_freespace_worker =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2204) 		btrfs_alloc_workqueue(fs_info, "freespace-write", flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2205) 				      max_active, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2206) 	fs_info->delayed_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2207) 		btrfs_alloc_workqueue(fs_info, "delayed-meta", flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2208) 				      max_active, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2209) 	fs_info->readahead_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2210) 		btrfs_alloc_workqueue(fs_info, "readahead", flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2211) 				      max_active, 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2212) 	fs_info->qgroup_rescan_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2213) 		btrfs_alloc_workqueue(fs_info, "qgroup-rescan", flags, 1, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2214) 	fs_info->discard_ctl.discard_workers =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2215) 		alloc_workqueue("btrfs_discard", WQ_UNBOUND | WQ_FREEZABLE, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2216) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2217) 	if (!(fs_info->workers && fs_info->delalloc_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2218) 	      fs_info->flush_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2219) 	      fs_info->endio_workers && fs_info->endio_meta_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2220) 	      fs_info->endio_meta_write_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2221) 	      fs_info->endio_write_workers && fs_info->endio_raid56_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2222) 	      fs_info->endio_freespace_worker && fs_info->rmw_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2223) 	      fs_info->caching_workers && fs_info->readahead_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2224) 	      fs_info->fixup_workers && fs_info->delayed_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2225) 	      fs_info->qgroup_rescan_workers &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2226) 	      fs_info->discard_ctl.discard_workers)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2227) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2228) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2229) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2230) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2231) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2232) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2233) static int btrfs_init_csum_hash(struct btrfs_fs_info *fs_info, u16 csum_type)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2234) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2235) 	struct crypto_shash *csum_shash;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2236) 	const char *csum_driver = btrfs_super_csum_driver(csum_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2237) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2238) 	csum_shash = crypto_alloc_shash(csum_driver, 0, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2239) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2240) 	if (IS_ERR(csum_shash)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2241) 		btrfs_err(fs_info, "error allocating %s hash for checksum",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2242) 			  csum_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2243) 		return PTR_ERR(csum_shash);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2244) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2245) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2246) 	fs_info->csum_shash = csum_shash;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2247) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2248) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2249) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2250) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2251) static int btrfs_replay_log(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2252) 			    struct btrfs_fs_devices *fs_devices)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2253) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2254) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2255) 	struct btrfs_root *log_tree_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2256) 	struct btrfs_super_block *disk_super = fs_info->super_copy;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2257) 	u64 bytenr = btrfs_super_log_root(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2258) 	int level = btrfs_super_log_root_level(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2259) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2260) 	if (fs_devices->rw_devices == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2261) 		btrfs_warn(fs_info, "log replay required on RO media");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2262) 		return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2263) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2264) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2265) 	log_tree_root = btrfs_alloc_root(fs_info, BTRFS_TREE_LOG_OBJECTID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2266) 					 GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2267) 	if (!log_tree_root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2268) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2269) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2270) 	log_tree_root->node = read_tree_block(fs_info, bytenr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2271) 					      fs_info->generation + 1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2272) 					      level, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2273) 	if (IS_ERR(log_tree_root->node)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2274) 		btrfs_warn(fs_info, "failed to read log tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2275) 		ret = PTR_ERR(log_tree_root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2276) 		log_tree_root->node = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2277) 		btrfs_put_root(log_tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2278) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2279) 	} else if (!extent_buffer_uptodate(log_tree_root->node)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2280) 		btrfs_err(fs_info, "failed to read log tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2281) 		btrfs_put_root(log_tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2282) 		return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2283) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2284) 	/* returns with log_tree_root freed on success */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2285) 	ret = btrfs_recover_log_trees(log_tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2286) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2287) 		btrfs_handle_fs_error(fs_info, ret,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2288) 				      "Failed to recover log tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2289) 		btrfs_put_root(log_tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2290) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2291) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2292) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2293) 	if (sb_rdonly(fs_info->sb)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2294) 		ret = btrfs_commit_super(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2295) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2296) 			return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2297) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2298) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2299) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2300) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2301) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2302) static int btrfs_read_roots(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2303) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2304) 	struct btrfs_root *tree_root = fs_info->tree_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2305) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2306) 	struct btrfs_key location;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2307) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2308) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2309) 	BUG_ON(!fs_info->tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2310) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2311) 	location.objectid = BTRFS_EXTENT_TREE_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2312) 	location.type = BTRFS_ROOT_ITEM_KEY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2313) 	location.offset = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2314) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2315) 	root = btrfs_read_tree_root(tree_root, &location);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2316) 	if (IS_ERR(root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2317) 		ret = PTR_ERR(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2318) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2319) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2320) 	set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2321) 	fs_info->extent_root = root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2322) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2323) 	location.objectid = BTRFS_DEV_TREE_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2324) 	root = btrfs_read_tree_root(tree_root, &location);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2325) 	if (IS_ERR(root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2326) 		ret = PTR_ERR(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2327) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2328) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2329) 	set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2330) 	fs_info->dev_root = root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2331) 	btrfs_init_devices_late(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2332) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2333) 	location.objectid = BTRFS_CSUM_TREE_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2334) 	root = btrfs_read_tree_root(tree_root, &location);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2335) 	if (IS_ERR(root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2336) 		ret = PTR_ERR(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2337) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2338) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2339) 	set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2340) 	fs_info->csum_root = root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2341) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2342) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2343) 	 * This tree can share blocks with some other fs tree during relocation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2344) 	 * and we need a proper setup by btrfs_get_fs_root
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2345) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2346) 	root = btrfs_get_fs_root(tree_root->fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2347) 				 BTRFS_DATA_RELOC_TREE_OBJECTID, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2348) 	if (IS_ERR(root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2349) 		ret = PTR_ERR(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2350) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2351) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2352) 	set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2353) 	fs_info->data_reloc_root = root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2354) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2355) 	location.objectid = BTRFS_QUOTA_TREE_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2356) 	root = btrfs_read_tree_root(tree_root, &location);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2357) 	if (!IS_ERR(root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2358) 		set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2359) 		set_bit(BTRFS_FS_QUOTA_ENABLED, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2360) 		fs_info->quota_root = root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2361) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2362) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2363) 	location.objectid = BTRFS_UUID_TREE_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2364) 	root = btrfs_read_tree_root(tree_root, &location);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2365) 	if (IS_ERR(root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2366) 		ret = PTR_ERR(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2367) 		if (ret != -ENOENT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2368) 			goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2369) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2370) 		set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2371) 		fs_info->uuid_root = root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2372) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2373) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2374) 	if (btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2375) 		location.objectid = BTRFS_FREE_SPACE_TREE_OBJECTID;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2376) 		root = btrfs_read_tree_root(tree_root, &location);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2377) 		if (IS_ERR(root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2378) 			ret = PTR_ERR(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2379) 			goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2380) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2381) 		set_bit(BTRFS_ROOT_TRACK_DIRTY, &root->state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2382) 		fs_info->free_space_root = root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2383) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2384) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2385) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2386) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2387) 	btrfs_warn(fs_info, "failed to read root (objectid=%llu): %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2388) 		   location.objectid, ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2389) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2390) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2391) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2392) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2393)  * Real super block validation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2394)  * NOTE: super csum type and incompat features will not be checked here.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2395)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2396)  * @sb:		super block to check
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2397)  * @mirror_num:	the super block number to check its bytenr:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2398)  * 		0	the primary (1st) sb
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2399)  * 		1, 2	2nd and 3rd backup copy
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2400)  * 	       -1	skip bytenr check
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2401)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2402) static int validate_super(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2403) 			    struct btrfs_super_block *sb, int mirror_num)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2404) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2405) 	u64 nodesize = btrfs_super_nodesize(sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2406) 	u64 sectorsize = btrfs_super_sectorsize(sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2407) 	int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2408) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2409) 	if (btrfs_super_magic(sb) != BTRFS_MAGIC) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2410) 		btrfs_err(fs_info, "no valid FS found");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2411) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2412) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2413) 	if (btrfs_super_flags(sb) & ~BTRFS_SUPER_FLAG_SUPP) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2414) 		btrfs_err(fs_info, "unrecognized or unsupported super flag: %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2415) 				btrfs_super_flags(sb) & ~BTRFS_SUPER_FLAG_SUPP);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2416) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2417) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2418) 	if (btrfs_super_root_level(sb) >= BTRFS_MAX_LEVEL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2419) 		btrfs_err(fs_info, "tree_root level too big: %d >= %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2420) 				btrfs_super_root_level(sb), BTRFS_MAX_LEVEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2421) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2422) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2423) 	if (btrfs_super_chunk_root_level(sb) >= BTRFS_MAX_LEVEL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2424) 		btrfs_err(fs_info, "chunk_root level too big: %d >= %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2425) 				btrfs_super_chunk_root_level(sb), BTRFS_MAX_LEVEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2426) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2427) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2428) 	if (btrfs_super_log_root_level(sb) >= BTRFS_MAX_LEVEL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2429) 		btrfs_err(fs_info, "log_root level too big: %d >= %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2430) 				btrfs_super_log_root_level(sb), BTRFS_MAX_LEVEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2431) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2432) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2433) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2434) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2435) 	 * Check sectorsize and nodesize first, other check will need it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2436) 	 * Check all possible sectorsize(4K, 8K, 16K, 32K, 64K) here.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2437) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2438) 	if (!is_power_of_2(sectorsize) || sectorsize < 4096 ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2439) 	    sectorsize > BTRFS_MAX_METADATA_BLOCKSIZE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2440) 		btrfs_err(fs_info, "invalid sectorsize %llu", sectorsize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2441) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2442) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2443) 	/* Only PAGE SIZE is supported yet */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2444) 	if (sectorsize != PAGE_SIZE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2445) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2446) 			"sectorsize %llu not supported yet, only support %lu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2447) 			sectorsize, PAGE_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2448) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2449) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2450) 	if (!is_power_of_2(nodesize) || nodesize < sectorsize ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2451) 	    nodesize > BTRFS_MAX_METADATA_BLOCKSIZE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2452) 		btrfs_err(fs_info, "invalid nodesize %llu", nodesize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2453) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2454) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2455) 	if (nodesize != le32_to_cpu(sb->__unused_leafsize)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2456) 		btrfs_err(fs_info, "invalid leafsize %u, should be %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2457) 			  le32_to_cpu(sb->__unused_leafsize), nodesize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2458) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2459) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2460) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2461) 	/* Root alignment check */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2462) 	if (!IS_ALIGNED(btrfs_super_root(sb), sectorsize)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2463) 		btrfs_warn(fs_info, "tree_root block unaligned: %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2464) 			   btrfs_super_root(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2465) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2466) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2467) 	if (!IS_ALIGNED(btrfs_super_chunk_root(sb), sectorsize)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2468) 		btrfs_warn(fs_info, "chunk_root block unaligned: %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2469) 			   btrfs_super_chunk_root(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2470) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2471) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2472) 	if (!IS_ALIGNED(btrfs_super_log_root(sb), sectorsize)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2473) 		btrfs_warn(fs_info, "log_root block unaligned: %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2474) 			   btrfs_super_log_root(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2475) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2476) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2477) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2478) 	if (memcmp(fs_info->fs_devices->fsid, fs_info->super_copy->fsid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2479) 		   BTRFS_FSID_SIZE)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2480) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2481) 		"superblock fsid doesn't match fsid of fs_devices: %pU != %pU",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2482) 			fs_info->super_copy->fsid, fs_info->fs_devices->fsid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2483) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2484) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2485) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2486) 	if (btrfs_fs_incompat(fs_info, METADATA_UUID) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2487) 	    memcmp(fs_info->fs_devices->metadata_uuid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2488) 		   fs_info->super_copy->metadata_uuid, BTRFS_FSID_SIZE)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2489) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2490) "superblock metadata_uuid doesn't match metadata uuid of fs_devices: %pU != %pU",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2491) 			fs_info->super_copy->metadata_uuid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2492) 			fs_info->fs_devices->metadata_uuid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2493) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2494) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2495) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2496) 	if (memcmp(fs_info->fs_devices->metadata_uuid, sb->dev_item.fsid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2497) 		   BTRFS_FSID_SIZE) != 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2498) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2499) 			"dev_item UUID does not match metadata fsid: %pU != %pU",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2500) 			fs_info->fs_devices->metadata_uuid, sb->dev_item.fsid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2501) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2502) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2503) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2504) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2505) 	 * Hint to catch really bogus numbers, bitflips or so, more exact checks are
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2506) 	 * done later
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2507) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2508) 	if (btrfs_super_bytes_used(sb) < 6 * btrfs_super_nodesize(sb)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2509) 		btrfs_err(fs_info, "bytes_used is too small %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2510) 			  btrfs_super_bytes_used(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2511) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2512) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2513) 	if (!is_power_of_2(btrfs_super_stripesize(sb))) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2514) 		btrfs_err(fs_info, "invalid stripesize %u",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2515) 			  btrfs_super_stripesize(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2516) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2517) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2518) 	if (btrfs_super_num_devices(sb) > (1UL << 31))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2519) 		btrfs_warn(fs_info, "suspicious number of devices: %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2520) 			   btrfs_super_num_devices(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2521) 	if (btrfs_super_num_devices(sb) == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2522) 		btrfs_err(fs_info, "number of devices is 0");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2523) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2524) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2525) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2526) 	if (mirror_num >= 0 &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2527) 	    btrfs_super_bytenr(sb) != btrfs_sb_offset(mirror_num)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2528) 		btrfs_err(fs_info, "super offset mismatch %llu != %u",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2529) 			  btrfs_super_bytenr(sb), BTRFS_SUPER_INFO_OFFSET);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2530) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2531) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2532) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2533) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2534) 	 * Obvious sys_chunk_array corruptions, it must hold at least one key
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2535) 	 * and one chunk
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2536) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2537) 	if (btrfs_super_sys_array_size(sb) > BTRFS_SYSTEM_CHUNK_ARRAY_SIZE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2538) 		btrfs_err(fs_info, "system chunk array too big %u > %u",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2539) 			  btrfs_super_sys_array_size(sb),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2540) 			  BTRFS_SYSTEM_CHUNK_ARRAY_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2541) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2542) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2543) 	if (btrfs_super_sys_array_size(sb) < sizeof(struct btrfs_disk_key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2544) 			+ sizeof(struct btrfs_chunk)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2545) 		btrfs_err(fs_info, "system chunk array too small %u < %zu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2546) 			  btrfs_super_sys_array_size(sb),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2547) 			  sizeof(struct btrfs_disk_key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2548) 			  + sizeof(struct btrfs_chunk));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2549) 		ret = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2550) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2551) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2552) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2553) 	 * The generation is a global counter, we'll trust it more than the others
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2554) 	 * but it's still possible that it's the one that's wrong.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2555) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2556) 	if (btrfs_super_generation(sb) < btrfs_super_chunk_root_generation(sb))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2557) 		btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2558) 			"suspicious: generation < chunk_root_generation: %llu < %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2559) 			btrfs_super_generation(sb),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2560) 			btrfs_super_chunk_root_generation(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2561) 	if (btrfs_super_generation(sb) < btrfs_super_cache_generation(sb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2562) 	    && btrfs_super_cache_generation(sb) != (u64)-1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2563) 		btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2564) 			"suspicious: generation < cache_generation: %llu < %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2565) 			btrfs_super_generation(sb),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2566) 			btrfs_super_cache_generation(sb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2567) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2568) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2569) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2570) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2571) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2572)  * Validation of super block at mount time.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2573)  * Some checks already done early at mount time, like csum type and incompat
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2574)  * flags will be skipped.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2575)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2576) static int btrfs_validate_mount_super(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2577) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2578) 	return validate_super(fs_info, fs_info->super_copy, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2579) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2580) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2581) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2582)  * Validation of super block at write time.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2583)  * Some checks like bytenr check will be skipped as their values will be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2584)  * overwritten soon.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2585)  * Extra checks like csum type and incompat flags will be done here.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2586)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2587) static int btrfs_validate_write_super(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2588) 				      struct btrfs_super_block *sb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2589) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2590) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2591) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2592) 	ret = validate_super(fs_info, sb, -1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2593) 	if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2594) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2595) 	if (!btrfs_supported_super_csum(btrfs_super_csum_type(sb))) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2596) 		ret = -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2597) 		btrfs_err(fs_info, "invalid csum type, has %u want %u",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2598) 			  btrfs_super_csum_type(sb), BTRFS_CSUM_TYPE_CRC32);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2599) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2600) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2601) 	if (btrfs_super_incompat_flags(sb) & ~BTRFS_FEATURE_INCOMPAT_SUPP) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2602) 		ret = -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2603) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2604) 		"invalid incompat flags, has 0x%llx valid mask 0x%llx",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2605) 			  btrfs_super_incompat_flags(sb),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2606) 			  (unsigned long long)BTRFS_FEATURE_INCOMPAT_SUPP);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2607) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2608) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2609) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2610) 	if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2611) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2612) 		"super block corruption detected before writing it to disk");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2613) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2614) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2615) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2616) static int __cold init_tree_roots(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2617) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2618) 	int backup_index = find_newest_super_backup(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2619) 	struct btrfs_super_block *sb = fs_info->super_copy;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2620) 	struct btrfs_root *tree_root = fs_info->tree_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2621) 	bool handle_error = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2622) 	int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2623) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2624) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2625) 	for (i = 0; i < BTRFS_NUM_BACKUP_ROOTS; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2626) 		u64 generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2627) 		int level;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2628) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2629) 		if (handle_error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2630) 			if (!IS_ERR(tree_root->node))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2631) 				free_extent_buffer(tree_root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2632) 			tree_root->node = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2633) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2634) 			if (!btrfs_test_opt(fs_info, USEBACKUPROOT))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2635) 				break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2636) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2637) 			free_root_pointers(fs_info, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2638) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2639) 			/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2640) 			 * Don't use the log in recovery mode, it won't be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2641) 			 * valid
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2642) 			 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2643) 			btrfs_set_super_log_root(sb, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2644) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2645) 			/* We can't trust the free space cache either */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2646) 			btrfs_set_opt(fs_info->mount_opt, CLEAR_CACHE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2647) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2648) 			ret = read_backup_root(fs_info, i);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2649) 			backup_index = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2650) 			if (ret < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2651) 				return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2652) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2653) 		generation = btrfs_super_generation(sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2654) 		level = btrfs_super_root_level(sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2655) 		tree_root->node = read_tree_block(fs_info, btrfs_super_root(sb),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2656) 						  generation, level, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2657) 		if (IS_ERR(tree_root->node)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2658) 			handle_error = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2659) 			ret = PTR_ERR(tree_root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2660) 			tree_root->node = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2661) 			btrfs_warn(fs_info, "couldn't read tree root");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2662) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2663) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2664) 		} else if (!extent_buffer_uptodate(tree_root->node)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2665) 			handle_error = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2666) 			ret = -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2667) 			btrfs_warn(fs_info, "error while reading tree root");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2668) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2669) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2670) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2671) 		btrfs_set_root_node(&tree_root->root_item, tree_root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2672) 		tree_root->commit_root = btrfs_root_node(tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2673) 		btrfs_set_root_refs(&tree_root->root_item, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2674) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2675) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2676) 		 * No need to hold btrfs_root::objectid_mutex since the fs
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2677) 		 * hasn't been fully initialised and we are the only user
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2678) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2679) 		ret = btrfs_find_highest_objectid(tree_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2680) 						&tree_root->highest_objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2681) 		if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2682) 			handle_error = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2683) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2684) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2685) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2686) 		ASSERT(tree_root->highest_objectid <= BTRFS_LAST_FREE_OBJECTID);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2687) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2688) 		ret = btrfs_read_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2689) 		if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2690) 			handle_error = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2691) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2692) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2693) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2694) 		/* All successful */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2695) 		fs_info->generation = generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2696) 		fs_info->last_trans_committed = generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2697) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2698) 		/* Always begin writing backup roots after the one being used */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2699) 		if (backup_index < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2700) 			fs_info->backup_root_index = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2701) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2702) 			fs_info->backup_root_index = backup_index + 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2703) 			fs_info->backup_root_index %= BTRFS_NUM_BACKUP_ROOTS;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2704) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2705) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2706) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2707) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2708) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2709) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2710) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2711) void btrfs_init_fs_info(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2712) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2713) 	INIT_RADIX_TREE(&fs_info->fs_roots_radix, GFP_ATOMIC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2714) 	INIT_RADIX_TREE(&fs_info->buffer_radix, GFP_ATOMIC);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2715) 	INIT_LIST_HEAD(&fs_info->trans_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2716) 	INIT_LIST_HEAD(&fs_info->dead_roots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2717) 	INIT_LIST_HEAD(&fs_info->delayed_iputs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2718) 	INIT_LIST_HEAD(&fs_info->delalloc_roots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2719) 	INIT_LIST_HEAD(&fs_info->caching_block_groups);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2720) 	spin_lock_init(&fs_info->delalloc_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2721) 	spin_lock_init(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2722) 	spin_lock_init(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2723) 	spin_lock_init(&fs_info->delayed_iput_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2724) 	spin_lock_init(&fs_info->defrag_inodes_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2725) 	spin_lock_init(&fs_info->super_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2726) 	spin_lock_init(&fs_info->buffer_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2727) 	spin_lock_init(&fs_info->unused_bgs_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2728) 	rwlock_init(&fs_info->tree_mod_log_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2729) 	mutex_init(&fs_info->unused_bg_unpin_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2730) 	mutex_init(&fs_info->delete_unused_bgs_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2731) 	mutex_init(&fs_info->reloc_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2732) 	mutex_init(&fs_info->delalloc_root_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2733) 	seqlock_init(&fs_info->profiles_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2734) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2735) 	INIT_LIST_HEAD(&fs_info->dirty_cowonly_roots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2736) 	INIT_LIST_HEAD(&fs_info->space_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2737) 	INIT_LIST_HEAD(&fs_info->tree_mod_seq_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2738) 	INIT_LIST_HEAD(&fs_info->unused_bgs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2739) #ifdef CONFIG_BTRFS_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2740) 	INIT_LIST_HEAD(&fs_info->allocated_roots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2741) 	INIT_LIST_HEAD(&fs_info->allocated_ebs);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2742) 	spin_lock_init(&fs_info->eb_leak_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2743) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2744) 	extent_map_tree_init(&fs_info->mapping_tree);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2745) 	btrfs_init_block_rsv(&fs_info->global_block_rsv,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2746) 			     BTRFS_BLOCK_RSV_GLOBAL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2747) 	btrfs_init_block_rsv(&fs_info->trans_block_rsv, BTRFS_BLOCK_RSV_TRANS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2748) 	btrfs_init_block_rsv(&fs_info->chunk_block_rsv, BTRFS_BLOCK_RSV_CHUNK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2749) 	btrfs_init_block_rsv(&fs_info->empty_block_rsv, BTRFS_BLOCK_RSV_EMPTY);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2750) 	btrfs_init_block_rsv(&fs_info->delayed_block_rsv,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2751) 			     BTRFS_BLOCK_RSV_DELOPS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2752) 	btrfs_init_block_rsv(&fs_info->delayed_refs_rsv,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2753) 			     BTRFS_BLOCK_RSV_DELREFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2754) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2755) 	atomic_set(&fs_info->async_delalloc_pages, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2756) 	atomic_set(&fs_info->defrag_running, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2757) 	atomic_set(&fs_info->reada_works_cnt, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2758) 	atomic_set(&fs_info->nr_delayed_iputs, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2759) 	atomic64_set(&fs_info->tree_mod_seq, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2760) 	fs_info->max_inline = BTRFS_DEFAULT_MAX_INLINE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2761) 	fs_info->metadata_ratio = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2762) 	fs_info->defrag_inodes = RB_ROOT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2763) 	atomic64_set(&fs_info->free_chunk_space, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2764) 	fs_info->tree_mod_log = RB_ROOT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2765) 	fs_info->commit_interval = BTRFS_DEFAULT_COMMIT_INTERVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2766) 	fs_info->avg_delayed_ref_runtime = NSEC_PER_SEC >> 6; /* div by 64 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2767) 	/* readahead state */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2768) 	INIT_RADIX_TREE(&fs_info->reada_tree, GFP_NOFS & ~__GFP_DIRECT_RECLAIM);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2769) 	spin_lock_init(&fs_info->reada_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2770) 	btrfs_init_ref_verify(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2771) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2772) 	fs_info->thread_pool_size = min_t(unsigned long,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2773) 					  num_online_cpus() + 2, 8);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2774) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2775) 	INIT_LIST_HEAD(&fs_info->ordered_roots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2776) 	spin_lock_init(&fs_info->ordered_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2777) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2778) 	btrfs_init_scrub(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2779) #ifdef CONFIG_BTRFS_FS_CHECK_INTEGRITY
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2780) 	fs_info->check_integrity_print_mask = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2781) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2782) 	btrfs_init_balance(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2783) 	btrfs_init_async_reclaim_work(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2784) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2785) 	spin_lock_init(&fs_info->block_group_cache_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2786) 	fs_info->block_group_cache_tree = RB_ROOT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2787) 	fs_info->first_logical_byte = (u64)-1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2788) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2789) 	extent_io_tree_init(fs_info, &fs_info->excluded_extents,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2790) 			    IO_TREE_FS_EXCLUDED_EXTENTS, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2791) 	set_bit(BTRFS_FS_BARRIER, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2792) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2793) 	mutex_init(&fs_info->ordered_operations_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2794) 	mutex_init(&fs_info->tree_log_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2795) 	mutex_init(&fs_info->chunk_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2796) 	mutex_init(&fs_info->transaction_kthread_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2797) 	mutex_init(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2798) 	mutex_init(&fs_info->ro_block_group_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2799) 	init_rwsem(&fs_info->commit_root_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2800) 	init_rwsem(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2801) 	init_rwsem(&fs_info->subvol_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2802) 	sema_init(&fs_info->uuid_tree_rescan_sem, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2803) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2804) 	btrfs_init_dev_replace_locks(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2805) 	btrfs_init_qgroup(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2806) 	btrfs_discard_init(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2807) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2808) 	btrfs_init_free_cluster(&fs_info->meta_alloc_cluster);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2809) 	btrfs_init_free_cluster(&fs_info->data_alloc_cluster);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2810) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2811) 	init_waitqueue_head(&fs_info->transaction_throttle);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2812) 	init_waitqueue_head(&fs_info->transaction_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2813) 	init_waitqueue_head(&fs_info->transaction_blocked_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2814) 	init_waitqueue_head(&fs_info->async_submit_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2815) 	init_waitqueue_head(&fs_info->delayed_iputs_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2816) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2817) 	/* Usable values until the real ones are cached from the superblock */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2818) 	fs_info->nodesize = 4096;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2819) 	fs_info->sectorsize = 4096;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2820) 	fs_info->stripesize = 4096;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2821) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2822) 	spin_lock_init(&fs_info->swapfile_pins_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2823) 	fs_info->swapfile_pins = RB_ROOT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2824) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2825) 	fs_info->send_in_progress = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2826) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2827) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2828) static int init_mount_fs_info(struct btrfs_fs_info *fs_info, struct super_block *sb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2829) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2830) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2831) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2832) 	fs_info->sb = sb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2833) 	sb->s_blocksize = BTRFS_BDEV_BLOCKSIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2834) 	sb->s_blocksize_bits = blksize_bits(BTRFS_BDEV_BLOCKSIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2835) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2836) 	ret = percpu_counter_init(&fs_info->dio_bytes, 0, GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2837) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2838) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2839) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2840) 	ret = percpu_counter_init(&fs_info->dirty_metadata_bytes, 0, GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2841) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2842) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2843) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2844) 	fs_info->dirty_metadata_batch = PAGE_SIZE *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2845) 					(1 + ilog2(nr_cpu_ids));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2846) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2847) 	ret = percpu_counter_init(&fs_info->delalloc_bytes, 0, GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2848) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2849) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2850) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2851) 	ret = percpu_counter_init(&fs_info->dev_replace.bio_counter, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2852) 			GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2853) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2854) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2855) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2856) 	fs_info->delayed_root = kmalloc(sizeof(struct btrfs_delayed_root),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2857) 					GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2858) 	if (!fs_info->delayed_root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2859) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2860) 	btrfs_init_delayed_root(fs_info->delayed_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2861) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2862) 	return btrfs_alloc_stripe_hash_table(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2863) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2864) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2865) static int btrfs_uuid_rescan_kthread(void *data)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2866) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2867) 	struct btrfs_fs_info *fs_info = (struct btrfs_fs_info *)data;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2868) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2869) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2870) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2871) 	 * 1st step is to iterate through the existing UUID tree and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2872) 	 * to delete all entries that contain outdated data.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2873) 	 * 2nd step is to add all missing entries to the UUID tree.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2874) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2875) 	ret = btrfs_uuid_tree_iterate(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2876) 	if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2877) 		if (ret != -EINTR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2878) 			btrfs_warn(fs_info, "iterating uuid_tree failed %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2879) 				   ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2880) 		up(&fs_info->uuid_tree_rescan_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2881) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2882) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2883) 	return btrfs_uuid_scan_kthread(data);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2884) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2885) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2886) static int btrfs_check_uuid_tree(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2887) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2888) 	struct task_struct *task;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2889) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2890) 	down(&fs_info->uuid_tree_rescan_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2891) 	task = kthread_run(btrfs_uuid_rescan_kthread, fs_info, "btrfs-uuid");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2892) 	if (IS_ERR(task)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2893) 		/* fs_info->update_uuid_tree_gen remains 0 in all error case */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2894) 		btrfs_warn(fs_info, "failed to start uuid_rescan task");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2895) 		up(&fs_info->uuid_tree_rescan_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2896) 		return PTR_ERR(task);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2897) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2898) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2899) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2900) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2901) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2902) int __cold open_ctree(struct super_block *sb, struct btrfs_fs_devices *fs_devices,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2903) 		      char *options)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2904) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2905) 	u32 sectorsize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2906) 	u32 nodesize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2907) 	u32 stripesize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2908) 	u64 generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2909) 	u64 features;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2910) 	u16 csum_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2911) 	struct btrfs_super_block *disk_super;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2912) 	struct btrfs_fs_info *fs_info = btrfs_sb(sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2913) 	struct btrfs_root *tree_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2914) 	struct btrfs_root *chunk_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2915) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2916) 	int err = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2917) 	int clear_free_space_tree = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2918) 	int level;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2919) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2920) 	ret = init_mount_fs_info(fs_info, sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2921) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2922) 		err = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2923) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2924) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2925) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2926) 	/* These need to be init'ed before we start creating inodes and such. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2927) 	tree_root = btrfs_alloc_root(fs_info, BTRFS_ROOT_TREE_OBJECTID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2928) 				     GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2929) 	fs_info->tree_root = tree_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2930) 	chunk_root = btrfs_alloc_root(fs_info, BTRFS_CHUNK_TREE_OBJECTID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2931) 				      GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2932) 	fs_info->chunk_root = chunk_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2933) 	if (!tree_root || !chunk_root) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2934) 		err = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2935) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2936) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2937) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2938) 	fs_info->btree_inode = new_inode(sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2939) 	if (!fs_info->btree_inode) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2940) 		err = -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2941) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2942) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2943) 	mapping_set_gfp_mask(fs_info->btree_inode->i_mapping, GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2944) 	btrfs_init_btree_inode(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2945) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2946) 	invalidate_bdev(fs_devices->latest_bdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2947) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2948) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2949) 	 * Read super block and check the signature bytes only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2950) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2951) 	disk_super = btrfs_read_dev_super(fs_devices->latest_bdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2952) 	if (IS_ERR(disk_super)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2953) 		err = PTR_ERR(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2954) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2955) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2956) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2957) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2958) 	 * Verify the type first, if that or the checksum value are
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2959) 	 * corrupted, we'll find out
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2960) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2961) 	csum_type = btrfs_super_csum_type(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2962) 	if (!btrfs_supported_super_csum(csum_type)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2963) 		btrfs_err(fs_info, "unsupported checksum algorithm: %u",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2964) 			  csum_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2965) 		err = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2966) 		btrfs_release_disk_super(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2967) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2968) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2969) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2970) 	ret = btrfs_init_csum_hash(fs_info, csum_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2971) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2972) 		err = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2973) 		btrfs_release_disk_super(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2974) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2975) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2976) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2977) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2978) 	 * We want to check superblock checksum, the type is stored inside.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2979) 	 * Pass the whole disk block of size BTRFS_SUPER_INFO_SIZE (4k).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2980) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2981) 	if (btrfs_check_super_csum(fs_info, (u8 *)disk_super)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2982) 		btrfs_err(fs_info, "superblock checksum mismatch");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2983) 		err = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2984) 		btrfs_release_disk_super(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2985) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2986) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2987) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2988) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2989) 	 * super_copy is zeroed at allocation time and we never touch the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2990) 	 * following bytes up to INFO_SIZE, the checksum is calculated from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2991) 	 * the whole block of INFO_SIZE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2992) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2993) 	memcpy(fs_info->super_copy, disk_super, sizeof(*fs_info->super_copy));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2994) 	btrfs_release_disk_super(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2995) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2996) 	disk_super = fs_info->super_copy;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2997) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2998) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2999) 	features = btrfs_super_flags(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3000) 	if (features & BTRFS_SUPER_FLAG_CHANGING_FSID_V2) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3001) 		features &= ~BTRFS_SUPER_FLAG_CHANGING_FSID_V2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3002) 		btrfs_set_super_flags(disk_super, features);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3003) 		btrfs_info(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3004) 			"found metadata UUID change in progress flag, clearing");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3005) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3006) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3007) 	memcpy(fs_info->super_for_commit, fs_info->super_copy,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3008) 	       sizeof(*fs_info->super_for_commit));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3009) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3010) 	ret = btrfs_validate_mount_super(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3011) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3012) 		btrfs_err(fs_info, "superblock contains fatal errors");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3013) 		err = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3014) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3015) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3016) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3017) 	if (!btrfs_super_root(disk_super))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3018) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3019) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3020) 	/* check FS state, whether FS is broken. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3021) 	if (btrfs_super_flags(disk_super) & BTRFS_SUPER_FLAG_ERROR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3022) 		set_bit(BTRFS_FS_STATE_ERROR, &fs_info->fs_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3023) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3024) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3025) 	 * In the long term, we'll store the compression type in the super
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3026) 	 * block, and it'll be used for per file compression control.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3027) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3028) 	fs_info->compress_type = BTRFS_COMPRESS_ZLIB;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3029) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3030) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3031) 	 * Flag our filesystem as having big metadata blocks if they are bigger
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3032) 	 * than the page size
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3033) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3034) 	if (btrfs_super_nodesize(disk_super) > PAGE_SIZE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3035) 		if (!(features & BTRFS_FEATURE_INCOMPAT_BIG_METADATA))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3036) 			btrfs_info(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3037) 				"flagging fs with big metadata feature");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3038) 		features |= BTRFS_FEATURE_INCOMPAT_BIG_METADATA;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3039) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3040) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3041) 	/* Set up fs_info before parsing mount options */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3042) 	nodesize = btrfs_super_nodesize(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3043) 	sectorsize = btrfs_super_sectorsize(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3044) 	stripesize = sectorsize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3045) 	fs_info->dirty_metadata_batch = nodesize * (1 + ilog2(nr_cpu_ids));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3046) 	fs_info->delalloc_batch = sectorsize * 512 * (1 + ilog2(nr_cpu_ids));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3047) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3048) 	/* Cache block sizes */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3049) 	fs_info->nodesize = nodesize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3050) 	fs_info->sectorsize = sectorsize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3051) 	fs_info->stripesize = stripesize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3052) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3053) 	ret = btrfs_parse_options(fs_info, options, sb->s_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3054) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3055) 		err = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3056) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3057) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3058) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3059) 	features = btrfs_super_incompat_flags(disk_super) &
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3060) 		~BTRFS_FEATURE_INCOMPAT_SUPP;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3061) 	if (features) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3062) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3063) 		    "cannot mount because of unsupported optional features (%llx)",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3064) 		    features);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3065) 		err = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3066) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3067) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3068) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3069) 	features = btrfs_super_incompat_flags(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3070) 	features |= BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3071) 	if (fs_info->compress_type == BTRFS_COMPRESS_LZO)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3072) 		features |= BTRFS_FEATURE_INCOMPAT_COMPRESS_LZO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3073) 	else if (fs_info->compress_type == BTRFS_COMPRESS_ZSTD)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3074) 		features |= BTRFS_FEATURE_INCOMPAT_COMPRESS_ZSTD;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3075) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3076) 	if (features & BTRFS_FEATURE_INCOMPAT_SKINNY_METADATA)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3077) 		btrfs_info(fs_info, "has skinny extents");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3078) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3079) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3080) 	 * mixed block groups end up with duplicate but slightly offset
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3081) 	 * extent buffers for the same range.  It leads to corruptions
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3082) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3083) 	if ((features & BTRFS_FEATURE_INCOMPAT_MIXED_GROUPS) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3084) 	    (sectorsize != nodesize)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3085) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3086) "unequal nodesize/sectorsize (%u != %u) are not allowed for mixed block groups",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3087) 			nodesize, sectorsize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3088) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3089) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3090) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3091) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3092) 	 * Needn't use the lock because there is no other task which will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3093) 	 * update the flag.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3094) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3095) 	btrfs_set_super_incompat_flags(disk_super, features);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3096) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3097) 	features = btrfs_super_compat_ro_flags(disk_super) &
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3098) 		~BTRFS_FEATURE_COMPAT_RO_SUPP;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3099) 	if (!sb_rdonly(sb) && features) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3100) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3101) 	"cannot mount read-write because of unsupported optional features (%llx)",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3102) 		       features);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3103) 		err = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3104) 		goto fail_alloc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3105) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3106) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3107) 	ret = btrfs_init_workqueues(fs_info, fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3108) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3109) 		err = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3110) 		goto fail_sb_buffer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3111) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3112) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3113) 	sb->s_bdi->ra_pages *= btrfs_super_num_devices(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3114) 	sb->s_bdi->ra_pages = max(sb->s_bdi->ra_pages, SZ_4M / PAGE_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3115) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3116) 	sb->s_blocksize = sectorsize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3117) 	sb->s_blocksize_bits = blksize_bits(sectorsize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3118) 	memcpy(&sb->s_uuid, fs_info->fs_devices->fsid, BTRFS_FSID_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3119) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3120) 	mutex_lock(&fs_info->chunk_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3121) 	ret = btrfs_read_sys_array(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3122) 	mutex_unlock(&fs_info->chunk_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3123) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3124) 		btrfs_err(fs_info, "failed to read the system array: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3125) 		goto fail_sb_buffer;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3126) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3127) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3128) 	generation = btrfs_super_chunk_root_generation(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3129) 	level = btrfs_super_chunk_root_level(disk_super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3130) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3131) 	chunk_root->node = read_tree_block(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3132) 					   btrfs_super_chunk_root(disk_super),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3133) 					   generation, level, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3134) 	if (IS_ERR(chunk_root->node) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3135) 	    !extent_buffer_uptodate(chunk_root->node)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3136) 		btrfs_err(fs_info, "failed to read chunk root");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3137) 		if (!IS_ERR(chunk_root->node))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3138) 			free_extent_buffer(chunk_root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3139) 		chunk_root->node = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3140) 		goto fail_tree_roots;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3141) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3142) 	btrfs_set_root_node(&chunk_root->root_item, chunk_root->node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3143) 	chunk_root->commit_root = btrfs_root_node(chunk_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3144) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3145) 	read_extent_buffer(chunk_root->node, fs_info->chunk_tree_uuid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3146) 			   offsetof(struct btrfs_header, chunk_tree_uuid),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3147) 			   BTRFS_UUID_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3148) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3149) 	ret = btrfs_read_chunk_tree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3150) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3151) 		btrfs_err(fs_info, "failed to read chunk tree: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3152) 		goto fail_tree_roots;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3153) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3154) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3155) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3156) 	 * Keep the devid that is marked to be the target device for the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3157) 	 * device replace procedure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3158) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3159) 	btrfs_free_extra_devids(fs_devices, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3160) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3161) 	if (!fs_devices->latest_bdev) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3162) 		btrfs_err(fs_info, "failed to read devices");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3163) 		goto fail_tree_roots;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3164) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3165) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3166) 	ret = init_tree_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3167) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3168) 		goto fail_tree_roots;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3169) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3170) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3171) 	 * If we have a uuid root and we're not being told to rescan we need to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3172) 	 * check the generation here so we can set the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3173) 	 * BTRFS_FS_UPDATE_UUID_TREE_GEN bit.  Otherwise we could commit the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3174) 	 * transaction during a balance or the log replay without updating the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3175) 	 * uuid generation, and then if we crash we would rescan the uuid tree,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3176) 	 * even though it was perfectly fine.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3177) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3178) 	if (fs_info->uuid_root && !btrfs_test_opt(fs_info, RESCAN_UUID_TREE) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3179) 	    fs_info->generation == btrfs_super_uuid_tree_generation(disk_super))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3180) 		set_bit(BTRFS_FS_UPDATE_UUID_TREE_GEN, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3181) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3182) 	ret = btrfs_verify_dev_extents(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3183) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3184) 		btrfs_err(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3185) 			  "failed to verify dev extents against chunks: %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3186) 			  ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3187) 		goto fail_block_groups;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3188) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3189) 	ret = btrfs_recover_balance(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3190) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3191) 		btrfs_err(fs_info, "failed to recover balance: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3192) 		goto fail_block_groups;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3193) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3194) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3195) 	ret = btrfs_init_dev_stats(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3196) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3197) 		btrfs_err(fs_info, "failed to init dev_stats: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3198) 		goto fail_block_groups;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3199) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3200) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3201) 	ret = btrfs_init_dev_replace(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3202) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3203) 		btrfs_err(fs_info, "failed to init dev_replace: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3204) 		goto fail_block_groups;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3205) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3206) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3207) 	btrfs_free_extra_devids(fs_devices, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3208) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3209) 	ret = btrfs_sysfs_add_fsid(fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3210) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3211) 		btrfs_err(fs_info, "failed to init sysfs fsid interface: %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3212) 				ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3213) 		goto fail_block_groups;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3214) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3215) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3216) 	ret = btrfs_sysfs_add_mounted(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3217) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3218) 		btrfs_err(fs_info, "failed to init sysfs interface: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3219) 		goto fail_fsdev_sysfs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3220) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3221) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3222) 	ret = btrfs_init_space_info(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3223) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3224) 		btrfs_err(fs_info, "failed to initialize space info: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3225) 		goto fail_sysfs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3226) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3227) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3228) 	ret = btrfs_read_block_groups(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3229) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3230) 		btrfs_err(fs_info, "failed to read block groups: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3231) 		goto fail_sysfs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3232) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3233) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3234) 	if (!sb_rdonly(sb) && fs_info->fs_devices->missing_devices &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3235) 	    !btrfs_check_rw_degradable(fs_info, NULL)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3236) 		btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3237) 		"writable mount is not allowed due to too many missing devices");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3238) 		goto fail_sysfs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3239) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3240) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3241) 	fs_info->cleaner_kthread = kthread_run(cleaner_kthread, tree_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3242) 					       "btrfs-cleaner");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3243) 	if (IS_ERR(fs_info->cleaner_kthread))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3244) 		goto fail_sysfs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3245) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3246) 	fs_info->transaction_kthread = kthread_run(transaction_kthread,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3247) 						   tree_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3248) 						   "btrfs-transaction");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3249) 	if (IS_ERR(fs_info->transaction_kthread))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3250) 		goto fail_cleaner;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3251) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3252) 	if (!btrfs_test_opt(fs_info, NOSSD) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3253) 	    !fs_info->fs_devices->rotating) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3254) 		btrfs_set_and_info(fs_info, SSD, "enabling ssd optimizations");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3255) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3256) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3257) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3258) 	 * Mount does not set all options immediately, we can do it now and do
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3259) 	 * not have to wait for transaction commit
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3260) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3261) 	btrfs_apply_pending_changes(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3262) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3263) #ifdef CONFIG_BTRFS_FS_CHECK_INTEGRITY
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3264) 	if (btrfs_test_opt(fs_info, CHECK_INTEGRITY)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3265) 		ret = btrfsic_mount(fs_info, fs_devices,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3266) 				    btrfs_test_opt(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3267) 					CHECK_INTEGRITY_INCLUDING_EXTENT_DATA) ?
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3268) 				    1 : 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3269) 				    fs_info->check_integrity_print_mask);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3270) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3271) 			btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3272) 				"failed to initialize integrity check module: %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3273) 				ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3274) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3275) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3276) 	ret = btrfs_read_qgroup_config(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3277) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3278) 		goto fail_trans_kthread;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3279) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3280) 	if (btrfs_build_ref_tree(fs_info))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3281) 		btrfs_err(fs_info, "couldn't build ref tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3282) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3283) 	/* do not make disk changes in broken FS or nologreplay is given */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3284) 	if (btrfs_super_log_root(disk_super) != 0 &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3285) 	    !btrfs_test_opt(fs_info, NOLOGREPLAY)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3286) 		btrfs_info(fs_info, "start tree-log replay");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3287) 		ret = btrfs_replay_log(fs_info, fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3288) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3289) 			err = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3290) 			goto fail_qgroup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3291) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3292) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3293) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3294) 	ret = btrfs_find_orphan_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3295) 	if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3296) 		goto fail_qgroup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3297) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3298) 	if (!sb_rdonly(sb)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3299) 		ret = btrfs_cleanup_fs_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3300) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3301) 			goto fail_qgroup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3302) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3303) 		mutex_lock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3304) 		ret = btrfs_recover_relocation(tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3305) 		mutex_unlock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3306) 		if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3307) 			btrfs_warn(fs_info, "failed to recover relocation: %d",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3308) 					ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3309) 			err = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3310) 			goto fail_qgroup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3311) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3312) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3313) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3314) 	fs_info->fs_root = btrfs_get_fs_root(fs_info, BTRFS_FS_TREE_OBJECTID, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3315) 	if (IS_ERR(fs_info->fs_root)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3316) 		err = PTR_ERR(fs_info->fs_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3317) 		btrfs_warn(fs_info, "failed to read fs tree: %d", err);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3318) 		fs_info->fs_root = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3319) 		goto fail_qgroup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3320) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3321) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3322) 	if (sb_rdonly(sb))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3323) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3324) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3325) 	if (btrfs_test_opt(fs_info, CLEAR_CACHE) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3326) 	    btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3327) 		clear_free_space_tree = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3328) 	} else if (btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3329) 		   !btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE_VALID)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3330) 		btrfs_warn(fs_info, "free space tree is invalid");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3331) 		clear_free_space_tree = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3332) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3333) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3334) 	if (clear_free_space_tree) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3335) 		btrfs_info(fs_info, "clearing free space tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3336) 		ret = btrfs_clear_free_space_tree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3337) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3338) 			btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3339) 				   "failed to clear free space tree: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3340) 			close_ctree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3341) 			return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3342) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3343) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3344) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3345) 	if (btrfs_test_opt(fs_info, FREE_SPACE_TREE) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3346) 	    !btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3347) 		btrfs_info(fs_info, "creating free space tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3348) 		ret = btrfs_create_free_space_tree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3349) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3350) 			btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3351) 				"failed to create free space tree: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3352) 			close_ctree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3353) 			return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3354) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3355) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3356) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3357) 	down_read(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3358) 	if ((ret = btrfs_orphan_cleanup(fs_info->fs_root)) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3359) 	    (ret = btrfs_orphan_cleanup(fs_info->tree_root))) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3360) 		up_read(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3361) 		close_ctree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3362) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3363) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3364) 	up_read(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3365) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3366) 	ret = btrfs_resume_balance_async(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3367) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3368) 		btrfs_warn(fs_info, "failed to resume balance: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3369) 		close_ctree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3370) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3371) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3372) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3373) 	ret = btrfs_resume_dev_replace_async(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3374) 	if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3375) 		btrfs_warn(fs_info, "failed to resume device replace: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3376) 		close_ctree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3377) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3378) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3379) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3380) 	btrfs_qgroup_rescan_resume(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3381) 	btrfs_discard_resume(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3382) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3383) 	if (!fs_info->uuid_root) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3384) 		btrfs_info(fs_info, "creating UUID tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3385) 		ret = btrfs_create_uuid_tree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3386) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3387) 			btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3388) 				"failed to create the UUID tree: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3389) 			close_ctree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3390) 			return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3391) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3392) 	} else if (btrfs_test_opt(fs_info, RESCAN_UUID_TREE) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3393) 		   fs_info->generation !=
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3394) 				btrfs_super_uuid_tree_generation(disk_super)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3395) 		btrfs_info(fs_info, "checking UUID tree");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3396) 		ret = btrfs_check_uuid_tree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3397) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3398) 			btrfs_warn(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3399) 				"failed to check the UUID tree: %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3400) 			close_ctree(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3401) 			return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3402) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3403) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3404) 	set_bit(BTRFS_FS_OPEN, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3405) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3406) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3407) 	 * backuproot only affect mount behavior, and if open_ctree succeeded,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3408) 	 * no need to keep the flag
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3409) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3410) 	btrfs_clear_opt(fs_info->mount_opt, USEBACKUPROOT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3411) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3412) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3413) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3414) fail_qgroup:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3415) 	btrfs_free_qgroup_config(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3416) fail_trans_kthread:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3417) 	kthread_stop(fs_info->transaction_kthread);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3418) 	btrfs_cleanup_transaction(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3419) 	btrfs_free_fs_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3420) fail_cleaner:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3421) 	kthread_stop(fs_info->cleaner_kthread);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3422) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3423) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3424) 	 * make sure we're done with the btree inode before we stop our
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3425) 	 * kthreads
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3426) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3427) 	filemap_write_and_wait(fs_info->btree_inode->i_mapping);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3428) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3429) fail_sysfs:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3430) 	btrfs_sysfs_remove_mounted(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3431) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3432) fail_fsdev_sysfs:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3433) 	btrfs_sysfs_remove_fsid(fs_info->fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3434) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3435) fail_block_groups:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3436) 	btrfs_put_block_group_cache(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3437) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3438) fail_tree_roots:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3439) 	if (fs_info->data_reloc_root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3440) 		btrfs_drop_and_free_fs_root(fs_info, fs_info->data_reloc_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3441) 	free_root_pointers(fs_info, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3442) 	invalidate_inode_pages2(fs_info->btree_inode->i_mapping);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3443) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3444) fail_sb_buffer:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3445) 	btrfs_stop_all_workers(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3446) 	btrfs_free_block_groups(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3447) fail_alloc:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3448) 	btrfs_mapping_tree_free(&fs_info->mapping_tree);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3449) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3450) 	iput(fs_info->btree_inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3451) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3452) 	btrfs_close_devices(fs_info->fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3453) 	return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3454) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3455) ALLOW_ERROR_INJECTION(open_ctree, ERRNO);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3456) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3457) static void btrfs_end_super_write(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3458) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3459) 	struct btrfs_device *device = bio->bi_private;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3460) 	struct bio_vec *bvec;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3461) 	struct bvec_iter_all iter_all;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3462) 	struct page *page;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3463) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3464) 	bio_for_each_segment_all(bvec, bio, iter_all) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3465) 		page = bvec->bv_page;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3466) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3467) 		if (bio->bi_status) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3468) 			btrfs_warn_rl_in_rcu(device->fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3469) 				"lost page write due to IO error on %s (%d)",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3470) 				rcu_str_deref(device->name),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3471) 				blk_status_to_errno(bio->bi_status));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3472) 			ClearPageUptodate(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3473) 			SetPageError(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3474) 			btrfs_dev_stat_inc_and_print(device,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3475) 						     BTRFS_DEV_STAT_WRITE_ERRS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3476) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3477) 			SetPageUptodate(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3478) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3479) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3480) 		put_page(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3481) 		unlock_page(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3482) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3483) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3484) 	bio_put(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3485) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3486) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3487) struct btrfs_super_block *btrfs_read_dev_one_super(struct block_device *bdev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3488) 						   int copy_num)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3489) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3490) 	struct btrfs_super_block *super;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3491) 	struct page *page;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3492) 	u64 bytenr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3493) 	struct address_space *mapping = bdev->bd_inode->i_mapping;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3494) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3495) 	bytenr = btrfs_sb_offset(copy_num);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3496) 	if (bytenr + BTRFS_SUPER_INFO_SIZE >= i_size_read(bdev->bd_inode))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3497) 		return ERR_PTR(-EINVAL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3498) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3499) 	page = read_cache_page_gfp(mapping, bytenr >> PAGE_SHIFT, GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3500) 	if (IS_ERR(page))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3501) 		return ERR_CAST(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3502) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3503) 	super = page_address(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3504) 	if (btrfs_super_magic(super) != BTRFS_MAGIC) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3505) 		btrfs_release_disk_super(super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3506) 		return ERR_PTR(-ENODATA);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3507) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3508) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3509) 	if (btrfs_super_bytenr(super) != bytenr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3510) 		btrfs_release_disk_super(super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3511) 		return ERR_PTR(-EINVAL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3512) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3513) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3514) 	return super;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3515) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3516) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3517) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3518) struct btrfs_super_block *btrfs_read_dev_super(struct block_device *bdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3519) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3520) 	struct btrfs_super_block *super, *latest = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3521) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3522) 	u64 transid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3523) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3524) 	/* we would like to check all the supers, but that would make
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3525) 	 * a btrfs mount succeed after a mkfs from a different FS.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3526) 	 * So, we need to add a special mount option to scan for
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3527) 	 * later supers, using BTRFS_SUPER_MIRROR_MAX instead
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3528) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3529) 	for (i = 0; i < 1; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3530) 		super = btrfs_read_dev_one_super(bdev, i);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3531) 		if (IS_ERR(super))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3532) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3533) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3534) 		if (!latest || btrfs_super_generation(super) > transid) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3535) 			if (latest)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3536) 				btrfs_release_disk_super(super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3537) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3538) 			latest = super;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3539) 			transid = btrfs_super_generation(super);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3540) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3541) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3542) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3543) 	return super;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3544) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3545) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3546) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3547)  * Write superblock @sb to the @device. Do not wait for completion, all the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3548)  * pages we use for writing are locked.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3549)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3550)  * Write @max_mirrors copies of the superblock, where 0 means default that fit
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3551)  * the expected device size at commit time. Note that max_mirrors must be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3552)  * same for write and wait phases.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3553)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3554)  * Return number of errors when page is not found or submission fails.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3555)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3556) static int write_dev_supers(struct btrfs_device *device,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3557) 			    struct btrfs_super_block *sb, int max_mirrors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3558) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3559) 	struct btrfs_fs_info *fs_info = device->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3560) 	struct address_space *mapping = device->bdev->bd_inode->i_mapping;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3561) 	SHASH_DESC_ON_STACK(shash, fs_info->csum_shash);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3562) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3563) 	int errors = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3564) 	u64 bytenr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3565) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3566) 	if (max_mirrors == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3567) 		max_mirrors = BTRFS_SUPER_MIRROR_MAX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3568) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3569) 	shash->tfm = fs_info->csum_shash;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3570) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3571) 	for (i = 0; i < max_mirrors; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3572) 		struct page *page;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3573) 		struct bio *bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3574) 		struct btrfs_super_block *disk_super;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3575) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3576) 		bytenr = btrfs_sb_offset(i);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3577) 		if (bytenr + BTRFS_SUPER_INFO_SIZE >=
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3578) 		    device->commit_total_bytes)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3579) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3580) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3581) 		btrfs_set_super_bytenr(sb, bytenr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3582) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3583) 		crypto_shash_digest(shash, (const char *)sb + BTRFS_CSUM_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3584) 				    BTRFS_SUPER_INFO_SIZE - BTRFS_CSUM_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3585) 				    sb->csum);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3586) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3587) 		page = find_or_create_page(mapping, bytenr >> PAGE_SHIFT,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3588) 					   GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3589) 		if (!page) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3590) 			btrfs_err(device->fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3591) 			    "couldn't get super block page for bytenr %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3592) 			    bytenr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3593) 			errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3594) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3595) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3596) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3597) 		/* Bump the refcount for wait_dev_supers() */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3598) 		get_page(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3599) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3600) 		disk_super = page_address(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3601) 		memcpy(disk_super, sb, BTRFS_SUPER_INFO_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3602) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3603) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3604) 		 * Directly use bios here instead of relying on the page cache
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3605) 		 * to do I/O, so we don't lose the ability to do integrity
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3606) 		 * checking.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3607) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3608) 		bio = bio_alloc(GFP_NOFS, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3609) 		bio_set_dev(bio, device->bdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3610) 		bio->bi_iter.bi_sector = bytenr >> SECTOR_SHIFT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3611) 		bio->bi_private = device;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3612) 		bio->bi_end_io = btrfs_end_super_write;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3613) 		__bio_add_page(bio, page, BTRFS_SUPER_INFO_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3614) 			       offset_in_page(bytenr));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3615) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3616) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3617) 		 * We FUA only the first super block.  The others we allow to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3618) 		 * go down lazy and there's a short window where the on-disk
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3619) 		 * copies might still contain the older version.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3620) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3621) 		bio->bi_opf = REQ_OP_WRITE | REQ_SYNC | REQ_META | REQ_PRIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3622) 		if (i == 0 && !btrfs_test_opt(device->fs_info, NOBARRIER))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3623) 			bio->bi_opf |= REQ_FUA;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3624) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3625) 		btrfsic_submit_bio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3626) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3627) 	return errors < i ? 0 : -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3628) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3629) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3630) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3631)  * Wait for write completion of superblocks done by write_dev_supers,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3632)  * @max_mirrors same for write and wait phases.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3633)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3634)  * Return number of errors when page is not found or not marked up to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3635)  * date.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3636)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3637) static int wait_dev_supers(struct btrfs_device *device, int max_mirrors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3638) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3639) 	int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3640) 	int errors = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3641) 	bool primary_failed = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3642) 	u64 bytenr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3643) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3644) 	if (max_mirrors == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3645) 		max_mirrors = BTRFS_SUPER_MIRROR_MAX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3646) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3647) 	for (i = 0; i < max_mirrors; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3648) 		struct page *page;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3649) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3650) 		bytenr = btrfs_sb_offset(i);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3651) 		if (bytenr + BTRFS_SUPER_INFO_SIZE >=
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3652) 		    device->commit_total_bytes)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3653) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3654) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3655) 		page = find_get_page(device->bdev->bd_inode->i_mapping,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3656) 				     bytenr >> PAGE_SHIFT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3657) 		if (!page) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3658) 			errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3659) 			if (i == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3660) 				primary_failed = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3661) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3662) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3663) 		/* Page is submitted locked and unlocked once the IO completes */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3664) 		wait_on_page_locked(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3665) 		if (PageError(page)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3666) 			errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3667) 			if (i == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3668) 				primary_failed = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3669) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3670) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3671) 		/* Drop our reference */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3672) 		put_page(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3673) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3674) 		/* Drop the reference from the writing run */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3675) 		put_page(page);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3676) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3677) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3678) 	/* log error, force error return */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3679) 	if (primary_failed) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3680) 		btrfs_err(device->fs_info, "error writing primary super block to device %llu",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3681) 			  device->devid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3682) 		return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3683) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3684) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3685) 	return errors < i ? 0 : -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3686) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3687) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3688) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3689)  * endio for the write_dev_flush, this will wake anyone waiting
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3690)  * for the barrier when it is done
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3691)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3692) static void btrfs_end_empty_barrier(struct bio *bio)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3693) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3694) 	complete(bio->bi_private);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3695) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3696) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3697) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3698)  * Submit a flush request to the device if it supports it. Error handling is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3699)  * done in the waiting counterpart.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3700)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3701) static void write_dev_flush(struct btrfs_device *device)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3702) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3703) 	struct bio *bio = device->flush_bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3704) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3705) #ifndef CONFIG_BTRFS_FS_CHECK_INTEGRITY
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3706) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3707) 	 * When a disk has write caching disabled, we skip submission of a bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3708) 	 * with flush and sync requests before writing the superblock, since
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3709) 	 * it's not needed. However when the integrity checker is enabled, this
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3710) 	 * results in reports that there are metadata blocks referred by a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3711) 	 * superblock that were not properly flushed. So don't skip the bio
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3712) 	 * submission only when the integrity checker is enabled for the sake
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3713) 	 * of simplicity, since this is a debug tool and not meant for use in
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3714) 	 * non-debug builds.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3715) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3716) 	struct request_queue *q = bdev_get_queue(device->bdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3717) 	if (!test_bit(QUEUE_FLAG_WC, &q->queue_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3718) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3719) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3720) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3721) 	bio_reset(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3722) 	bio->bi_end_io = btrfs_end_empty_barrier;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3723) 	bio_set_dev(bio, device->bdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3724) 	bio->bi_opf = REQ_OP_WRITE | REQ_SYNC | REQ_PREFLUSH;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3725) 	init_completion(&device->flush_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3726) 	bio->bi_private = &device->flush_wait;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3727) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3728) 	btrfsic_submit_bio(bio);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3729) 	set_bit(BTRFS_DEV_STATE_FLUSH_SENT, &device->dev_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3730) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3731) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3732) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3733)  * If the flush bio has been submitted by write_dev_flush, wait for it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3734)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3735) static blk_status_t wait_dev_flush(struct btrfs_device *device)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3736) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3737) 	struct bio *bio = device->flush_bio;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3738) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3739) 	if (!test_bit(BTRFS_DEV_STATE_FLUSH_SENT, &device->dev_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3740) 		return BLK_STS_OK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3741) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3742) 	clear_bit(BTRFS_DEV_STATE_FLUSH_SENT, &device->dev_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3743) 	wait_for_completion_io(&device->flush_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3744) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3745) 	return bio->bi_status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3746) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3747) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3748) static int check_barrier_error(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3749) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3750) 	if (!btrfs_check_rw_degradable(fs_info, NULL))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3751) 		return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3752) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3753) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3754) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3755) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3756)  * send an empty flush down to each device in parallel,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3757)  * then wait for them
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3758)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3759) static int barrier_all_devices(struct btrfs_fs_info *info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3760) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3761) 	struct list_head *head;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3762) 	struct btrfs_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3763) 	int errors_wait = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3764) 	blk_status_t ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3765) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3766) 	lockdep_assert_held(&info->fs_devices->device_list_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3767) 	/* send down all the barriers */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3768) 	head = &info->fs_devices->devices;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3769) 	list_for_each_entry(dev, head, dev_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3770) 		if (test_bit(BTRFS_DEV_STATE_MISSING, &dev->dev_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3771) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3772) 		if (!dev->bdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3773) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3774) 		if (!test_bit(BTRFS_DEV_STATE_IN_FS_METADATA, &dev->dev_state) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3775) 		    !test_bit(BTRFS_DEV_STATE_WRITEABLE, &dev->dev_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3776) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3777) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3778) 		write_dev_flush(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3779) 		dev->last_flush_error = BLK_STS_OK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3780) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3781) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3782) 	/* wait for all the barriers */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3783) 	list_for_each_entry(dev, head, dev_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3784) 		if (test_bit(BTRFS_DEV_STATE_MISSING, &dev->dev_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3785) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3786) 		if (!dev->bdev) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3787) 			errors_wait++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3788) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3789) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3790) 		if (!test_bit(BTRFS_DEV_STATE_IN_FS_METADATA, &dev->dev_state) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3791) 		    !test_bit(BTRFS_DEV_STATE_WRITEABLE, &dev->dev_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3792) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3793) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3794) 		ret = wait_dev_flush(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3795) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3796) 			dev->last_flush_error = ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3797) 			btrfs_dev_stat_inc_and_print(dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3798) 					BTRFS_DEV_STAT_FLUSH_ERRS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3799) 			errors_wait++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3800) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3801) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3802) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3803) 	if (errors_wait) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3804) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3805) 		 * At some point we need the status of all disks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3806) 		 * to arrive at the volume status. So error checking
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3807) 		 * is being pushed to a separate loop.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3808) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3809) 		return check_barrier_error(info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3810) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3811) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3812) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3813) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3814) int btrfs_get_num_tolerated_disk_barrier_failures(u64 flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3815) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3816) 	int raid_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3817) 	int min_tolerated = INT_MAX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3818) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3819) 	if ((flags & BTRFS_BLOCK_GROUP_PROFILE_MASK) == 0 ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3820) 	    (flags & BTRFS_AVAIL_ALLOC_BIT_SINGLE))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3821) 		min_tolerated = min_t(int, min_tolerated,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3822) 				    btrfs_raid_array[BTRFS_RAID_SINGLE].
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3823) 				    tolerated_failures);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3824) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3825) 	for (raid_type = 0; raid_type < BTRFS_NR_RAID_TYPES; raid_type++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3826) 		if (raid_type == BTRFS_RAID_SINGLE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3827) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3828) 		if (!(flags & btrfs_raid_array[raid_type].bg_flag))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3829) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3830) 		min_tolerated = min_t(int, min_tolerated,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3831) 				    btrfs_raid_array[raid_type].
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3832) 				    tolerated_failures);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3833) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3834) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3835) 	if (min_tolerated == INT_MAX) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3836) 		pr_warn("BTRFS: unknown raid flag: %llu", flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3837) 		min_tolerated = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3838) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3839) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3840) 	return min_tolerated;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3841) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3842) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3843) int write_all_supers(struct btrfs_fs_info *fs_info, int max_mirrors)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3844) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3845) 	struct list_head *head;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3846) 	struct btrfs_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3847) 	struct btrfs_super_block *sb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3848) 	struct btrfs_dev_item *dev_item;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3849) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3850) 	int do_barriers;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3851) 	int max_errors;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3852) 	int total_errors = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3853) 	u64 flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3854) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3855) 	do_barriers = !btrfs_test_opt(fs_info, NOBARRIER);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3856) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3857) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3858) 	 * max_mirrors == 0 indicates we're from commit_transaction,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3859) 	 * not from fsync where the tree roots in fs_info have not
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3860) 	 * been consistent on disk.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3861) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3862) 	if (max_mirrors == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3863) 		backup_super_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3864) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3865) 	sb = fs_info->super_for_commit;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3866) 	dev_item = &sb->dev_item;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3867) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3868) 	mutex_lock(&fs_info->fs_devices->device_list_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3869) 	head = &fs_info->fs_devices->devices;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3870) 	max_errors = btrfs_super_num_devices(fs_info->super_copy) - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3871) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3872) 	if (do_barriers) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3873) 		ret = barrier_all_devices(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3874) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3875) 			mutex_unlock(
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3876) 				&fs_info->fs_devices->device_list_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3877) 			btrfs_handle_fs_error(fs_info, ret,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3878) 					      "errors while submitting device barriers.");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3879) 			return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3880) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3881) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3882) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3883) 	list_for_each_entry(dev, head, dev_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3884) 		if (!dev->bdev) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3885) 			total_errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3886) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3887) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3888) 		if (!test_bit(BTRFS_DEV_STATE_IN_FS_METADATA, &dev->dev_state) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3889) 		    !test_bit(BTRFS_DEV_STATE_WRITEABLE, &dev->dev_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3890) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3891) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3892) 		btrfs_set_stack_device_generation(dev_item, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3893) 		btrfs_set_stack_device_type(dev_item, dev->type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3894) 		btrfs_set_stack_device_id(dev_item, dev->devid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3895) 		btrfs_set_stack_device_total_bytes(dev_item,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3896) 						   dev->commit_total_bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3897) 		btrfs_set_stack_device_bytes_used(dev_item,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3898) 						  dev->commit_bytes_used);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3899) 		btrfs_set_stack_device_io_align(dev_item, dev->io_align);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3900) 		btrfs_set_stack_device_io_width(dev_item, dev->io_width);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3901) 		btrfs_set_stack_device_sector_size(dev_item, dev->sector_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3902) 		memcpy(dev_item->uuid, dev->uuid, BTRFS_UUID_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3903) 		memcpy(dev_item->fsid, dev->fs_devices->metadata_uuid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3904) 		       BTRFS_FSID_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3905) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3906) 		flags = btrfs_super_flags(sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3907) 		btrfs_set_super_flags(sb, flags | BTRFS_HEADER_FLAG_WRITTEN);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3908) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3909) 		ret = btrfs_validate_write_super(fs_info, sb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3910) 		if (ret < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3911) 			mutex_unlock(&fs_info->fs_devices->device_list_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3912) 			btrfs_handle_fs_error(fs_info, -EUCLEAN,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3913) 				"unexpected superblock corruption detected");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3914) 			return -EUCLEAN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3915) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3916) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3917) 		ret = write_dev_supers(dev, sb, max_mirrors);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3918) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3919) 			total_errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3920) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3921) 	if (total_errors > max_errors) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3922) 		btrfs_err(fs_info, "%d errors while writing supers",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3923) 			  total_errors);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3924) 		mutex_unlock(&fs_info->fs_devices->device_list_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3925) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3926) 		/* FUA is masked off if unsupported and can't be the reason */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3927) 		btrfs_handle_fs_error(fs_info, -EIO,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3928) 				      "%d errors while writing supers",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3929) 				      total_errors);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3930) 		return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3931) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3932) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3933) 	total_errors = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3934) 	list_for_each_entry(dev, head, dev_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3935) 		if (!dev->bdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3936) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3937) 		if (!test_bit(BTRFS_DEV_STATE_IN_FS_METADATA, &dev->dev_state) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3938) 		    !test_bit(BTRFS_DEV_STATE_WRITEABLE, &dev->dev_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3939) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3940) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3941) 		ret = wait_dev_supers(dev, max_mirrors);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3942) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3943) 			total_errors++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3944) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3945) 	mutex_unlock(&fs_info->fs_devices->device_list_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3946) 	if (total_errors > max_errors) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3947) 		btrfs_handle_fs_error(fs_info, -EIO,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3948) 				      "%d errors while writing supers",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3949) 				      total_errors);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3950) 		return -EIO;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3951) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3952) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3953) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3954) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3955) /* Drop a fs root from the radix tree and free it. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3956) void btrfs_drop_and_free_fs_root(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3957) 				  struct btrfs_root *root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3958) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3959) 	bool drop_ref = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3960) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3961) 	spin_lock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3962) 	radix_tree_delete(&fs_info->fs_roots_radix,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3963) 			  (unsigned long)root->root_key.objectid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3964) 	if (test_and_clear_bit(BTRFS_ROOT_IN_RADIX, &root->state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3965) 		drop_ref = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3966) 	spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3967) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3968) 	if (test_bit(BTRFS_FS_STATE_ERROR, &fs_info->fs_state)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3969) 		ASSERT(root->log_root == NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3970) 		if (root->reloc_root) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3971) 			btrfs_put_root(root->reloc_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3972) 			root->reloc_root = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3973) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3974) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3975) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3976) 	if (root->free_ino_pinned)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3977) 		__btrfs_remove_free_space_cache(root->free_ino_pinned);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3978) 	if (root->free_ino_ctl)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3979) 		__btrfs_remove_free_space_cache(root->free_ino_ctl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3980) 	if (root->ino_cache_inode) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3981) 		iput(root->ino_cache_inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3982) 		root->ino_cache_inode = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3983) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3984) 	if (drop_ref)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3985) 		btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3986) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3987) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3988) int btrfs_cleanup_fs_roots(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3989) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3990) 	u64 root_objectid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3991) 	struct btrfs_root *gang[8];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3992) 	int i = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3993) 	int err = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3994) 	unsigned int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3995) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3996) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3997) 		spin_lock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3998) 		ret = radix_tree_gang_lookup(&fs_info->fs_roots_radix,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3999) 					     (void **)gang, root_objectid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4000) 					     ARRAY_SIZE(gang));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4001) 		if (!ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4002) 			spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4003) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4004) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4005) 		root_objectid = gang[ret - 1]->root_key.objectid + 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4006) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4007) 		for (i = 0; i < ret; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4008) 			/* Avoid to grab roots in dead_roots */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4009) 			if (btrfs_root_refs(&gang[i]->root_item) == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4010) 				gang[i] = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4011) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4012) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4013) 			/* grab all the search result for later use */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4014) 			gang[i] = btrfs_grab_root(gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4015) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4016) 		spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4017) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4018) 		for (i = 0; i < ret; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4019) 			if (!gang[i])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4020) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4021) 			root_objectid = gang[i]->root_key.objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4022) 			err = btrfs_orphan_cleanup(gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4023) 			if (err)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4024) 				break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4025) 			btrfs_put_root(gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4026) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4027) 		root_objectid++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4028) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4029) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4030) 	/* release the uncleaned roots due to error */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4031) 	for (; i < ret; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4032) 		if (gang[i])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4033) 			btrfs_put_root(gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4034) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4035) 	return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4036) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4037) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4038) int btrfs_commit_super(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4039) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4040) 	struct btrfs_root *root = fs_info->tree_root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4041) 	struct btrfs_trans_handle *trans;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4042) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4043) 	mutex_lock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4044) 	btrfs_run_delayed_iputs(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4045) 	mutex_unlock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4046) 	wake_up_process(fs_info->cleaner_kthread);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4047) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4048) 	/* wait until ongoing cleanup work done */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4049) 	down_write(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4050) 	up_write(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4051) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4052) 	trans = btrfs_join_transaction(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4053) 	if (IS_ERR(trans))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4054) 		return PTR_ERR(trans);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4055) 	return btrfs_commit_transaction(trans);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4056) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4057) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4058) void __cold close_ctree(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4059) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4060) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4061) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4062) 	set_bit(BTRFS_FS_CLOSING_START, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4063) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4064) 	 * We don't want the cleaner to start new transactions, add more delayed
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4065) 	 * iputs, etc. while we're closing. We can't use kthread_stop() yet
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4066) 	 * because that frees the task_struct, and the transaction kthread might
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4067) 	 * still try to wake up the cleaner.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4068) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4069) 	kthread_park(fs_info->cleaner_kthread);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4070) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4071) 	/* wait for the qgroup rescan worker to stop */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4072) 	btrfs_qgroup_wait_for_completion(fs_info, false);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4073) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4074) 	/* wait for the uuid_scan task to finish */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4075) 	down(&fs_info->uuid_tree_rescan_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4076) 	/* avoid complains from lockdep et al., set sem back to initial state */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4077) 	up(&fs_info->uuid_tree_rescan_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4078) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4079) 	/* pause restriper - we want to resume on mount */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4080) 	btrfs_pause_balance(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4081) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4082) 	btrfs_dev_replace_suspend_for_unmount(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4083) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4084) 	btrfs_scrub_cancel(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4085) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4086) 	/* wait for any defraggers to finish */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4087) 	wait_event(fs_info->transaction_wait,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4088) 		   (atomic_read(&fs_info->defrag_running) == 0));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4089) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4090) 	/* clear out the rbtree of defraggable inodes */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4091) 	btrfs_cleanup_defrag_inodes(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4092) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4093) 	cancel_work_sync(&fs_info->async_reclaim_work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4094) 	cancel_work_sync(&fs_info->async_data_reclaim_work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4095) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4096) 	/* Cancel or finish ongoing discard work */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4097) 	btrfs_discard_cleanup(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4098) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4099) 	if (!sb_rdonly(fs_info->sb)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4100) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4101) 		 * The cleaner kthread is stopped, so do one final pass over
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4102) 		 * unused block groups.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4103) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4104) 		btrfs_delete_unused_bgs(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4105) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4106) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4107) 		 * There might be existing delayed inode workers still running
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4108) 		 * and holding an empty delayed inode item. We must wait for
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4109) 		 * them to complete first because they can create a transaction.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4110) 		 * This happens when someone calls btrfs_balance_delayed_items()
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4111) 		 * and then a transaction commit runs the same delayed nodes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4112) 		 * before any delayed worker has done something with the nodes.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4113) 		 * We must wait for any worker here and not at transaction
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4114) 		 * commit time since that could cause a deadlock.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4115) 		 * This is a very rare case.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4116) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4117) 		btrfs_flush_workqueue(fs_info->delayed_workers);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4118) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4119) 		ret = btrfs_commit_super(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4120) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4121) 			btrfs_err(fs_info, "commit super ret %d", ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4122) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4123) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4124) 	if (test_bit(BTRFS_FS_STATE_ERROR, &fs_info->fs_state) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4125) 	    test_bit(BTRFS_FS_STATE_TRANS_ABORTED, &fs_info->fs_state))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4126) 		btrfs_error_commit_super(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4127) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4128) 	kthread_stop(fs_info->transaction_kthread);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4129) 	kthread_stop(fs_info->cleaner_kthread);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4130) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4131) 	ASSERT(list_empty(&fs_info->delayed_iputs));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4132) 	set_bit(BTRFS_FS_CLOSING_DONE, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4133) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4134) 	if (btrfs_check_quota_leak(fs_info)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4135) 		WARN_ON(IS_ENABLED(CONFIG_BTRFS_DEBUG));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4136) 		btrfs_err(fs_info, "qgroup reserved space leaked");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4137) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4138) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4139) 	btrfs_free_qgroup_config(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4140) 	ASSERT(list_empty(&fs_info->delalloc_roots));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4141) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4142) 	if (percpu_counter_sum(&fs_info->delalloc_bytes)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4143) 		btrfs_info(fs_info, "at unmount delalloc count %lld",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4144) 		       percpu_counter_sum(&fs_info->delalloc_bytes));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4145) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4146) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4147) 	if (percpu_counter_sum(&fs_info->dio_bytes))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4148) 		btrfs_info(fs_info, "at unmount dio bytes count %lld",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4149) 			   percpu_counter_sum(&fs_info->dio_bytes));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4150) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4151) 	btrfs_sysfs_remove_mounted(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4152) 	btrfs_sysfs_remove_fsid(fs_info->fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4153) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4154) 	btrfs_put_block_group_cache(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4155) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4156) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4157) 	 * we must make sure there is not any read request to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4158) 	 * submit after we stopping all workers.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4159) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4160) 	invalidate_inode_pages2(fs_info->btree_inode->i_mapping);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4161) 	btrfs_stop_all_workers(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4162) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4163) 	clear_bit(BTRFS_FS_OPEN, &fs_info->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4164) 	free_root_pointers(fs_info, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4165) 	btrfs_free_fs_roots(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4166) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4167) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4168) 	 * We must free the block groups after dropping the fs_roots as we could
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4169) 	 * have had an IO error and have left over tree log blocks that aren't
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4170) 	 * cleaned up until the fs roots are freed.  This makes the block group
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4171) 	 * accounting appear to be wrong because there's pending reserved bytes,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4172) 	 * so make sure we do the block group cleanup afterwards.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4173) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4174) 	btrfs_free_block_groups(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4175) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4176) 	iput(fs_info->btree_inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4177) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4178) #ifdef CONFIG_BTRFS_FS_CHECK_INTEGRITY
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4179) 	if (btrfs_test_opt(fs_info, CHECK_INTEGRITY))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4180) 		btrfsic_unmount(fs_info->fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4181) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4182) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4183) 	btrfs_mapping_tree_free(&fs_info->mapping_tree);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4184) 	btrfs_close_devices(fs_info->fs_devices);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4185) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4186) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4187) int btrfs_buffer_uptodate(struct extent_buffer *buf, u64 parent_transid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4188) 			  int atomic)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4189) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4190) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4191) 	struct inode *btree_inode = buf->pages[0]->mapping->host;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4192) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4193) 	ret = extent_buffer_uptodate(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4194) 	if (!ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4195) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4196) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4197) 	ret = verify_parent_transid(&BTRFS_I(btree_inode)->io_tree, buf,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4198) 				    parent_transid, atomic);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4199) 	if (ret == -EAGAIN)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4200) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4201) 	return !ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4202) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4203) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4204) void btrfs_mark_buffer_dirty(struct extent_buffer *buf)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4205) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4206) 	struct btrfs_fs_info *fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4207) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4208) 	u64 transid = btrfs_header_generation(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4209) 	int was_dirty;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4210) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4211) #ifdef CONFIG_BTRFS_FS_RUN_SANITY_TESTS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4212) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4213) 	 * This is a fast path so only do this check if we have sanity tests
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4214) 	 * enabled.  Normal people shouldn't be using unmapped buffers as dirty
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4215) 	 * outside of the sanity tests.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4216) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4217) 	if (unlikely(test_bit(EXTENT_BUFFER_UNMAPPED, &buf->bflags)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4218) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4219) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4220) 	root = BTRFS_I(buf->pages[0]->mapping->host)->root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4221) 	fs_info = root->fs_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4222) 	btrfs_assert_tree_locked(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4223) 	if (transid != fs_info->generation)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4224) 		WARN(1, KERN_CRIT "btrfs transid mismatch buffer %llu, found %llu running %llu\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4225) 			buf->start, transid, fs_info->generation);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4226) 	was_dirty = set_extent_buffer_dirty(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4227) 	if (!was_dirty)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4228) 		percpu_counter_add_batch(&fs_info->dirty_metadata_bytes,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4229) 					 buf->len,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4230) 					 fs_info->dirty_metadata_batch);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4231) #ifdef CONFIG_BTRFS_FS_CHECK_INTEGRITY
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4232) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4233) 	 * Since btrfs_mark_buffer_dirty() can be called with item pointer set
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4234) 	 * but item data not updated.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4235) 	 * So here we should only check item pointers, not item data.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4236) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4237) 	if (btrfs_header_level(buf) == 0 &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4238) 	    btrfs_check_leaf_relaxed(buf)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4239) 		btrfs_print_leaf(buf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4240) 		ASSERT(0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4241) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4242) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4243) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4244) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4245) static void __btrfs_btree_balance_dirty(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4246) 					int flush_delayed)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4247) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4248) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4249) 	 * looks as though older kernels can get into trouble with
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4250) 	 * this code, they end up stuck in balance_dirty_pages forever
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4251) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4252) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4253) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4254) 	if (current->flags & PF_MEMALLOC)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4255) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4256) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4257) 	if (flush_delayed)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4258) 		btrfs_balance_delayed_items(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4259) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4260) 	ret = __percpu_counter_compare(&fs_info->dirty_metadata_bytes,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4261) 				     BTRFS_DIRTY_METADATA_THRESH,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4262) 				     fs_info->dirty_metadata_batch);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4263) 	if (ret > 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4264) 		balance_dirty_pages_ratelimited(fs_info->btree_inode->i_mapping);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4265) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4266) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4267) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4268) void btrfs_btree_balance_dirty(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4269) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4270) 	__btrfs_btree_balance_dirty(fs_info, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4271) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4272) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4273) void btrfs_btree_balance_dirty_nodelay(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4274) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4275) 	__btrfs_btree_balance_dirty(fs_info, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4276) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4277) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4278) int btrfs_read_buffer(struct extent_buffer *buf, u64 parent_transid, int level,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4279) 		      struct btrfs_key *first_key)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4280) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4281) 	return btree_read_extent_buffer_pages(buf, parent_transid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4282) 					      level, first_key);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4283) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4284) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4285) static void btrfs_error_commit_super(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4286) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4287) 	/* cleanup FS via transaction */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4288) 	btrfs_cleanup_transaction(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4289) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4290) 	mutex_lock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4291) 	btrfs_run_delayed_iputs(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4292) 	mutex_unlock(&fs_info->cleaner_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4293) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4294) 	down_write(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4295) 	up_write(&fs_info->cleanup_work_sem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4296) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4297) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4298) static void btrfs_drop_all_logs(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4299) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4300) 	struct btrfs_root *gang[8];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4301) 	u64 root_objectid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4302) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4303) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4304) 	spin_lock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4305) 	while ((ret = radix_tree_gang_lookup(&fs_info->fs_roots_radix,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4306) 					     (void **)gang, root_objectid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4307) 					     ARRAY_SIZE(gang))) != 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4308) 		int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4309) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4310) 		for (i = 0; i < ret; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4311) 			gang[i] = btrfs_grab_root(gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4312) 		spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4313) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4314) 		for (i = 0; i < ret; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4315) 			if (!gang[i])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4316) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4317) 			root_objectid = gang[i]->root_key.objectid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4318) 			btrfs_free_log(NULL, gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4319) 			btrfs_put_root(gang[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4320) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4321) 		root_objectid++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4322) 		spin_lock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4323) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4324) 	spin_unlock(&fs_info->fs_roots_radix_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4325) 	btrfs_free_log_root_tree(NULL, fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4326) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4327) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4328) static void btrfs_destroy_ordered_extents(struct btrfs_root *root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4329) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4330) 	struct btrfs_ordered_extent *ordered;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4331) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4332) 	spin_lock(&root->ordered_extent_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4333) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4334) 	 * This will just short circuit the ordered completion stuff which will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4335) 	 * make sure the ordered extent gets properly cleaned up.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4336) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4337) 	list_for_each_entry(ordered, &root->ordered_extents,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4338) 			    root_extent_list)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4339) 		set_bit(BTRFS_ORDERED_IOERR, &ordered->flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4340) 	spin_unlock(&root->ordered_extent_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4341) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4342) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4343) static void btrfs_destroy_all_ordered_extents(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4344) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4345) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4346) 	struct list_head splice;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4347) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4348) 	INIT_LIST_HEAD(&splice);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4349) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4350) 	spin_lock(&fs_info->ordered_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4351) 	list_splice_init(&fs_info->ordered_roots, &splice);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4352) 	while (!list_empty(&splice)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4353) 		root = list_first_entry(&splice, struct btrfs_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4354) 					ordered_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4355) 		list_move_tail(&root->ordered_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4356) 			       &fs_info->ordered_roots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4357) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4358) 		spin_unlock(&fs_info->ordered_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4359) 		btrfs_destroy_ordered_extents(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4360) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4361) 		cond_resched();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4362) 		spin_lock(&fs_info->ordered_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4363) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4364) 	spin_unlock(&fs_info->ordered_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4365) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4366) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4367) 	 * We need this here because if we've been flipped read-only we won't
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4368) 	 * get sync() from the umount, so we need to make sure any ordered
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4369) 	 * extents that haven't had their dirty pages IO start writeout yet
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4370) 	 * actually get run and error out properly.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4371) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4372) 	btrfs_wait_ordered_roots(fs_info, U64_MAX, 0, (u64)-1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4373) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4374) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4375) static int btrfs_destroy_delayed_refs(struct btrfs_transaction *trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4376) 				      struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4377) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4378) 	struct rb_node *node;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4379) 	struct btrfs_delayed_ref_root *delayed_refs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4380) 	struct btrfs_delayed_ref_node *ref;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4381) 	int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4382) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4383) 	delayed_refs = &trans->delayed_refs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4384) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4385) 	spin_lock(&delayed_refs->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4386) 	if (atomic_read(&delayed_refs->num_entries) == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4387) 		spin_unlock(&delayed_refs->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4388) 		btrfs_debug(fs_info, "delayed_refs has NO entry");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4389) 		return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4390) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4391) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4392) 	while ((node = rb_first_cached(&delayed_refs->href_root)) != NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4393) 		struct btrfs_delayed_ref_head *head;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4394) 		struct rb_node *n;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4395) 		bool pin_bytes = false;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4396) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4397) 		head = rb_entry(node, struct btrfs_delayed_ref_head,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4398) 				href_node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4399) 		if (btrfs_delayed_ref_lock(delayed_refs, head))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4400) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4401) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4402) 		spin_lock(&head->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4403) 		while ((n = rb_first_cached(&head->ref_tree)) != NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4404) 			ref = rb_entry(n, struct btrfs_delayed_ref_node,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4405) 				       ref_node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4406) 			ref->in_tree = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4407) 			rb_erase_cached(&ref->ref_node, &head->ref_tree);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4408) 			RB_CLEAR_NODE(&ref->ref_node);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4409) 			if (!list_empty(&ref->add_list))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4410) 				list_del(&ref->add_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4411) 			atomic_dec(&delayed_refs->num_entries);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4412) 			btrfs_put_delayed_ref(ref);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4413) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4414) 		if (head->must_insert_reserved)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4415) 			pin_bytes = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4416) 		btrfs_free_delayed_extent_op(head->extent_op);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4417) 		btrfs_delete_ref_head(delayed_refs, head);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4418) 		spin_unlock(&head->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4419) 		spin_unlock(&delayed_refs->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4420) 		mutex_unlock(&head->mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4421) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4422) 		if (pin_bytes) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4423) 			struct btrfs_block_group *cache;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4424) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4425) 			cache = btrfs_lookup_block_group(fs_info, head->bytenr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4426) 			BUG_ON(!cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4427) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4428) 			spin_lock(&cache->space_info->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4429) 			spin_lock(&cache->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4430) 			cache->pinned += head->num_bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4431) 			btrfs_space_info_update_bytes_pinned(fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4432) 				cache->space_info, head->num_bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4433) 			cache->reserved -= head->num_bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4434) 			cache->space_info->bytes_reserved -= head->num_bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4435) 			spin_unlock(&cache->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4436) 			spin_unlock(&cache->space_info->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4437) 			percpu_counter_add_batch(
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4438) 				&cache->space_info->total_bytes_pinned,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4439) 				head->num_bytes, BTRFS_TOTAL_BYTES_PINNED_BATCH);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4440) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4441) 			btrfs_put_block_group(cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4442) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4443) 			btrfs_error_unpin_extent_range(fs_info, head->bytenr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4444) 				head->bytenr + head->num_bytes - 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4445) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4446) 		btrfs_cleanup_ref_head_accounting(fs_info, delayed_refs, head);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4447) 		btrfs_put_delayed_ref_head(head);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4448) 		cond_resched();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4449) 		spin_lock(&delayed_refs->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4450) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4451) 	btrfs_qgroup_destroy_extent_records(trans);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4452) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4453) 	spin_unlock(&delayed_refs->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4454) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4455) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4456) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4457) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4458) static void btrfs_destroy_delalloc_inodes(struct btrfs_root *root)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4459) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4460) 	struct btrfs_inode *btrfs_inode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4461) 	struct list_head splice;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4462) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4463) 	INIT_LIST_HEAD(&splice);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4464) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4465) 	spin_lock(&root->delalloc_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4466) 	list_splice_init(&root->delalloc_inodes, &splice);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4467) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4468) 	while (!list_empty(&splice)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4469) 		struct inode *inode = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4470) 		btrfs_inode = list_first_entry(&splice, struct btrfs_inode,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4471) 					       delalloc_inodes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4472) 		__btrfs_del_delalloc_inode(root, btrfs_inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4473) 		spin_unlock(&root->delalloc_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4474) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4475) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4476) 		 * Make sure we get a live inode and that it'll not disappear
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4477) 		 * meanwhile.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4478) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4479) 		inode = igrab(&btrfs_inode->vfs_inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4480) 		if (inode) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4481) 			invalidate_inode_pages2(inode->i_mapping);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4482) 			iput(inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4483) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4484) 		spin_lock(&root->delalloc_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4485) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4486) 	spin_unlock(&root->delalloc_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4487) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4488) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4489) static void btrfs_destroy_all_delalloc_inodes(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4490) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4491) 	struct btrfs_root *root;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4492) 	struct list_head splice;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4493) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4494) 	INIT_LIST_HEAD(&splice);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4495) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4496) 	spin_lock(&fs_info->delalloc_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4497) 	list_splice_init(&fs_info->delalloc_roots, &splice);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4498) 	while (!list_empty(&splice)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4499) 		root = list_first_entry(&splice, struct btrfs_root,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4500) 					 delalloc_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4501) 		root = btrfs_grab_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4502) 		BUG_ON(!root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4503) 		spin_unlock(&fs_info->delalloc_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4504) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4505) 		btrfs_destroy_delalloc_inodes(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4506) 		btrfs_put_root(root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4507) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4508) 		spin_lock(&fs_info->delalloc_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4509) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4510) 	spin_unlock(&fs_info->delalloc_root_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4511) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4512) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4513) static int btrfs_destroy_marked_extents(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4514) 					struct extent_io_tree *dirty_pages,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4515) 					int mark)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4516) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4517) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4518) 	struct extent_buffer *eb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4519) 	u64 start = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4520) 	u64 end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4521) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4522) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4523) 		ret = find_first_extent_bit(dirty_pages, start, &start, &end,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4524) 					    mark, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4525) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4526) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4527) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4528) 		clear_extent_bits(dirty_pages, start, end, mark);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4529) 		while (start <= end) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4530) 			eb = find_extent_buffer(fs_info, start);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4531) 			start += fs_info->nodesize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4532) 			if (!eb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4533) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4534) 			wait_on_extent_buffer_writeback(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4535) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4536) 			if (test_and_clear_bit(EXTENT_BUFFER_DIRTY,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4537) 					       &eb->bflags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4538) 				clear_extent_buffer_dirty(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4539) 			free_extent_buffer_stale(eb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4540) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4541) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4542) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4543) 	return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4544) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4545) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4546) static int btrfs_destroy_pinned_extent(struct btrfs_fs_info *fs_info,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4547) 				       struct extent_io_tree *unpin)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4548) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4549) 	u64 start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4550) 	u64 end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4551) 	int ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4552) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4553) 	while (1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4554) 		struct extent_state *cached_state = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4555) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4556) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4557) 		 * The btrfs_finish_extent_commit() may get the same range as
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4558) 		 * ours between find_first_extent_bit and clear_extent_dirty.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4559) 		 * Hence, hold the unused_bg_unpin_mutex to avoid double unpin
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4560) 		 * the same extent range.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4561) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4562) 		mutex_lock(&fs_info->unused_bg_unpin_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4563) 		ret = find_first_extent_bit(unpin, 0, &start, &end,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4564) 					    EXTENT_DIRTY, &cached_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4565) 		if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4566) 			mutex_unlock(&fs_info->unused_bg_unpin_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4567) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4568) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4569) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4570) 		clear_extent_dirty(unpin, start, end, &cached_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4571) 		free_extent_state(cached_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4572) 		btrfs_error_unpin_extent_range(fs_info, start, end);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4573) 		mutex_unlock(&fs_info->unused_bg_unpin_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4574) 		cond_resched();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4575) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4576) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4577) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4578) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4579) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4580) static void btrfs_cleanup_bg_io(struct btrfs_block_group *cache)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4581) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4582) 	struct inode *inode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4583) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4584) 	inode = cache->io_ctl.inode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4585) 	if (inode) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4586) 		invalidate_inode_pages2(inode->i_mapping);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4587) 		BTRFS_I(inode)->generation = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4588) 		cache->io_ctl.inode = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4589) 		iput(inode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4590) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4591) 	ASSERT(cache->io_ctl.pages == NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4592) 	btrfs_put_block_group(cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4593) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4594) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4595) void btrfs_cleanup_dirty_bgs(struct btrfs_transaction *cur_trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4596) 			     struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4597) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4598) 	struct btrfs_block_group *cache;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4599) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4600) 	spin_lock(&cur_trans->dirty_bgs_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4601) 	while (!list_empty(&cur_trans->dirty_bgs)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4602) 		cache = list_first_entry(&cur_trans->dirty_bgs,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4603) 					 struct btrfs_block_group,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4604) 					 dirty_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4605) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4606) 		if (!list_empty(&cache->io_list)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4607) 			spin_unlock(&cur_trans->dirty_bgs_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4608) 			list_del_init(&cache->io_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4609) 			btrfs_cleanup_bg_io(cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4610) 			spin_lock(&cur_trans->dirty_bgs_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4611) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4612) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4613) 		list_del_init(&cache->dirty_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4614) 		spin_lock(&cache->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4615) 		cache->disk_cache_state = BTRFS_DC_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4616) 		spin_unlock(&cache->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4617) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4618) 		spin_unlock(&cur_trans->dirty_bgs_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4619) 		btrfs_put_block_group(cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4620) 		btrfs_delayed_refs_rsv_release(fs_info, 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4621) 		spin_lock(&cur_trans->dirty_bgs_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4622) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4623) 	spin_unlock(&cur_trans->dirty_bgs_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4624) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4625) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4626) 	 * Refer to the definition of io_bgs member for details why it's safe
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4627) 	 * to use it without any locking
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4628) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4629) 	while (!list_empty(&cur_trans->io_bgs)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4630) 		cache = list_first_entry(&cur_trans->io_bgs,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4631) 					 struct btrfs_block_group,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4632) 					 io_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4633) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4634) 		list_del_init(&cache->io_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4635) 		spin_lock(&cache->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4636) 		cache->disk_cache_state = BTRFS_DC_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4637) 		spin_unlock(&cache->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4638) 		btrfs_cleanup_bg_io(cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4639) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4640) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4641) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4642) void btrfs_cleanup_one_transaction(struct btrfs_transaction *cur_trans,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4643) 				   struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4644) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4645) 	struct btrfs_device *dev, *tmp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4646) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4647) 	btrfs_cleanup_dirty_bgs(cur_trans, fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4648) 	ASSERT(list_empty(&cur_trans->dirty_bgs));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4649) 	ASSERT(list_empty(&cur_trans->io_bgs));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4650) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4651) 	list_for_each_entry_safe(dev, tmp, &cur_trans->dev_update_list,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4652) 				 post_commit_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4653) 		list_del_init(&dev->post_commit_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4654) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4655) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4656) 	btrfs_destroy_delayed_refs(cur_trans, fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4657) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4658) 	cur_trans->state = TRANS_STATE_COMMIT_START;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4659) 	wake_up(&fs_info->transaction_blocked_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4660) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4661) 	cur_trans->state = TRANS_STATE_UNBLOCKED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4662) 	wake_up(&fs_info->transaction_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4663) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4664) 	btrfs_destroy_delayed_inodes(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4665) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4666) 	btrfs_destroy_marked_extents(fs_info, &cur_trans->dirty_pages,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4667) 				     EXTENT_DIRTY);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4668) 	btrfs_destroy_pinned_extent(fs_info, &cur_trans->pinned_extents);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4669) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4670) 	cur_trans->state =TRANS_STATE_COMPLETED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4671) 	wake_up(&cur_trans->commit_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4672) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4673) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4674) static int btrfs_cleanup_transaction(struct btrfs_fs_info *fs_info)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4675) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4676) 	struct btrfs_transaction *t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4677) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4678) 	mutex_lock(&fs_info->transaction_kthread_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4679) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4680) 	spin_lock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4681) 	while (!list_empty(&fs_info->trans_list)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4682) 		t = list_first_entry(&fs_info->trans_list,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4683) 				     struct btrfs_transaction, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4684) 		if (t->state >= TRANS_STATE_COMMIT_START) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4685) 			refcount_inc(&t->use_count);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4686) 			spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4687) 			btrfs_wait_for_commit(fs_info, t->transid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4688) 			btrfs_put_transaction(t);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4689) 			spin_lock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4690) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4691) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4692) 		if (t == fs_info->running_transaction) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4693) 			t->state = TRANS_STATE_COMMIT_DOING;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4694) 			spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4695) 			/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4696) 			 * We wait for 0 num_writers since we don't hold a trans
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4697) 			 * handle open currently for this transaction.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4698) 			 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4699) 			wait_event(t->writer_wait,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4700) 				   atomic_read(&t->num_writers) == 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4701) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4702) 			spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4703) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4704) 		btrfs_cleanup_one_transaction(t, fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4705) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4706) 		spin_lock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4707) 		if (t == fs_info->running_transaction)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4708) 			fs_info->running_transaction = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4709) 		list_del_init(&t->list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4710) 		spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4711) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4712) 		btrfs_put_transaction(t);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4713) 		trace_btrfs_transaction_commit(fs_info->tree_root);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4714) 		spin_lock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4715) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4716) 	spin_unlock(&fs_info->trans_lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4717) 	btrfs_destroy_all_ordered_extents(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4718) 	btrfs_destroy_delayed_inodes(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4719) 	btrfs_assert_delayed_root_empty(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4720) 	btrfs_destroy_all_delalloc_inodes(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4721) 	btrfs_drop_all_logs(fs_info);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4722) 	mutex_unlock(&fs_info->transaction_kthread_mutex);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4723) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4724) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4725) }