Orange Pi5 kernel

Deprecated Linux kernel 5.10.110 for OrangePi 5/5B/5+ boards

3 Commits   0 Branches   0 Tags
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    1) // SPDX-License-Identifier: GPL-2.0-only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    3)  * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    4)  * Copyright 2004-2011 Red Hat, Inc.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    5)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    6) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    7) #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    8) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300    9) #include <linux/fs.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   10) #include <linux/dlm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   11) #include <linux/slab.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   12) #include <linux/types.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   13) #include <linux/delay.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   14) #include <linux/gfs2_ondisk.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   15) #include <linux/sched/signal.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   16) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   17) #include "incore.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   18) #include "glock.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   19) #include "glops.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   20) #include "recovery.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   21) #include "util.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   22) #include "sys.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   23) #include "trace_gfs2.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   24) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   25) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   26)  * gfs2_update_stats - Update time based stats
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   27)  * @mv: Pointer to mean/variance structure to update
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   28)  * @sample: New data to include
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   29)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   30)  * @delta is the difference between the current rtt sample and the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   31)  * running average srtt. We add 1/8 of that to the srtt in order to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   32)  * update the current srtt estimate. The variance estimate is a bit
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   33)  * more complicated. We subtract the current variance estimate from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   34)  * the abs value of the @delta and add 1/4 of that to the running
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   35)  * total.  That's equivalent to 3/4 of the current variance
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   36)  * estimate plus 1/4 of the abs of @delta.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   37)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   38)  * Note that the index points at the array entry containing the smoothed
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   39)  * mean value, and the variance is always in the following entry
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   40)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   41)  * Reference: TCP/IP Illustrated, vol 2, p. 831,832
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   42)  * All times are in units of integer nanoseconds. Unlike the TCP/IP case,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   43)  * they are not scaled fixed point.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   44)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   45) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   46) static inline void gfs2_update_stats(struct gfs2_lkstats *s, unsigned index,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   47) 				     s64 sample)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   48) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   49) 	s64 delta = sample - s->stats[index];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   50) 	s->stats[index] += (delta >> 3);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   51) 	index++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   52) 	s->stats[index] += (s64)(abs(delta) - s->stats[index]) >> 2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   53) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   54) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   55) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   56)  * gfs2_update_reply_times - Update locking statistics
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   57)  * @gl: The glock to update
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   58)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   59)  * This assumes that gl->gl_dstamp has been set earlier.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   60)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   61)  * The rtt (lock round trip time) is an estimate of the time
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   62)  * taken to perform a dlm lock request. We update it on each
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   63)  * reply from the dlm.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   64)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   65)  * The blocking flag is set on the glock for all dlm requests
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   66)  * which may potentially block due to lock requests from other nodes.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   67)  * DLM requests where the current lock state is exclusive, the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   68)  * requested state is null (or unlocked) or where the TRY or
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   69)  * TRY_1CB flags are set are classified as non-blocking. All
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   70)  * other DLM requests are counted as (potentially) blocking.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   71)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   72) static inline void gfs2_update_reply_times(struct gfs2_glock *gl)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   73) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   74) 	struct gfs2_pcpu_lkstats *lks;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   75) 	const unsigned gltype = gl->gl_name.ln_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   76) 	unsigned index = test_bit(GLF_BLOCKING, &gl->gl_flags) ?
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   77) 			 GFS2_LKS_SRTTB : GFS2_LKS_SRTT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   78) 	s64 rtt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   79) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   80) 	preempt_disable();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   81) 	rtt = ktime_to_ns(ktime_sub(ktime_get_real(), gl->gl_dstamp));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   82) 	lks = this_cpu_ptr(gl->gl_name.ln_sbd->sd_lkstats);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   83) 	gfs2_update_stats(&gl->gl_stats, index, rtt);		/* Local */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   84) 	gfs2_update_stats(&lks->lkstats[gltype], index, rtt);	/* Global */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   85) 	preempt_enable();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   86) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   87) 	trace_gfs2_glock_lock_time(gl, rtt);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   88) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   89) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   90) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   91)  * gfs2_update_request_times - Update locking statistics
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   92)  * @gl: The glock to update
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   93)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   94)  * The irt (lock inter-request times) measures the average time
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   95)  * between requests to the dlm. It is updated immediately before
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   96)  * each dlm call.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   97)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   98) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300   99) static inline void gfs2_update_request_times(struct gfs2_glock *gl)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  100) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  101) 	struct gfs2_pcpu_lkstats *lks;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  102) 	const unsigned gltype = gl->gl_name.ln_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  103) 	ktime_t dstamp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  104) 	s64 irt;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  105) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  106) 	preempt_disable();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  107) 	dstamp = gl->gl_dstamp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  108) 	gl->gl_dstamp = ktime_get_real();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  109) 	irt = ktime_to_ns(ktime_sub(gl->gl_dstamp, dstamp));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  110) 	lks = this_cpu_ptr(gl->gl_name.ln_sbd->sd_lkstats);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  111) 	gfs2_update_stats(&gl->gl_stats, GFS2_LKS_SIRT, irt);		/* Local */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  112) 	gfs2_update_stats(&lks->lkstats[gltype], GFS2_LKS_SIRT, irt);	/* Global */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  113) 	preempt_enable();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  114) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  115)  
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  116) static void gdlm_ast(void *arg)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  117) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  118) 	struct gfs2_glock *gl = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  119) 	unsigned ret = gl->gl_state;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  120) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  121) 	gfs2_update_reply_times(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  122) 	BUG_ON(gl->gl_lksb.sb_flags & DLM_SBF_DEMOTED);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  123) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  124) 	if ((gl->gl_lksb.sb_flags & DLM_SBF_VALNOTVALID) && gl->gl_lksb.sb_lvbptr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  125) 		memset(gl->gl_lksb.sb_lvbptr, 0, GDLM_LVB_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  126) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  127) 	switch (gl->gl_lksb.sb_status) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  128) 	case -DLM_EUNLOCK: /* Unlocked, so glock can be freed */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  129) 		if (gl->gl_ops->go_free)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  130) 			gl->gl_ops->go_free(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  131) 		gfs2_glock_free(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  132) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  133) 	case -DLM_ECANCEL: /* Cancel while getting lock */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  134) 		ret |= LM_OUT_CANCELED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  135) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  136) 	case -EAGAIN: /* Try lock fails */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  137) 	case -EDEADLK: /* Deadlock detected */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  138) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  139) 	case -ETIMEDOUT: /* Canceled due to timeout */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  140) 		ret |= LM_OUT_ERROR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  141) 		goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  142) 	case 0: /* Success */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  143) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  144) 	default: /* Something unexpected */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  145) 		BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  146) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  147) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  148) 	ret = gl->gl_req;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  149) 	if (gl->gl_lksb.sb_flags & DLM_SBF_ALTMODE) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  150) 		if (gl->gl_req == LM_ST_SHARED)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  151) 			ret = LM_ST_DEFERRED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  152) 		else if (gl->gl_req == LM_ST_DEFERRED)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  153) 			ret = LM_ST_SHARED;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  154) 		else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  155) 			BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  156) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  157) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  158) 	set_bit(GLF_INITIAL, &gl->gl_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  159) 	gfs2_glock_complete(gl, ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  160) 	return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  161) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  162) 	if (!test_bit(GLF_INITIAL, &gl->gl_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  163) 		gl->gl_lksb.sb_lkid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  164) 	gfs2_glock_complete(gl, ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  165) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  166) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  167) static void gdlm_bast(void *arg, int mode)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  168) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  169) 	struct gfs2_glock *gl = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  170) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  171) 	switch (mode) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  172) 	case DLM_LOCK_EX:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  173) 		gfs2_glock_cb(gl, LM_ST_UNLOCKED);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  174) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  175) 	case DLM_LOCK_CW:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  176) 		gfs2_glock_cb(gl, LM_ST_DEFERRED);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  177) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  178) 	case DLM_LOCK_PR:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  179) 		gfs2_glock_cb(gl, LM_ST_SHARED);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  180) 		break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  181) 	default:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  182) 		fs_err(gl->gl_name.ln_sbd, "unknown bast mode %d\n", mode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  183) 		BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  184) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  185) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  186) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  187) /* convert gfs lock-state to dlm lock-mode */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  188) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  189) static int make_mode(struct gfs2_sbd *sdp, const unsigned int lmstate)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  190) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  191) 	switch (lmstate) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  192) 	case LM_ST_UNLOCKED:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  193) 		return DLM_LOCK_NL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  194) 	case LM_ST_EXCLUSIVE:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  195) 		return DLM_LOCK_EX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  196) 	case LM_ST_DEFERRED:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  197) 		return DLM_LOCK_CW;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  198) 	case LM_ST_SHARED:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  199) 		return DLM_LOCK_PR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  200) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  201) 	fs_err(sdp, "unknown LM state %d\n", lmstate);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  202) 	BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  203) 	return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  204) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  205) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  206) static u32 make_flags(struct gfs2_glock *gl, const unsigned int gfs_flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  207) 		      const int req)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  208) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  209) 	u32 lkf = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  210) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  211) 	if (gl->gl_lksb.sb_lvbptr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  212) 		lkf |= DLM_LKF_VALBLK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  213) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  214) 	if (gfs_flags & LM_FLAG_TRY)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  215) 		lkf |= DLM_LKF_NOQUEUE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  216) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  217) 	if (gfs_flags & LM_FLAG_TRY_1CB) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  218) 		lkf |= DLM_LKF_NOQUEUE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  219) 		lkf |= DLM_LKF_NOQUEUEBAST;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  220) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  221) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  222) 	if (gfs_flags & LM_FLAG_PRIORITY) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  223) 		lkf |= DLM_LKF_NOORDER;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  224) 		lkf |= DLM_LKF_HEADQUE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  225) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  226) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  227) 	if (gfs_flags & LM_FLAG_ANY) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  228) 		if (req == DLM_LOCK_PR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  229) 			lkf |= DLM_LKF_ALTCW;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  230) 		else if (req == DLM_LOCK_CW)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  231) 			lkf |= DLM_LKF_ALTPR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  232) 		else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  233) 			BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  234) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  235) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  236) 	if (gl->gl_lksb.sb_lkid != 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  237) 		lkf |= DLM_LKF_CONVERT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  238) 		if (test_bit(GLF_BLOCKING, &gl->gl_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  239) 			lkf |= DLM_LKF_QUECVT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  240) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  241) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  242) 	return lkf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  243) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  244) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  245) static void gfs2_reverse_hex(char *c, u64 value)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  246) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  247) 	*c = '0';
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  248) 	while (value) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  249) 		*c-- = hex_asc[value & 0x0f];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  250) 		value >>= 4;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  251) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  252) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  253) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  254) static int gdlm_lock(struct gfs2_glock *gl, unsigned int req_state,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  255) 		     unsigned int flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  256) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  257) 	struct lm_lockstruct *ls = &gl->gl_name.ln_sbd->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  258) 	int req;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  259) 	u32 lkf;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  260) 	char strname[GDLM_STRNAME_BYTES] = "";
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  261) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  262) 	req = make_mode(gl->gl_name.ln_sbd, req_state);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  263) 	lkf = make_flags(gl, flags, req);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  264) 	gfs2_glstats_inc(gl, GFS2_LKS_DCOUNT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  265) 	gfs2_sbstats_inc(gl, GFS2_LKS_DCOUNT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  266) 	if (gl->gl_lksb.sb_lkid) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  267) 		gfs2_update_request_times(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  268) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  269) 		memset(strname, ' ', GDLM_STRNAME_BYTES - 1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  270) 		strname[GDLM_STRNAME_BYTES - 1] = '\0';
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  271) 		gfs2_reverse_hex(strname + 7, gl->gl_name.ln_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  272) 		gfs2_reverse_hex(strname + 23, gl->gl_name.ln_number);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  273) 		gl->gl_dstamp = ktime_get_real();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  274) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  275) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  276) 	 * Submit the actual lock request.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  277) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  278) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  279) 	return dlm_lock(ls->ls_dlm, req, &gl->gl_lksb, lkf, strname,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  280) 			GDLM_STRNAME_BYTES - 1, 0, gdlm_ast, gl, gdlm_bast);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  281) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  282) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  283) static void gdlm_put_lock(struct gfs2_glock *gl)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  284) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  285) 	struct gfs2_sbd *sdp = gl->gl_name.ln_sbd;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  286) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  287) 	int error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  288) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  289) 	if (gl->gl_lksb.sb_lkid == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  290) 		gfs2_glock_free(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  291) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  292) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  293) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  294) 	clear_bit(GLF_BLOCKING, &gl->gl_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  295) 	gfs2_glstats_inc(gl, GFS2_LKS_DCOUNT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  296) 	gfs2_sbstats_inc(gl, GFS2_LKS_DCOUNT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  297) 	gfs2_update_request_times(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  298) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  299) 	/* don't want to call dlm if we've unmounted the lock protocol */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  300) 	if (test_bit(DFL_UNMOUNT, &ls->ls_recover_flags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  301) 		gfs2_glock_free(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  302) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  303) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  304) 	/* don't want to skip dlm_unlock writing the lvb when lock has one */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  305) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  306) 	if (test_bit(SDF_SKIP_DLM_UNLOCK, &sdp->sd_flags) &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  307) 	    !gl->gl_lksb.sb_lvbptr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  308) 		gfs2_glock_free(gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  309) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  310) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  311) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  312) 	error = dlm_unlock(ls->ls_dlm, gl->gl_lksb.sb_lkid, DLM_LKF_VALBLK,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  313) 			   NULL, gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  314) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  315) 		fs_err(sdp, "gdlm_unlock %x,%llx err=%d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  316) 		       gl->gl_name.ln_type,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  317) 		       (unsigned long long)gl->gl_name.ln_number, error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  318) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  319) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  320) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  321) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  322) static void gdlm_cancel(struct gfs2_glock *gl)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  323) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  324) 	struct lm_lockstruct *ls = &gl->gl_name.ln_sbd->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  325) 	dlm_unlock(ls->ls_dlm, gl->gl_lksb.sb_lkid, DLM_LKF_CANCEL, NULL, gl);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  326) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  327) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  328) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  329)  * dlm/gfs2 recovery coordination using dlm_recover callbacks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  330)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  331)  *  0. gfs2 checks for another cluster node withdraw, needing journal replay
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  332)  *  1. dlm_controld sees lockspace members change
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  333)  *  2. dlm_controld blocks dlm-kernel locking activity
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  334)  *  3. dlm_controld within dlm-kernel notifies gfs2 (recover_prep)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  335)  *  4. dlm_controld starts and finishes its own user level recovery
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  336)  *  5. dlm_controld starts dlm-kernel dlm_recoverd to do kernel recovery
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  337)  *  6. dlm_recoverd notifies gfs2 of failed nodes (recover_slot)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  338)  *  7. dlm_recoverd does its own lock recovery
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  339)  *  8. dlm_recoverd unblocks dlm-kernel locking activity
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  340)  *  9. dlm_recoverd notifies gfs2 when done (recover_done with new generation)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  341)  * 10. gfs2_control updates control_lock lvb with new generation and jid bits
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  342)  * 11. gfs2_control enqueues journals for gfs2_recover to recover (maybe none)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  343)  * 12. gfs2_recover dequeues and recovers journals of failed nodes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  344)  * 13. gfs2_recover provides recovery results to gfs2_control (recovery_result)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  345)  * 14. gfs2_control updates control_lock lvb jid bits for recovered journals
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  346)  * 15. gfs2_control unblocks normal locking when all journals are recovered
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  347)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  348)  * - failures during recovery
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  349)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  350)  * recover_prep() may set BLOCK_LOCKS (step 3) again before gfs2_control
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  351)  * clears BLOCK_LOCKS (step 15), e.g. another node fails while still
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  352)  * recovering for a prior failure.  gfs2_control needs a way to detect
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  353)  * this so it can leave BLOCK_LOCKS set in step 15.  This is managed using
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  354)  * the recover_block and recover_start values.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  355)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  356)  * recover_done() provides a new lockspace generation number each time it
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  357)  * is called (step 9).  This generation number is saved as recover_start.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  358)  * When recover_prep() is called, it sets BLOCK_LOCKS and sets
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  359)  * recover_block = recover_start.  So, while recover_block is equal to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  360)  * recover_start, BLOCK_LOCKS should remain set.  (recover_spin must
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  361)  * be held around the BLOCK_LOCKS/recover_block/recover_start logic.)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  362)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  363)  * - more specific gfs2 steps in sequence above
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  364)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  365)  *  3. recover_prep sets BLOCK_LOCKS and sets recover_block = recover_start
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  366)  *  6. recover_slot records any failed jids (maybe none)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  367)  *  9. recover_done sets recover_start = new generation number
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  368)  * 10. gfs2_control sets control_lock lvb = new gen + bits for failed jids
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  369)  * 12. gfs2_recover does journal recoveries for failed jids identified above
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  370)  * 14. gfs2_control clears control_lock lvb bits for recovered jids
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  371)  * 15. gfs2_control checks if recover_block == recover_start (step 3 occured
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  372)  *     again) then do nothing, otherwise if recover_start > recover_block
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  373)  *     then clear BLOCK_LOCKS.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  374)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  375)  * - parallel recovery steps across all nodes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  376)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  377)  * All nodes attempt to update the control_lock lvb with the new generation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  378)  * number and jid bits, but only the first to get the control_lock EX will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  379)  * do so; others will see that it's already done (lvb already contains new
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  380)  * generation number.)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  381)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  382)  * . All nodes get the same recover_prep/recover_slot/recover_done callbacks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  383)  * . All nodes attempt to set control_lock lvb gen + bits for the new gen
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  384)  * . One node gets control_lock first and writes the lvb, others see it's done
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  385)  * . All nodes attempt to recover jids for which they see control_lock bits set
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  386)  * . One node succeeds for a jid, and that one clears the jid bit in the lvb
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  387)  * . All nodes will eventually see all lvb bits clear and unblock locks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  388)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  389)  * - is there a problem with clearing an lvb bit that should be set
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  390)  *   and missing a journal recovery?
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  391)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  392)  * 1. jid fails
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  393)  * 2. lvb bit set for step 1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  394)  * 3. jid recovered for step 1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  395)  * 4. jid taken again (new mount)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  396)  * 5. jid fails (for step 4)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  397)  * 6. lvb bit set for step 5 (will already be set)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  398)  * 7. lvb bit cleared for step 3
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  399)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  400)  * This is not a problem because the failure in step 5 does not
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  401)  * require recovery, because the mount in step 4 could not have
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  402)  * progressed far enough to unblock locks and access the fs.  The
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  403)  * control_mount() function waits for all recoveries to be complete
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  404)  * for the latest lockspace generation before ever unblocking locks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  405)  * and returning.  The mount in step 4 waits until the recovery in
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  406)  * step 1 is done.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  407)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  408)  * - special case of first mounter: first node to mount the fs
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  409)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  410)  * The first node to mount a gfs2 fs needs to check all the journals
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  411)  * and recover any that need recovery before other nodes are allowed
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  412)  * to mount the fs.  (Others may begin mounting, but they must wait
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  413)  * for the first mounter to be done before taking locks on the fs
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  414)  * or accessing the fs.)  This has two parts:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  415)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  416)  * 1. The mounted_lock tells a node it's the first to mount the fs.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  417)  * Each node holds the mounted_lock in PR while it's mounted.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  418)  * Each node tries to acquire the mounted_lock in EX when it mounts.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  419)  * If a node is granted the mounted_lock EX it means there are no
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  420)  * other mounted nodes (no PR locks exist), and it is the first mounter.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  421)  * The mounted_lock is demoted to PR when first recovery is done, so
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  422)  * others will fail to get an EX lock, but will get a PR lock.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  423)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  424)  * 2. The control_lock blocks others in control_mount() while the first
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  425)  * mounter is doing first mount recovery of all journals.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  426)  * A mounting node needs to acquire control_lock in EX mode before
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  427)  * it can proceed.  The first mounter holds control_lock in EX while doing
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  428)  * the first mount recovery, blocking mounts from other nodes, then demotes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  429)  * control_lock to NL when it's done (others_may_mount/first_done),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  430)  * allowing other nodes to continue mounting.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  431)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  432)  * first mounter:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  433)  * control_lock EX/NOQUEUE success
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  434)  * mounted_lock EX/NOQUEUE success (no other PR, so no other mounters)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  435)  * set first=1
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  436)  * do first mounter recovery
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  437)  * mounted_lock EX->PR
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  438)  * control_lock EX->NL, write lvb generation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  439)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  440)  * other mounter:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  441)  * control_lock EX/NOQUEUE success (if fail -EAGAIN, retry)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  442)  * mounted_lock EX/NOQUEUE fail -EAGAIN (expected due to other mounters PR)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  443)  * mounted_lock PR/NOQUEUE success
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  444)  * read lvb generation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  445)  * control_lock EX->NL
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  446)  * set first=0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  447)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  448)  * - mount during recovery
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  449)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  450)  * If a node mounts while others are doing recovery (not first mounter),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  451)  * the mounting node will get its initial recover_done() callback without
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  452)  * having seen any previous failures/callbacks.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  453)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  454)  * It must wait for all recoveries preceding its mount to be finished
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  455)  * before it unblocks locks.  It does this by repeating the "other mounter"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  456)  * steps above until the lvb generation number is >= its mount generation
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  457)  * number (from initial recover_done) and all lvb bits are clear.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  458)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  459)  * - control_lock lvb format
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  460)  *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  461)  * 4 bytes generation number: the latest dlm lockspace generation number
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  462)  * from recover_done callback.  Indicates the jid bitmap has been updated
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  463)  * to reflect all slot failures through that generation.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  464)  * 4 bytes unused.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  465)  * GDLM_LVB_SIZE-8 bytes of jid bit map. If bit N is set, it indicates
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  466)  * that jid N needs recovery.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  467)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  468) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  469) #define JID_BITMAP_OFFSET 8 /* 4 byte generation number + 4 byte unused */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  470) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  471) static void control_lvb_read(struct lm_lockstruct *ls, uint32_t *lvb_gen,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  472) 			     char *lvb_bits)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  473) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  474) 	__le32 gen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  475) 	memcpy(lvb_bits, ls->ls_control_lvb, GDLM_LVB_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  476) 	memcpy(&gen, lvb_bits, sizeof(__le32));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  477) 	*lvb_gen = le32_to_cpu(gen);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  478) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  479) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  480) static void control_lvb_write(struct lm_lockstruct *ls, uint32_t lvb_gen,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  481) 			      char *lvb_bits)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  482) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  483) 	__le32 gen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  484) 	memcpy(ls->ls_control_lvb, lvb_bits, GDLM_LVB_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  485) 	gen = cpu_to_le32(lvb_gen);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  486) 	memcpy(ls->ls_control_lvb, &gen, sizeof(__le32));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  487) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  488) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  489) static int all_jid_bits_clear(char *lvb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  490) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  491) 	return !memchr_inv(lvb + JID_BITMAP_OFFSET, 0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  492) 			GDLM_LVB_SIZE - JID_BITMAP_OFFSET);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  493) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  494) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  495) static void sync_wait_cb(void *arg)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  496) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  497) 	struct lm_lockstruct *ls = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  498) 	complete(&ls->ls_sync_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  499) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  500) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  501) static int sync_unlock(struct gfs2_sbd *sdp, struct dlm_lksb *lksb, char *name)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  502) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  503) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  504) 	int error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  505) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  506) 	error = dlm_unlock(ls->ls_dlm, lksb->sb_lkid, 0, lksb, ls);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  507) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  508) 		fs_err(sdp, "%s lkid %x error %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  509) 		       name, lksb->sb_lkid, error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  510) 		return error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  511) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  512) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  513) 	wait_for_completion(&ls->ls_sync_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  514) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  515) 	if (lksb->sb_status != -DLM_EUNLOCK) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  516) 		fs_err(sdp, "%s lkid %x status %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  517) 		       name, lksb->sb_lkid, lksb->sb_status);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  518) 		return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  519) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  520) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  521) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  522) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  523) static int sync_lock(struct gfs2_sbd *sdp, int mode, uint32_t flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  524) 		     unsigned int num, struct dlm_lksb *lksb, char *name)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  525) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  526) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  527) 	char strname[GDLM_STRNAME_BYTES];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  528) 	int error, status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  529) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  530) 	memset(strname, 0, GDLM_STRNAME_BYTES);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  531) 	snprintf(strname, GDLM_STRNAME_BYTES, "%8x%16x", LM_TYPE_NONDISK, num);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  532) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  533) 	error = dlm_lock(ls->ls_dlm, mode, lksb, flags,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  534) 			 strname, GDLM_STRNAME_BYTES - 1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  535) 			 0, sync_wait_cb, ls, NULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  536) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  537) 		fs_err(sdp, "%s lkid %x flags %x mode %d error %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  538) 		       name, lksb->sb_lkid, flags, mode, error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  539) 		return error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  540) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  541) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  542) 	wait_for_completion(&ls->ls_sync_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  543) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  544) 	status = lksb->sb_status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  545) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  546) 	if (status && status != -EAGAIN) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  547) 		fs_err(sdp, "%s lkid %x flags %x mode %d status %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  548) 		       name, lksb->sb_lkid, flags, mode, status);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  549) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  550) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  551) 	return status;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  552) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  553) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  554) static int mounted_unlock(struct gfs2_sbd *sdp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  555) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  556) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  557) 	return sync_unlock(sdp, &ls->ls_mounted_lksb, "mounted_lock");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  558) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  559) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  560) static int mounted_lock(struct gfs2_sbd *sdp, int mode, uint32_t flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  561) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  562) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  563) 	return sync_lock(sdp, mode, flags, GFS2_MOUNTED_LOCK,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  564) 			 &ls->ls_mounted_lksb, "mounted_lock");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  565) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  566) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  567) static int control_unlock(struct gfs2_sbd *sdp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  568) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  569) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  570) 	return sync_unlock(sdp, &ls->ls_control_lksb, "control_lock");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  571) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  572) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  573) static int control_lock(struct gfs2_sbd *sdp, int mode, uint32_t flags)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  574) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  575) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  576) 	return sync_lock(sdp, mode, flags, GFS2_CONTROL_LOCK,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  577) 			 &ls->ls_control_lksb, "control_lock");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  578) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  579) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  580) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  581)  * remote_withdraw - react to a node withdrawing from the file system
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  582)  * @sdp: The superblock
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  583)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  584) static void remote_withdraw(struct gfs2_sbd *sdp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  585) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  586) 	struct gfs2_jdesc *jd;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  587) 	int ret = 0, count = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  588) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  589) 	list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  590) 		if (jd->jd_jid == sdp->sd_lockstruct.ls_jid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  591) 			continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  592) 		ret = gfs2_recover_journal(jd, true);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  593) 		if (ret)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  594) 			break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  595) 		count++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  596) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  597) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  598) 	/* Now drop the additional reference we acquired */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  599) 	fs_err(sdp, "Journals checked: %d, ret = %d.\n", count, ret);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  600) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  601) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  602) static void gfs2_control_func(struct work_struct *work)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  603) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  604) 	struct gfs2_sbd *sdp = container_of(work, struct gfs2_sbd, sd_control_work.work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  605) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  606) 	uint32_t block_gen, start_gen, lvb_gen, flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  607) 	int recover_set = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  608) 	int write_lvb = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  609) 	int recover_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  610) 	int i, error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  611) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  612) 	/* First check for other nodes that may have done a withdraw. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  613) 	if (test_bit(SDF_REMOTE_WITHDRAW, &sdp->sd_flags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  614) 		remote_withdraw(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  615) 		clear_bit(SDF_REMOTE_WITHDRAW, &sdp->sd_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  616) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  617) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  618) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  619) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  620) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  621) 	 * No MOUNT_DONE means we're still mounting; control_mount()
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  622) 	 * will set this flag, after which this thread will take over
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  623) 	 * all further clearing of BLOCK_LOCKS.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  624) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  625) 	 * FIRST_MOUNT means this node is doing first mounter recovery,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  626) 	 * for which recovery control is handled by
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  627) 	 * control_mount()/control_first_done(), not this thread.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  628) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  629) 	if (!test_bit(DFL_MOUNT_DONE, &ls->ls_recover_flags) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  630) 	     test_bit(DFL_FIRST_MOUNT, &ls->ls_recover_flags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  631) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  632) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  633) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  634) 	block_gen = ls->ls_recover_block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  635) 	start_gen = ls->ls_recover_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  636) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  637) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  638) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  639) 	 * Equal block_gen and start_gen implies we are between
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  640) 	 * recover_prep and recover_done callbacks, which means
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  641) 	 * dlm recovery is in progress and dlm locking is blocked.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  642) 	 * There's no point trying to do any work until recover_done.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  643) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  644) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  645) 	if (block_gen == start_gen)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  646) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  647) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  648) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  649) 	 * Propagate recover_submit[] and recover_result[] to lvb:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  650) 	 * dlm_recoverd adds to recover_submit[] jids needing recovery
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  651) 	 * gfs2_recover adds to recover_result[] journal recovery results
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  652) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  653) 	 * set lvb bit for jids in recover_submit[] if the lvb has not
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  654) 	 * yet been updated for the generation of the failure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  655) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  656) 	 * clear lvb bit for jids in recover_result[] if the result of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  657) 	 * the journal recovery is SUCCESS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  658) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  659) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  660) 	error = control_lock(sdp, DLM_LOCK_EX, DLM_LKF_CONVERT|DLM_LKF_VALBLK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  661) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  662) 		fs_err(sdp, "control lock EX error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  663) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  664) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  665) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  666) 	control_lvb_read(ls, &lvb_gen, ls->ls_lvb_bits);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  667) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  668) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  669) 	if (block_gen != ls->ls_recover_block ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  670) 	    start_gen != ls->ls_recover_start) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  671) 		fs_info(sdp, "recover generation %u block1 %u %u\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  672) 			start_gen, block_gen, ls->ls_recover_block);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  673) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  674) 		control_lock(sdp, DLM_LOCK_NL, DLM_LKF_CONVERT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  675) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  676) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  677) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  678) 	recover_size = ls->ls_recover_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  679) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  680) 	if (lvb_gen <= start_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  681) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  682) 		 * Clear lvb bits for jids we've successfully recovered.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  683) 		 * Because all nodes attempt to recover failed journals,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  684) 		 * a journal can be recovered multiple times successfully
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  685) 		 * in succession.  Only the first will really do recovery,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  686) 		 * the others find it clean, but still report a successful
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  687) 		 * recovery.  So, another node may have already recovered
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  688) 		 * the jid and cleared the lvb bit for it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  689) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  690) 		for (i = 0; i < recover_size; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  691) 			if (ls->ls_recover_result[i] != LM_RD_SUCCESS)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  692) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  693) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  694) 			ls->ls_recover_result[i] = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  695) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  696) 			if (!test_bit_le(i, ls->ls_lvb_bits + JID_BITMAP_OFFSET))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  697) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  698) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  699) 			__clear_bit_le(i, ls->ls_lvb_bits + JID_BITMAP_OFFSET);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  700) 			write_lvb = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  701) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  702) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  703) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  704) 	if (lvb_gen == start_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  705) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  706) 		 * Failed slots before start_gen are already set in lvb.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  707) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  708) 		for (i = 0; i < recover_size; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  709) 			if (!ls->ls_recover_submit[i])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  710) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  711) 			if (ls->ls_recover_submit[i] < lvb_gen)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  712) 				ls->ls_recover_submit[i] = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  713) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  714) 	} else if (lvb_gen < start_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  715) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  716) 		 * Failed slots before start_gen are not yet set in lvb.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  717) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  718) 		for (i = 0; i < recover_size; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  719) 			if (!ls->ls_recover_submit[i])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  720) 				continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  721) 			if (ls->ls_recover_submit[i] < start_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  722) 				ls->ls_recover_submit[i] = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  723) 				__set_bit_le(i, ls->ls_lvb_bits + JID_BITMAP_OFFSET);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  724) 			}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  725) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  726) 		/* even if there are no bits to set, we need to write the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  727) 		   latest generation to the lvb */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  728) 		write_lvb = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  729) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  730) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  731) 		 * we should be getting a recover_done() for lvb_gen soon
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  732) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  733) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  734) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  735) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  736) 	if (write_lvb) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  737) 		control_lvb_write(ls, start_gen, ls->ls_lvb_bits);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  738) 		flags = DLM_LKF_CONVERT | DLM_LKF_VALBLK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  739) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  740) 		flags = DLM_LKF_CONVERT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  741) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  742) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  743) 	error = control_lock(sdp, DLM_LOCK_NL, flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  744) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  745) 		fs_err(sdp, "control lock NL error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  746) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  747) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  748) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  749) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  750) 	 * Everyone will see jid bits set in the lvb, run gfs2_recover_set(),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  751) 	 * and clear a jid bit in the lvb if the recovery is a success.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  752) 	 * Eventually all journals will be recovered, all jid bits will
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  753) 	 * be cleared in the lvb, and everyone will clear BLOCK_LOCKS.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  754) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  755) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  756) 	for (i = 0; i < recover_size; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  757) 		if (test_bit_le(i, ls->ls_lvb_bits + JID_BITMAP_OFFSET)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  758) 			fs_info(sdp, "recover generation %u jid %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  759) 				start_gen, i);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  760) 			gfs2_recover_set(sdp, i);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  761) 			recover_set++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  762) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  763) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  764) 	if (recover_set)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  765) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  766) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  767) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  768) 	 * No more jid bits set in lvb, all recovery is done, unblock locks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  769) 	 * (unless a new recover_prep callback has occured blocking locks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  770) 	 * again while working above)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  771) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  772) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  773) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  774) 	if (ls->ls_recover_block == block_gen &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  775) 	    ls->ls_recover_start == start_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  776) 		clear_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  777) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  778) 		fs_info(sdp, "recover generation %u done\n", start_gen);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  779) 		gfs2_glock_thaw(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  780) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  781) 		fs_info(sdp, "recover generation %u block2 %u %u\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  782) 			start_gen, block_gen, ls->ls_recover_block);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  783) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  784) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  785) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  786) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  787) static int control_mount(struct gfs2_sbd *sdp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  788) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  789) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  790) 	uint32_t start_gen, block_gen, mount_gen, lvb_gen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  791) 	int mounted_mode;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  792) 	int retries = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  793) 	int error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  794) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  795) 	memset(&ls->ls_mounted_lksb, 0, sizeof(struct dlm_lksb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  796) 	memset(&ls->ls_control_lksb, 0, sizeof(struct dlm_lksb));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  797) 	memset(&ls->ls_control_lvb, 0, GDLM_LVB_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  798) 	ls->ls_control_lksb.sb_lvbptr = ls->ls_control_lvb;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  799) 	init_completion(&ls->ls_sync_wait);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  800) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  801) 	set_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  802) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  803) 	error = control_lock(sdp, DLM_LOCK_NL, DLM_LKF_VALBLK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  804) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  805) 		fs_err(sdp, "control_mount control_lock NL error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  806) 		return error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  807) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  808) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  809) 	error = mounted_lock(sdp, DLM_LOCK_NL, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  810) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  811) 		fs_err(sdp, "control_mount mounted_lock NL error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  812) 		control_unlock(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  813) 		return error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  814) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  815) 	mounted_mode = DLM_LOCK_NL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  816) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  817) restart:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  818) 	if (retries++ && signal_pending(current)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  819) 		error = -EINTR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  820) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  821) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  822) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  823) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  824) 	 * We always start with both locks in NL. control_lock is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  825) 	 * demoted to NL below so we don't need to do it here.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  826) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  827) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  828) 	if (mounted_mode != DLM_LOCK_NL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  829) 		error = mounted_lock(sdp, DLM_LOCK_NL, DLM_LKF_CONVERT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  830) 		if (error)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  831) 			goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  832) 		mounted_mode = DLM_LOCK_NL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  833) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  834) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  835) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  836) 	 * Other nodes need to do some work in dlm recovery and gfs2_control
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  837) 	 * before the recover_done and control_lock will be ready for us below.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  838) 	 * A delay here is not required but often avoids having to retry.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  839) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  840) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  841) 	msleep_interruptible(500);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  842) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  843) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  844) 	 * Acquire control_lock in EX and mounted_lock in either EX or PR.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  845) 	 * control_lock lvb keeps track of any pending journal recoveries.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  846) 	 * mounted_lock indicates if any other nodes have the fs mounted.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  847) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  848) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  849) 	error = control_lock(sdp, DLM_LOCK_EX, DLM_LKF_CONVERT|DLM_LKF_NOQUEUE|DLM_LKF_VALBLK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  850) 	if (error == -EAGAIN) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  851) 		goto restart;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  852) 	} else if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  853) 		fs_err(sdp, "control_mount control_lock EX error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  854) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  855) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  856) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  857) 	/**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  858) 	 * If we're a spectator, we don't want to take the lock in EX because
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  859) 	 * we cannot do the first-mount responsibility it implies: recovery.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  860) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  861) 	if (sdp->sd_args.ar_spectator)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  862) 		goto locks_done;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  863) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  864) 	error = mounted_lock(sdp, DLM_LOCK_EX, DLM_LKF_CONVERT|DLM_LKF_NOQUEUE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  865) 	if (!error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  866) 		mounted_mode = DLM_LOCK_EX;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  867) 		goto locks_done;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  868) 	} else if (error != -EAGAIN) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  869) 		fs_err(sdp, "control_mount mounted_lock EX error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  870) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  871) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  872) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  873) 	error = mounted_lock(sdp, DLM_LOCK_PR, DLM_LKF_CONVERT|DLM_LKF_NOQUEUE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  874) 	if (!error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  875) 		mounted_mode = DLM_LOCK_PR;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  876) 		goto locks_done;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  877) 	} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  878) 		/* not even -EAGAIN should happen here */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  879) 		fs_err(sdp, "control_mount mounted_lock PR error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  880) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  881) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  882) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  883) locks_done:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  884) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  885) 	 * If we got both locks above in EX, then we're the first mounter.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  886) 	 * If not, then we need to wait for the control_lock lvb to be
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  887) 	 * updated by other mounted nodes to reflect our mount generation.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  888) 	 *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  889) 	 * In simple first mounter cases, first mounter will see zero lvb_gen,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  890) 	 * but in cases where all existing nodes leave/fail before mounting
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  891) 	 * nodes finish control_mount, then all nodes will be mounting and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  892) 	 * lvb_gen will be non-zero.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  893) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  894) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  895) 	control_lvb_read(ls, &lvb_gen, ls->ls_lvb_bits);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  896) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  897) 	if (lvb_gen == 0xFFFFFFFF) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  898) 		/* special value to force mount attempts to fail */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  899) 		fs_err(sdp, "control_mount control_lock disabled\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  900) 		error = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  901) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  902) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  903) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  904) 	if (mounted_mode == DLM_LOCK_EX) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  905) 		/* first mounter, keep both EX while doing first recovery */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  906) 		spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  907) 		clear_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  908) 		set_bit(DFL_MOUNT_DONE, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  909) 		set_bit(DFL_FIRST_MOUNT, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  910) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  911) 		fs_info(sdp, "first mounter control generation %u\n", lvb_gen);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  912) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  913) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  914) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  915) 	error = control_lock(sdp, DLM_LOCK_NL, DLM_LKF_CONVERT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  916) 	if (error)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  917) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  918) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  919) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  920) 	 * We are not first mounter, now we need to wait for the control_lock
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  921) 	 * lvb generation to be >= the generation from our first recover_done
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  922) 	 * and all lvb bits to be clear (no pending journal recoveries.)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  923) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  924) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  925) 	if (!all_jid_bits_clear(ls->ls_lvb_bits)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  926) 		/* journals need recovery, wait until all are clear */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  927) 		fs_info(sdp, "control_mount wait for journal recovery\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  928) 		goto restart;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  929) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  930) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  931) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  932) 	block_gen = ls->ls_recover_block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  933) 	start_gen = ls->ls_recover_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  934) 	mount_gen = ls->ls_recover_mount;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  935) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  936) 	if (lvb_gen < mount_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  937) 		/* wait for mounted nodes to update control_lock lvb to our
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  938) 		   generation, which might include new recovery bits set */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  939) 		if (sdp->sd_args.ar_spectator) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  940) 			fs_info(sdp, "Recovery is required. Waiting for a "
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  941) 				"non-spectator to mount.\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  942) 			msleep_interruptible(1000);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  943) 		} else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  944) 			fs_info(sdp, "control_mount wait1 block %u start %u "
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  945) 				"mount %u lvb %u flags %lx\n", block_gen,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  946) 				start_gen, mount_gen, lvb_gen,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  947) 				ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  948) 		}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  949) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  950) 		goto restart;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  951) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  952) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  953) 	if (lvb_gen != start_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  954) 		/* wait for mounted nodes to update control_lock lvb to the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  955) 		   latest recovery generation */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  956) 		fs_info(sdp, "control_mount wait2 block %u start %u mount %u "
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  957) 			"lvb %u flags %lx\n", block_gen, start_gen, mount_gen,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  958) 			lvb_gen, ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  959) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  960) 		goto restart;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  961) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  962) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  963) 	if (block_gen == start_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  964) 		/* dlm recovery in progress, wait for it to finish */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  965) 		fs_info(sdp, "control_mount wait3 block %u start %u mount %u "
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  966) 			"lvb %u flags %lx\n", block_gen, start_gen, mount_gen,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  967) 			lvb_gen, ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  968) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  969) 		goto restart;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  970) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  971) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  972) 	clear_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  973) 	set_bit(DFL_MOUNT_DONE, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  974) 	memset(ls->ls_recover_submit, 0, ls->ls_recover_size*sizeof(uint32_t));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  975) 	memset(ls->ls_recover_result, 0, ls->ls_recover_size*sizeof(uint32_t));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  976) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  977) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  978) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  979) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  980) 	mounted_unlock(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  981) 	control_unlock(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  982) 	return error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  983) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  984) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  985) static int control_first_done(struct gfs2_sbd *sdp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  986) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  987) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  988) 	uint32_t start_gen, block_gen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  989) 	int error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  990) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  991) restart:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  992) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  993) 	start_gen = ls->ls_recover_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  994) 	block_gen = ls->ls_recover_block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  995) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  996) 	if (test_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  997) 	    !test_bit(DFL_MOUNT_DONE, &ls->ls_recover_flags) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  998) 	    !test_bit(DFL_FIRST_MOUNT, &ls->ls_recover_flags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300  999) 		/* sanity check, should not happen */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1000) 		fs_err(sdp, "control_first_done start %u block %u flags %lx\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1001) 		       start_gen, block_gen, ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1002) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1003) 		control_unlock(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1004) 		return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1005) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1006) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1007) 	if (start_gen == block_gen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1008) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1009) 		 * Wait for the end of a dlm recovery cycle to switch from
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1010) 		 * first mounter recovery.  We can ignore any recover_slot
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1011) 		 * callbacks between the recover_prep and next recover_done
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1012) 		 * because we are still the first mounter and any failed nodes
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1013) 		 * have not fully mounted, so they don't need recovery.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1014) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1015) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1016) 		fs_info(sdp, "control_first_done wait gen %u\n", start_gen);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1017) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1018) 		wait_on_bit(&ls->ls_recover_flags, DFL_DLM_RECOVERY,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1019) 			    TASK_UNINTERRUPTIBLE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1020) 		goto restart;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1021) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1022) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1023) 	clear_bit(DFL_FIRST_MOUNT, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1024) 	set_bit(DFL_FIRST_MOUNT_DONE, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1025) 	memset(ls->ls_recover_submit, 0, ls->ls_recover_size*sizeof(uint32_t));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1026) 	memset(ls->ls_recover_result, 0, ls->ls_recover_size*sizeof(uint32_t));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1027) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1028) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1029) 	memset(ls->ls_lvb_bits, 0, GDLM_LVB_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1030) 	control_lvb_write(ls, start_gen, ls->ls_lvb_bits);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1031) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1032) 	error = mounted_lock(sdp, DLM_LOCK_PR, DLM_LKF_CONVERT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1033) 	if (error)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1034) 		fs_err(sdp, "control_first_done mounted PR error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1035) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1036) 	error = control_lock(sdp, DLM_LOCK_NL, DLM_LKF_CONVERT|DLM_LKF_VALBLK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1037) 	if (error)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1038) 		fs_err(sdp, "control_first_done control NL error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1039) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1040) 	return error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1041) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1042) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1043) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1044)  * Expand static jid arrays if necessary (by increments of RECOVER_SIZE_INC)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1045)  * to accomodate the largest slot number.  (NB dlm slot numbers start at 1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1046)  * gfs2 jids start at 0, so jid = slot - 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1047)  */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1048) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1049) #define RECOVER_SIZE_INC 16
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1050) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1051) static int set_recover_size(struct gfs2_sbd *sdp, struct dlm_slot *slots,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1052) 			    int num_slots)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1053) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1054) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1055) 	uint32_t *submit = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1056) 	uint32_t *result = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1057) 	uint32_t old_size, new_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1058) 	int i, max_jid;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1059) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1060) 	if (!ls->ls_lvb_bits) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1061) 		ls->ls_lvb_bits = kzalloc(GDLM_LVB_SIZE, GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1062) 		if (!ls->ls_lvb_bits)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1063) 			return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1064) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1065) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1066) 	max_jid = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1067) 	for (i = 0; i < num_slots; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1068) 		if (max_jid < slots[i].slot - 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1069) 			max_jid = slots[i].slot - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1070) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1071) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1072) 	old_size = ls->ls_recover_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1073) 	new_size = old_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1074) 	while (new_size < max_jid + 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1075) 		new_size += RECOVER_SIZE_INC;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1076) 	if (new_size == old_size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1077) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1078) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1079) 	submit = kcalloc(new_size, sizeof(uint32_t), GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1080) 	result = kcalloc(new_size, sizeof(uint32_t), GFP_NOFS);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1081) 	if (!submit || !result) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1082) 		kfree(submit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1083) 		kfree(result);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1084) 		return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1085) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1086) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1087) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1088) 	memcpy(submit, ls->ls_recover_submit, old_size * sizeof(uint32_t));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1089) 	memcpy(result, ls->ls_recover_result, old_size * sizeof(uint32_t));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1090) 	kfree(ls->ls_recover_submit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1091) 	kfree(ls->ls_recover_result);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1092) 	ls->ls_recover_submit = submit;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1093) 	ls->ls_recover_result = result;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1094) 	ls->ls_recover_size = new_size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1095) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1096) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1097) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1098) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1099) static void free_recover_size(struct lm_lockstruct *ls)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1100) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1101) 	kfree(ls->ls_lvb_bits);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1102) 	kfree(ls->ls_recover_submit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1103) 	kfree(ls->ls_recover_result);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1104) 	ls->ls_recover_submit = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1105) 	ls->ls_recover_result = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1106) 	ls->ls_recover_size = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1107) 	ls->ls_lvb_bits = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1108) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1109) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1110) /* dlm calls before it does lock recovery */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1111) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1112) static void gdlm_recover_prep(void *arg)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1113) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1114) 	struct gfs2_sbd *sdp = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1115) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1116) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1117) 	if (gfs2_withdrawn(sdp)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1118) 		fs_err(sdp, "recover_prep ignored due to withdraw.\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1119) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1120) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1121) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1122) 	ls->ls_recover_block = ls->ls_recover_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1123) 	set_bit(DFL_DLM_RECOVERY, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1124) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1125) 	if (!test_bit(DFL_MOUNT_DONE, &ls->ls_recover_flags) ||
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1126) 	     test_bit(DFL_FIRST_MOUNT, &ls->ls_recover_flags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1127) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1128) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1129) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1130) 	set_bit(DFL_BLOCK_LOCKS, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1131) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1132) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1133) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1134) /* dlm calls after recover_prep has been completed on all lockspace members;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1135)    identifies slot/jid of failed member */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1136) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1137) static void gdlm_recover_slot(void *arg, struct dlm_slot *slot)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1138) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1139) 	struct gfs2_sbd *sdp = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1140) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1141) 	int jid = slot->slot - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1142) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1143) 	if (gfs2_withdrawn(sdp)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1144) 		fs_err(sdp, "recover_slot jid %d ignored due to withdraw.\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1145) 		       jid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1146) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1147) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1148) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1149) 	if (ls->ls_recover_size < jid + 1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1150) 		fs_err(sdp, "recover_slot jid %d gen %u short size %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1151) 		       jid, ls->ls_recover_block, ls->ls_recover_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1152) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1153) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1154) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1155) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1156) 	if (ls->ls_recover_submit[jid]) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1157) 		fs_info(sdp, "recover_slot jid %d gen %u prev %u\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1158) 			jid, ls->ls_recover_block, ls->ls_recover_submit[jid]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1159) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1160) 	ls->ls_recover_submit[jid] = ls->ls_recover_block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1161) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1162) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1163) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1164) /* dlm calls after recover_slot and after it completes lock recovery */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1165) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1166) static void gdlm_recover_done(void *arg, struct dlm_slot *slots, int num_slots,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1167) 			      int our_slot, uint32_t generation)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1168) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1169) 	struct gfs2_sbd *sdp = arg;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1170) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1171) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1172) 	if (gfs2_withdrawn(sdp)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1173) 		fs_err(sdp, "recover_done ignored due to withdraw.\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1174) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1175) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1176) 	/* ensure the ls jid arrays are large enough */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1177) 	set_recover_size(sdp, slots, num_slots);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1178) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1179) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1180) 	ls->ls_recover_start = generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1181) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1182) 	if (!ls->ls_recover_mount) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1183) 		ls->ls_recover_mount = generation;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1184) 		ls->ls_jid = our_slot - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1185) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1186) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1187) 	if (!test_bit(DFL_UNMOUNT, &ls->ls_recover_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1188) 		queue_delayed_work(gfs2_control_wq, &sdp->sd_control_work, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1189) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1190) 	clear_bit(DFL_DLM_RECOVERY, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1191) 	smp_mb__after_atomic();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1192) 	wake_up_bit(&ls->ls_recover_flags, DFL_DLM_RECOVERY);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1193) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1194) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1195) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1196) /* gfs2_recover thread has a journal recovery result */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1197) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1198) static void gdlm_recovery_result(struct gfs2_sbd *sdp, unsigned int jid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1199) 				 unsigned int result)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1200) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1201) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1202) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1203) 	if (gfs2_withdrawn(sdp)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1204) 		fs_err(sdp, "recovery_result jid %d ignored due to withdraw.\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1205) 		       jid);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1206) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1207) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1208) 	if (test_bit(DFL_NO_DLM_OPS, &ls->ls_recover_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1209) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1210) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1211) 	/* don't care about the recovery of own journal during mount */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1212) 	if (jid == ls->ls_jid)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1213) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1214) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1215) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1216) 	if (test_bit(DFL_FIRST_MOUNT, &ls->ls_recover_flags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1217) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1218) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1219) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1220) 	if (ls->ls_recover_size < jid + 1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1221) 		fs_err(sdp, "recovery_result jid %d short size %d\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1222) 		       jid, ls->ls_recover_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1223) 		spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1224) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1225) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1226) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1227) 	fs_info(sdp, "recover jid %d result %s\n", jid,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1228) 		result == LM_RD_GAVEUP ? "busy" : "success");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1229) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1230) 	ls->ls_recover_result[jid] = result;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1231) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1232) 	/* GAVEUP means another node is recovering the journal; delay our
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1233) 	   next attempt to recover it, to give the other node a chance to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1234) 	   finish before trying again */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1235) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1236) 	if (!test_bit(DFL_UNMOUNT, &ls->ls_recover_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1237) 		queue_delayed_work(gfs2_control_wq, &sdp->sd_control_work,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1238) 				   result == LM_RD_GAVEUP ? HZ : 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1239) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1240) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1241) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1242) static const struct dlm_lockspace_ops gdlm_lockspace_ops = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1243) 	.recover_prep = gdlm_recover_prep,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1244) 	.recover_slot = gdlm_recover_slot,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1245) 	.recover_done = gdlm_recover_done,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1246) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1247) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1248) static int gdlm_mount(struct gfs2_sbd *sdp, const char *table)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1249) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1250) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1251) 	char cluster[GFS2_LOCKNAME_LEN];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1252) 	const char *fsname;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1253) 	uint32_t flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1254) 	int error, ops_result;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1255) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1256) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1257) 	 * initialize everything
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1258) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1259) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1260) 	INIT_DELAYED_WORK(&sdp->sd_control_work, gfs2_control_func);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1261) 	spin_lock_init(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1262) 	ls->ls_recover_flags = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1263) 	ls->ls_recover_mount = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1264) 	ls->ls_recover_start = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1265) 	ls->ls_recover_block = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1266) 	ls->ls_recover_size = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1267) 	ls->ls_recover_submit = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1268) 	ls->ls_recover_result = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1269) 	ls->ls_lvb_bits = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1270) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1271) 	error = set_recover_size(sdp, NULL, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1272) 	if (error)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1273) 		goto fail;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1274) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1275) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1276) 	 * prepare dlm_new_lockspace args
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1277) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1278) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1279) 	fsname = strchr(table, ':');
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1280) 	if (!fsname) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1281) 		fs_info(sdp, "no fsname found\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1282) 		error = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1283) 		goto fail_free;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1284) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1285) 	memset(cluster, 0, sizeof(cluster));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1286) 	memcpy(cluster, table, strlen(table) - strlen(fsname));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1287) 	fsname++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1288) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1289) 	flags = DLM_LSFL_FS | DLM_LSFL_NEWEXCL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1290) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1291) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1292) 	 * create/join lockspace
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1293) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1294) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1295) 	error = dlm_new_lockspace(fsname, cluster, flags, GDLM_LVB_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1296) 				  &gdlm_lockspace_ops, sdp, &ops_result,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1297) 				  &ls->ls_dlm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1298) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1299) 		fs_err(sdp, "dlm_new_lockspace error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1300) 		goto fail_free;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1301) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1302) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1303) 	if (ops_result < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1304) 		/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1305) 		 * dlm does not support ops callbacks,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1306) 		 * old dlm_controld/gfs_controld are used, try without ops.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1307) 		 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1308) 		fs_info(sdp, "dlm lockspace ops not used\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1309) 		free_recover_size(ls);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1310) 		set_bit(DFL_NO_DLM_OPS, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1311) 		return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1312) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1313) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1314) 	if (!test_bit(SDF_NOJOURNALID, &sdp->sd_flags)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1315) 		fs_err(sdp, "dlm lockspace ops disallow jid preset\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1316) 		error = -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1317) 		goto fail_release;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1318) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1319) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1320) 	/*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1321) 	 * control_mount() uses control_lock to determine first mounter,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1322) 	 * and for later mounts, waits for any recoveries to be cleared.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1323) 	 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1324) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1325) 	error = control_mount(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1326) 	if (error) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1327) 		fs_err(sdp, "mount control error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1328) 		goto fail_release;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1329) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1330) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1331) 	ls->ls_first = !!test_bit(DFL_FIRST_MOUNT, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1332) 	clear_bit(SDF_NOJOURNALID, &sdp->sd_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1333) 	smp_mb__after_atomic();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1334) 	wake_up_bit(&sdp->sd_flags, SDF_NOJOURNALID);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1335) 	return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1336) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1337) fail_release:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1338) 	dlm_release_lockspace(ls->ls_dlm, 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1339) fail_free:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1340) 	free_recover_size(ls);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1341) fail:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1342) 	return error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1343) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1344) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1345) static void gdlm_first_done(struct gfs2_sbd *sdp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1346) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1347) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1348) 	int error;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1349) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1350) 	if (test_bit(DFL_NO_DLM_OPS, &ls->ls_recover_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1351) 		return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1352) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1353) 	error = control_first_done(sdp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1354) 	if (error)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1355) 		fs_err(sdp, "mount first_done error %d\n", error);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1356) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1357) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1358) static void gdlm_unmount(struct gfs2_sbd *sdp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1359) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1360) 	struct lm_lockstruct *ls = &sdp->sd_lockstruct;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1361) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1362) 	if (test_bit(DFL_NO_DLM_OPS, &ls->ls_recover_flags))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1363) 		goto release;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1364) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1365) 	/* wait for gfs2_control_wq to be done with this mount */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1366) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1367) 	spin_lock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1368) 	set_bit(DFL_UNMOUNT, &ls->ls_recover_flags);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1369) 	spin_unlock(&ls->ls_recover_spin);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1370) 	flush_delayed_work(&sdp->sd_control_work);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1371) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1372) 	/* mounted_lock and control_lock will be purged in dlm recovery */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1373) release:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1374) 	if (ls->ls_dlm) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1375) 		dlm_release_lockspace(ls->ls_dlm, 2);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1376) 		ls->ls_dlm = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1377) 	}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1378) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1379) 	free_recover_size(ls);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1380) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1381) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1382) static const match_table_t dlm_tokens = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1383) 	{ Opt_jid, "jid=%d"},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1384) 	{ Opt_id, "id=%d"},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1385) 	{ Opt_first, "first=%d"},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1386) 	{ Opt_nodir, "nodir=%d"},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1387) 	{ Opt_err, NULL },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1388) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1389) 
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1390) const struct lm_lockops gfs2_dlm_ops = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1391) 	.lm_proto_name = "lock_dlm",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1392) 	.lm_mount = gdlm_mount,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1393) 	.lm_first_done = gdlm_first_done,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1394) 	.lm_recovery_result = gdlm_recovery_result,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1395) 	.lm_unmount = gdlm_unmount,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1396) 	.lm_put_lock = gdlm_put_lock,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1397) 	.lm_lock = gdlm_lock,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1398) 	.lm_cancel = gdlm_cancel,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1399) 	.lm_tokens = &dlm_tokens,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1400) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1401)