| /* |
| * Copyright (C) 2001-2004 Sistina Software, Inc. All rights reserved. |
| * Copyright (C) 2004-2011 Red Hat, Inc. All rights reserved. |
| * |
| * This file is part of LVM2. |
| * |
| * This copyrighted material is made available to anyone wishing to use, |
| * modify, copy, or redistribute it subject to the terms and conditions |
| * of the GNU Lesser General Public License v.2.1. |
| * |
| * You should have received a copy of the GNU Lesser General Public License |
| * along with this program; if not, write to the Free Software Foundation, |
| * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| #ifndef _LVM_LOCKING_H |
| #define _LVM_LOCKING_H |
| |
| #include "uuid.h" |
| #include "config.h" |
| |
| struct logical_volume; |
| |
| int init_locking(int type, struct cmd_context *cmd, int suppress_messages); |
| void fin_locking(void); |
| void reset_locking(void); |
| int vg_write_lock_held(void); |
| int locking_is_clustered(void); |
| int locking_supports_remote_queries(void); |
| |
| #ifndef NODE_ALL |
| # define NODE_ALL "*" |
| # define NODE_LOCAL "." |
| # define NODE_REMOTE "^" |
| #endif |
| int cluster_lock_held(const char *vol, const char *node, int *exclusive); |
| |
| /* |
| * LCK_VG: |
| * Lock/unlock on-disk volume group data. |
| * Use VG_ORPHANS to lock all orphan PVs. |
| * Use VG_GLOBAL as a global lock and to wipe the internal cache. |
| * char *vol holds volume group name. |
| * Set LCK_CACHE flag when manipulating 'vol' metadata in the internal cache. |
| * (Like commit, revert or invalidate metadata.) |
| * If more than one lock needs to be held simultaneously, they must be |
| * acquired in alphabetical order of 'vol' (to avoid deadlocks), with |
| * VG_ORPHANS last. |
| * |
| * Use VG_SYNC_NAMES to ensure /dev is up-to-date for example, with udev, |
| * by waiting for any asynchronous events issued to have completed. |
| * |
| * LCK_LV: |
| * Lock/unlock an individual logical volume |
| * char *vol holds lvid |
| */ |
| int lock_vol(struct cmd_context *cmd, const char *vol, uint32_t flags, const struct logical_volume *lv); |
| |
| /* |
| * Internal locking representation. |
| * LCK_VG: Uses prefix V_ unless the vol begins with # (i.e. #global or #orphans) |
| * or the LCK_CACHE flag is set when it uses the prefix P_. |
| * If LCK_CACHE is set, we do not take out a real lock. |
| * NB In clustered situations, LCK_CACHE is not propagated directly to remote nodes. |
| * (It can be deduced from lock name.) |
| */ |
| |
| /* |
| * Does the LVM1 driver have this VG active? |
| */ |
| int check_lvm1_vg_inactive(struct cmd_context *cmd, const char *vgname); |
| |
| /* |
| * Lock type - these numbers are the same as VMS and the IBM DLM |
| */ |
| #define LCK_TYPE_MASK 0x00000007U |
| |
| #define LCK_NULL 0x00000000U /* LCK$_NLMODE (Deactivate) */ |
| #define LCK_READ 0x00000001U /* LCK$_CRMODE (Activate) */ |
| /* LCK$_CWMODE */ |
| #define LCK_PREAD 0x00000003U /* LCK$_PRMODE */ |
| #define LCK_WRITE 0x00000004U /* LCK$_PWMODE (Suspend) */ |
| #define LCK_EXCL 0x00000005U /* LCK$_EXMODE (Exclusive) */ |
| #define LCK_UNLOCK 0x00000006U /* This is ours (Resume) */ |
| |
| /* |
| * Lock flags - these numbers are the same as DLM |
| */ |
| #define LCKF_NOQUEUE 0x00000001U /* LKF$_NOQUEUE */ |
| #define LCKF_CONVERT 0x00000004U /* LKF$_CONVERT */ |
| |
| /* |
| * Lock scope |
| */ |
| #define LCK_SCOPE_MASK 0x00001008U |
| #define LCK_VG 0x00000000U /* Volume Group */ |
| #define LCK_LV 0x00000008U /* Logical Volume */ |
| #define LCK_ACTIVATION 0x00001000U /* Activation */ |
| |
| /* |
| * Lock bits. |
| * Bottom 8 bits except LCK_LOCAL form args[0] in cluster comms. |
| */ |
| #define LCK_NONBLOCK 0x00000010U /* Don't block waiting for lock? */ |
| #define LCK_HOLD 0x00000020U /* Hold lock when lock_vol returns? */ |
| #define LCK_CLUSTER_VG 0x00000080U /* VG is clustered */ |
| |
| #define LCK_LOCAL 0x00000040U /* Don't propagate to other nodes */ |
| #define LCK_REMOTE 0x00000800U /* Propagate to remote nodes only */ |
| #define LCK_CACHE 0x00000100U /* Operation on cache only using P_ lock */ |
| #define LCK_ORIGIN_ONLY 0x00000200U /* Operation should bypass any snapshots */ |
| #define LCK_REVERT 0x00000400U /* Revert any incomplete change */ |
| |
| /* |
| * Additional lock bits for cluster communication via args[1] |
| */ |
| #define LCK_PARTIAL_MODE 0x01 /* Partial activation? */ |
| #define LCK_MIRROR_NOSYNC_MODE 0x02 /* Mirrors don't require sync */ |
| #define LCK_DMEVENTD_MONITOR_MODE 0x04 /* Register with dmeventd */ |
| |
| /* Not yet used. */ |
| #define LCK_CONVERT_MODE 0x08 /* Convert existing lock */ |
| |
| #define LCK_TEST_MODE 0x10 /* Test mode: No activation */ |
| #define LCK_ORIGIN_ONLY_MODE 0x20 /* Same as above */ |
| #define LCK_DMEVENTD_MONITOR_IGNORE 0x40 /* Whether to ignore dmeventd */ |
| #define LCK_REVERT_MODE 0x80 /* Remove inactive tables */ |
| |
| /* |
| * Special cases of VG locks. |
| */ |
| #define VG_ORPHANS "#orphans" |
| #define VG_GLOBAL "#global" |
| #define VG_SYNC_NAMES "#sync_names" |
| |
| /* |
| * Common combinations |
| */ |
| #define LCK_NONE (LCK_VG | LCK_NULL) |
| |
| #define LCK_ACTIVATE_LOCK (LCK_ACTIVATION | LCK_WRITE | LCK_HOLD) |
| #define LCK_ACTIVATE_UNLOCK (LCK_ACTIVATION | LCK_UNLOCK) |
| |
| #define LCK_VG_READ (LCK_VG | LCK_READ | LCK_HOLD) |
| #define LCK_VG_WRITE (LCK_VG | LCK_WRITE | LCK_HOLD) |
| #define LCK_VG_UNLOCK (LCK_VG | LCK_UNLOCK) |
| #define LCK_VG_DROP_CACHE (LCK_VG | LCK_WRITE | LCK_CACHE) |
| |
| /* FIXME: LCK_HOLD abused here */ |
| #define LCK_VG_COMMIT (LCK_VG | LCK_WRITE | LCK_CACHE | LCK_HOLD) |
| #define LCK_VG_REVERT (LCK_VG | LCK_READ | LCK_CACHE | LCK_HOLD) |
| |
| #define LCK_VG_BACKUP (LCK_VG | LCK_CACHE) |
| |
| #define LCK_VG_SYNC (LCK_NONE | LCK_CACHE) |
| #define LCK_VG_SYNC_LOCAL (LCK_NONE | LCK_CACHE | LCK_LOCAL) |
| |
| #define LCK_LV_EXCLUSIVE (LCK_LV | LCK_EXCL) |
| #define LCK_LV_SUSPEND (LCK_LV | LCK_WRITE) |
| #define LCK_LV_RESUME (LCK_LV | LCK_UNLOCK) |
| #define LCK_LV_ACTIVATE (LCK_LV | LCK_READ) |
| #define LCK_LV_DEACTIVATE (LCK_LV | LCK_NULL) |
| |
| #define LCK_MASK (LCK_TYPE_MASK | LCK_SCOPE_MASK) |
| |
| #define LCK_LV_CLUSTERED(lv) \ |
| (vg_is_clustered((lv)->vg) ? LCK_CLUSTER_VG : 0) |
| |
| #define lock_lv_vol(cmd, lv, flags) \ |
| (find_replicator_vgs((lv)) ? \ |
| lock_vol(cmd, (lv)->lvid.s, flags | LCK_LV_CLUSTERED(lv), lv) : \ |
| 0) |
| |
| /* |
| * Activation locks are wrapped around activation commands that have to |
| * be processed atomically one-at-a-time. |
| * If a VG WRITE lock is held, an activation lock is redundant. |
| * |
| * FIXME Test and support this for thin and cache types. |
| * FIXME Add cluster support. |
| */ |
| #define lv_supports_activation_locking(lv) (!vg_is_clustered((lv)->vg) && !lv_is_thin_type(lv) && !lv_is_cache_type(lv)) |
| #define lock_activation(cmd, lv) (vg_write_lock_held() && lv_supports_activation_locking(lv) ? 1 : lock_vol(cmd, (lv)->lvid.s, LCK_ACTIVATE_LOCK, lv)) |
| #define unlock_activation(cmd, lv) (vg_write_lock_held() && lv_supports_activation_locking(lv) ? 1 : lock_vol(cmd, (lv)->lvid.s, LCK_ACTIVATE_UNLOCK, lv)) |
| |
| /* |
| * Place temporary exclusive 'activation' lock around an LV locking operation |
| * to serialise it. |
| */ |
| #define lock_lv_vol_serially(cmd, lv, flags) \ |
| ({ \ |
| int rr = 0; \ |
| \ |
| if (lock_activation((cmd), (lv))) { \ |
| rr = lock_lv_vol((cmd), (lv), (flags)); \ |
| unlock_activation((cmd), (lv)); \ |
| } \ |
| rr; \ |
| }) |
| |
| #define unlock_vg(cmd, vg, vol) \ |
| do { \ |
| if (vg && !lvmetad_vg_update_finish(vg)) \ |
| stack; \ |
| if (is_real_vg(vol) && !sync_dev_names(cmd)) \ |
| stack; \ |
| if (!lock_vol(cmd, vol, LCK_VG_UNLOCK, NULL)) \ |
| stack; \ |
| } while (0) |
| #define unlock_and_release_vg(cmd, vg, vol) \ |
| do { \ |
| unlock_vg(cmd, vg, vol); \ |
| release_vg(vg); \ |
| } while (0) |
| |
| #define resume_lv(cmd, lv) \ |
| ({ \ |
| int rr = lock_lv_vol((cmd), (lv), LCK_LV_RESUME); \ |
| unlock_activation((cmd), (lv)); \ |
| rr; \ |
| }) |
| #define resume_lv_origin(cmd, lv) lock_lv_vol(cmd, lv, LCK_LV_RESUME | LCK_ORIGIN_ONLY) |
| #define revert_lv(cmd, lv) \ |
| ({ \ |
| int rr = lock_lv_vol((cmd), (lv), LCK_LV_RESUME | LCK_REVERT); \ |
| \ |
| unlock_activation((cmd), (lv)); \ |
| rr; \ |
| }) |
| #define suspend_lv(cmd, lv) \ |
| (lock_activation((cmd), (lv)) ? lock_lv_vol((cmd), (lv), LCK_LV_SUSPEND | LCK_HOLD) : 0) |
| #define suspend_lv_origin(cmd, lv) lock_lv_vol(cmd, lv, LCK_LV_SUSPEND | LCK_HOLD | LCK_ORIGIN_ONLY) |
| #define deactivate_lv(cmd, lv) lock_lv_vol_serially(cmd, lv, LCK_LV_DEACTIVATE) |
| |
| #define activate_lv(cmd, lv) lock_lv_vol_serially(cmd, lv, LCK_LV_ACTIVATE | LCK_HOLD) |
| #define activate_lv_excl_local(cmd, lv) \ |
| lock_lv_vol_serially(cmd, lv, LCK_LV_EXCLUSIVE | LCK_HOLD | LCK_LOCAL) |
| #define activate_lv_excl_remote(cmd, lv) \ |
| lock_lv_vol(cmd, lv, LCK_LV_EXCLUSIVE | LCK_HOLD | LCK_REMOTE) |
| |
| struct logical_volume; |
| int activate_lv_excl(struct cmd_context *cmd, struct logical_volume *lv); |
| |
| #define activate_lv_local(cmd, lv) \ |
| lock_lv_vol_serially(cmd, lv, LCK_LV_ACTIVATE | LCK_HOLD | LCK_LOCAL) |
| #define deactivate_lv_local(cmd, lv) \ |
| lock_lv_vol_serially(cmd, lv, LCK_LV_DEACTIVATE | LCK_LOCAL) |
| #define drop_cached_metadata(vg) \ |
| lock_vol((vg)->cmd, (vg)->name, LCK_VG_DROP_CACHE, NULL) |
| #define remote_commit_cached_metadata(vg) \ |
| lock_vol((vg)->cmd, (vg)->name, LCK_VG_COMMIT, NULL) |
| #define remote_revert_cached_metadata(vg) \ |
| lock_vol((vg)->cmd, (vg)->name, LCK_VG_REVERT, NULL) |
| #define remote_backup_metadata(vg) \ |
| lock_vol((vg)->cmd, (vg)->name, LCK_VG_BACKUP, NULL) |
| |
| int sync_local_dev_names(struct cmd_context* cmd); |
| int sync_dev_names(struct cmd_context* cmd); |
| |
| /* Process list of LVs */ |
| struct volume_group; |
| int suspend_lvs(struct cmd_context *cmd, struct dm_list *lvs, |
| struct volume_group *vg_to_revert); |
| int resume_lvs(struct cmd_context *cmd, struct dm_list *lvs); |
| int revert_lvs(struct cmd_context *cmd, struct dm_list *lvs); |
| int activate_lvs(struct cmd_context *cmd, struct dm_list *lvs, unsigned exclusive); |
| |
| #endif |