/* * Copyright (C) 2002-2004 Sistina Software, Inc. All rights reserved. * Copyright (C) 2004-2018 Red Hat, Inc. All rights reserved. * * This file is part of LVM2. * * This copyrighted material is made available to anyone wishing to use, * modify, copy, or redistribute it subject to the terms and conditions * of the GNU Lesser General Public License v.2.1. * * You should have received a copy of the GNU Lesser General Public License * along with this program; if not, write to the Free Software Foundation, * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "lib/misc/lib.h" #include "dev_manager.h" #include "lib/misc/lvm-string.h" #include "fs.h" #include "lib/config/defaults.h" #include "lib/metadata/segtype.h" #include "lib/display/display.h" #include "lib/commands/toolcontext.h" #include "lib/activate/targets.h" #include "lib/config/config.h" #include "lib/activate/activate.h" #include "lib/misc/lvm-exec.h" #include "lib/datastruct/str_list.h" #include "lib/misc/lvm-signal.h" #include #include #define MAX_TARGET_PARAMSIZE 50000 #define LVM_UDEV_NOSCAN_FLAG DM_SUBSYSTEM_UDEV_FLAG0 #define CRYPT_TEMP "CRYPT-TEMP" #define CRYPT_SUBDEV "CRYPT-SUBDEV" #define STRATIS "stratis-" typedef enum { PRELOAD, ACTIVATE, DEACTIVATE, SUSPEND, SUSPEND_WITH_LOCKFS, CLEAN } action_t; /* This list must match lib/misc/lvm-string.c:build_dm_uuid(). */ const char *uuid_suffix_list[] = { "pool", "cdata", "cmeta", "cvol", "tdata", "tmeta", "vdata", "vpool", "imeta", NULL}; struct dlid_list { struct dm_list list; const char *dlid; const struct logical_volume *lv; }; struct dev_manager { struct dm_pool *mem; struct cmd_context *cmd; void *target_state; uint32_t pvmove_mirror_count; int flush_required; int activation; /* building activation tree */ int suspend; /* building suspend tree */ unsigned track_external_lv_deps; unsigned track_pending_delete; unsigned track_pvmove_deps; const char *vg_name; }; struct lv_layer { const struct logical_volume *lv; const char *old_name; int visible_component; }; int read_only_lv(const struct logical_volume *lv, const struct lv_activate_opts *laopts, const char *layer) { if (layer && lv_is_cow(lv)) return 0; /* Keep snapshot's COW volume writable */ if (lv_is_raid_image(lv) || lv_is_raid_metadata(lv)) return 0; /* Keep RAID SubLvs writable */ if (!layer) { if (lv_is_thin_pool(lv) || lv_is_vdo_pool(lv)) return 1; } return (laopts->read_only || !(lv->status & LVM_WRITE)); } /* * Low level device-layer operations. * * Unless task is DM_DEVICE_TARGET_MSG, also calls dm_task_run() */ static struct dm_task *_setup_task_run(int task, struct dm_info *info, const char *name, const char *uuid, uint32_t *event_nr, uint32_t major, uint32_t minor, int with_open_count, int with_flush, int query_inactive) { char vsn[80]; unsigned maj, min; struct dm_task *dmt; if (!(dmt = dm_task_create(task))) return_NULL; if (name && !dm_task_set_name(dmt, name)) goto_out; if (uuid && *uuid && !dm_task_set_uuid(dmt, uuid)) goto_out; if (event_nr && !dm_task_set_event_nr(dmt, *event_nr)) goto_out; if (major && !dm_task_set_major_minor(dmt, major, minor, 1)) goto_out; if (activation_checks() && !dm_task_enable_checks(dmt)) goto_out; if (query_inactive && !dm_task_query_inactive_table(dmt)) { log_error("Failed to set query_inactive_table."); goto out; } if (!with_open_count && !dm_task_no_open_count(dmt)) log_warn("WARNING: Failed to disable open_count."); if (!with_flush && !dm_task_no_flush(dmt)) log_warn("WARNING: Failed to set no_flush."); switch (task) { case DM_DEVICE_TARGET_MSG: return dmt; /* TARGET_MSG needs more local tweaking before task_run() */ case DM_DEVICE_LIST: /* Use 'newuuid' only with DM version that supports it */ if (driver_version(vsn, sizeof(vsn)) && (sscanf(vsn, "%u.%u", &maj, &min) == 2) && (maj == 4 ? min >= 19 : maj > 4) && !dm_task_set_newuuid(dmt, " ")) // new uuid has no meaning here log_warn("WARNING: Failed to query uuid with LIST."); break; default: break; } if (!dm_task_run(dmt)) goto_out; if (info && !dm_task_get_info(dmt, info)) goto_out; return dmt; out: dm_task_destroy(dmt); return NULL; } static int _get_segment_status_from_target_params(const char *target_name, const char *params, const struct dm_info *dminfo, struct lv_seg_status *seg_status) { const struct lv_segment *seg = seg_status->seg; const struct segment_type *segtype = seg->segtype; seg_status->type = SEG_STATUS_UNKNOWN; /* Parsing failed */ /* Switch to snapshot segtype status logic for merging origin */ /* This is 'dynamic' decision, both states are valid */ if (lv_is_merging_origin(seg->lv)) { if (!strcmp(target_name, TARGET_NAME_SNAPSHOT_ORIGIN)) { seg_status->type = SEG_STATUS_NONE; return 1; /* Merge has not yet started */ } if (!strcmp(target_name, TARGET_NAME_SNAPSHOT_MERGE) && !(segtype = get_segtype_from_string(seg->lv->vg->cmd, TARGET_NAME_SNAPSHOT))) return_0; /* Merging, parse 'snapshot' status of merge progress */ } if (!params) { log_warn("WARNING: Cannot find matching %s segment for %s.", segtype->name, display_lvname(seg_status->seg->lv)); return 0; } /* Validate target_name segtype from DM table with lvm2 metadata segtype */ if (!lv_is_locked(seg->lv) && strcmp(segtype->name, target_name) && /* If kernel's type isn't an exact match is it compatible? */ (!segtype->ops->target_status_compatible || !segtype->ops->target_status_compatible(target_name))) { log_warn("WARNING: Detected %s segment type does not match expected type %s for %s.", target_name, segtype->name, display_lvname(seg_status->seg->lv)); return 0; } /* TODO: move into segtype method */ if (segtype_is_cache(segtype)) { if (!dm_get_status_cache(seg_status->mem, params, &(seg_status->cache))) return_0; seg_status->type = SEG_STATUS_CACHE; } else if (segtype_is_raid(segtype)) { if (!dm_get_status_raid(seg_status->mem, params, &seg_status->raid)) return_0; seg_status->type = SEG_STATUS_RAID; } else if (segtype_is_thin_volume(segtype)) { if (!dm_get_status_thin(seg_status->mem, params, &seg_status->thin)) return_0; seg_status->type = SEG_STATUS_THIN; } else if (segtype_is_thin_pool(segtype)) { if (!dm_get_status_thin_pool(seg_status->mem, params, &seg_status->thin_pool)) return_0; seg_status->type = SEG_STATUS_THIN_POOL; } else if (segtype_is_snapshot(segtype)) { if (!dm_get_status_snapshot(seg_status->mem, params, &seg_status->snapshot)) return_0; seg_status->type = SEG_STATUS_SNAPSHOT; } else if (segtype_is_vdo_pool(segtype)) { if (!parse_vdo_pool_status(seg_status->mem, seg->lv, params, dminfo, &seg_status->vdo_pool)) return_0; seg_status->type = SEG_STATUS_VDO_POOL; } else if (segtype_is_writecache(segtype)) { if (!dm_get_status_writecache(seg_status->mem, params, &(seg_status->writecache))) return_0; seg_status->type = SEG_STATUS_WRITECACHE; } else if (segtype_is_integrity(segtype)) { if (!dm_get_status_integrity(seg_status->mem, params, &(seg_status->integrity))) return_0; seg_status->type = SEG_STATUS_INTEGRITY; } else /* * TODO: Add support for other segment types too! * Status not supported */ seg_status->type = SEG_STATUS_NONE; return 1; } typedef enum { INFO, /* DM_DEVICE_INFO ioctl */ STATUS, /* DM_DEVICE_STATUS ioctl */ } info_type_t; /* Return length of segment depending on type and reshape_len */ static uint32_t _seg_len(const struct lv_segment *seg) { uint32_t reshape_len = seg_is_raid(seg) ? ((seg->area_count - seg->segtype->parity_devs) * seg->reshape_len) : 0; return seg->len - reshape_len; } static int _info_run(const char *dlid, struct dm_info *dminfo, uint32_t *read_ahead, struct lv_seg_status *seg_status, const char *name_check, int with_open_count, int with_read_ahead, uint32_t major, uint32_t minor) { int r = 0; struct dm_task *dmt; int dmtask; int with_flush; /* TODO: arg for _info_run */ void *target = NULL; uint64_t target_start, target_length, start, extent_size, length, length_crop = 0; char *target_name, *target_params; const char *devname; if (seg_status) { dmtask = DM_DEVICE_STATUS; with_flush = 0; } else { dmtask = DM_DEVICE_INFO; with_flush = 1; /* doesn't really matter */ } if (!(dmt = _setup_task_run(dmtask, dminfo, NULL, dlid, 0, major, minor, with_open_count, with_flush, 0))) return_0; if (name_check && dminfo->exists && (devname = dm_task_get_name(dmt)) && (strcmp(name_check, devname) != 0)) dminfo->exists = 0; /* mismatching name -> device does not exist */ if (with_read_ahead && read_ahead && dminfo->exists) { if (!dm_task_get_read_ahead(dmt, read_ahead)) goto_out; } else if (read_ahead) *read_ahead = DM_READ_AHEAD_NONE; /* Query status only for active device */ if (seg_status && dminfo->exists) { extent_size = length = seg_status->seg->lv->vg->extent_size; start = extent_size * seg_status->seg->le; length *= _seg_len(seg_status->seg); /* Uses max DM_THIN_MAX_METADATA_SIZE sectors for metadata device */ if (lv_is_thin_pool_metadata(seg_status->seg->lv) && (length > DM_THIN_MAX_METADATA_SIZE)) length_crop = DM_THIN_MAX_METADATA_SIZE; /* Uses virtual size with headers for VDO pool device */ if (lv_is_vdo_pool(seg_status->seg->lv)) length = get_vdo_pool_virtual_size(seg_status->seg); if (lv_is_integrity(seg_status->seg->lv)) length = seg_status->seg->integrity_data_sectors; do { target = dm_get_next_target(dmt, target, &target_start, &target_length, &target_name, &target_params); if ((start == target_start) && ((length == target_length) || ((lv_is_vdo_pool(seg_status->seg->lv)) && /* should fit within extent size */ (length < target_length) && ((length + extent_size) > target_length)) || (length_crop && (length_crop == target_length)))) break; /* Keep target_params when matching segment is found */ target_params = NULL; /* Marking this target_params unusable */ } while (target); if (!target_name || !_get_segment_status_from_target_params(target_name, target_params, dminfo, seg_status)) stack; } r = 1; out: dm_task_destroy(dmt); return r; } /* * ignore_blocked_mirror_devices * @dev * @start * @length * @mirror_status_str * * When a DM 'mirror' target is created with 'block_on_error' or * 'handle_errors', it will block I/O if there is a device failure * until the mirror is reconfigured. Thus, LVM should never attempt * to read labels from a mirror that has a failed device. (LVM * commands are issued to repair mirrors; and if LVM is blocked * attempting to read a mirror, a circular dependency would be created.) * * This function is a slimmed-down version of lib/mirror/mirrored.c: * _mirrored_transient_status(). * * If a failed device is detected in the status string, then it must be * determined if 'block_on_error' or 'handle_errors' was used when * creating the mirror. This info can only be determined from the mirror * table. The 'dev', 'start', 'length' trio allow us to correlate the * 'mirror_status_str' with the correct device table in order to check * for blocking. * * Returns: 1 if mirror should be ignored, 0 if safe to use */ static int _ignore_blocked_mirror_devices(struct cmd_context *cmd, struct device *dev, uint64_t start, uint64_t length, char *mirror_status_str) { struct dm_pool *mem; struct dm_status_mirror *sm; unsigned i, check_for_blocking = 0; uint64_t s,l; char *p, *params, *target_type = NULL; void *next = NULL; struct dm_task *dmt = NULL; int r = 0; struct device *tmp_dev; char buf[16]; if (!(mem = dm_pool_create("blocked_mirrors", 128))) return_0; if (!dm_get_status_mirror(mem, mirror_status_str, &sm)) goto_out; for (i = 0; i < sm->dev_count; ++i) if (sm->devs[i].health != DM_STATUS_MIRROR_ALIVE) { log_debug_activation("%s: Mirror image %d marked as failed.", dev_name(dev), i); check_for_blocking = 1; } if (!check_for_blocking && sm->log_count) { if (sm->logs[0].health != DM_STATUS_MIRROR_ALIVE) { log_debug_activation("%s: Mirror log device marked as failed.", dev_name(dev)); check_for_blocking = 1; } else { if (dm_snprintf(buf, sizeof(buf), "%u:%u", sm->logs[0].major, sm->logs[0].minor) < 0) goto_out; if (!(tmp_dev = dev_create_file(buf, NULL, NULL, 0))) goto_out; tmp_dev->dev = MKDEV(sm->logs[0].major, sm->logs[0].minor); if (device_is_usable(cmd, tmp_dev, (struct dev_usable_check_params) { .check_empty = 1, .check_blocked = 1, .check_suspended = ignore_suspended_devices(), .check_error_target = 1, .check_reserved = 0 }, NULL)) goto out; /* safe to use */ stack; } } if (!check_for_blocking) { r = 1; goto out; } /* * We avoid another system call if we can, but if a device is * dead, we have no choice but to look up the table too. */ if (!(dmt = _setup_task_run(DM_DEVICE_TABLE, NULL, NULL, NULL, NULL, MAJOR(dev->dev), MINOR(dev->dev), 0, 1, 0))) goto_out; do { next = dm_get_next_target(dmt, next, &s, &l, &target_type, ¶ms); if ((s == start) && (l == length) && target_type && params) { if (strcmp(target_type, TARGET_NAME_MIRROR)) goto_out; if (((p = strstr(params, " block_on_error")) && (p[15] == '\0' || p[15] == ' ')) || ((p = strstr(params, " handle_errors")) && (p[14] == '\0' || p[14] == ' '))) { log_debug_activation("%s: I/O blocked to mirror device.", dev_name(dev)); goto out; } } } while (next); r = 1; out: if (dmt) dm_task_destroy(dmt); dm_pool_destroy(mem); return r; } static int _device_is_suspended(int major, int minor) { struct dm_task *dmt; struct dm_info info; if (!(dmt = _setup_task_run(DM_DEVICE_INFO, &info, NULL, NULL, NULL, major, minor, 0, 0, 0))) return_0; dm_task_destroy(dmt); return (info.exists && info.suspended); } static int _ignore_suspended_snapshot_component(struct device *dev) { struct dm_task *dmt; void *next = NULL; char *params, *target_type = NULL; uint64_t start, length; int major1, minor1, major2, minor2; int r = 0; if (!(dmt = _setup_task_run(DM_DEVICE_TABLE, NULL, NULL, NULL, NULL, MAJOR(dev->dev), MINOR(dev->dev), 0, 1, 0))) return_0; do { next = dm_get_next_target(dmt, next, &start, &length, &target_type, ¶ms); if (!target_type) continue; if (!strcmp(target_type, TARGET_NAME_SNAPSHOT)) { if (!params || sscanf(params, "%d:%d %d:%d", &major1, &minor1, &major2, &minor2) != 4) { log_warn("WARNING: Incorrect snapshot table found for %d:%d.", (int)MAJOR(dev->dev), (int)MINOR(dev->dev)); goto out; } r = r || _device_is_suspended(major1, minor1) || _device_is_suspended(major2, minor2); } else if (!strcmp(target_type, TARGET_NAME_SNAPSHOT_ORIGIN)) { if (!params || sscanf(params, "%d:%d", &major1, &minor1) != 2) { log_warn("WARNING: Incorrect snapshot-origin table found for %d:%d.", (int)MAJOR(dev->dev), (int)MINOR(dev->dev)); goto out; } r = r || _device_is_suspended(major1, minor1); } } while (next); out: dm_task_destroy(dmt); return r; } static int _ignore_unusable_thins(struct device *dev) { /* TODO make function for thin testing */ struct dm_pool *mem; struct dm_status_thin_pool *status; struct dm_task *dmt = NULL; void *next = NULL; uint64_t start, length; char *target_type = NULL; char *params; int minor, major; int r = 0; if (!(mem = dm_pool_create("unusable_thins", 128))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_TABLE, NULL, NULL, NULL, NULL, MAJOR(dev->dev), MINOR(dev->dev), 0, 1, 0))) goto_out; dm_get_next_target(dmt, next, &start, &length, &target_type, ¶ms); if (!params || sscanf(params, "%d:%d", &major, &minor) != 2) { log_warn("WARNING: Cannot get thin-pool major:minor for thin device %d:%d.", (int)MAJOR(dev->dev), (int)MINOR(dev->dev)); goto out; } dm_task_destroy(dmt); if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, NULL, NULL, NULL, NULL, major, minor, 0, 0, 0))) goto_out; dm_get_next_target(dmt, next, &start, &length, &target_type, ¶ms); if (!dm_get_status_thin_pool(mem, params, &status)) goto_out; if (status->read_only || status->out_of_data_space) { log_warn("WARNING: %s: Thin's thin-pool needs inspection.", dev_name(dev)); goto out; } r = 1; out: if (dmt) dm_task_destroy(dmt); dm_pool_destroy(mem); return r; } static int _ignore_invalid_snapshot(const char *params) { struct dm_status_snapshot *s; struct dm_pool *mem; int r = 0; if (!(mem = dm_pool_create("invalid snapshots", 128))) return_0; if (!dm_get_status_snapshot(mem, params, &s)) stack; else r = s->invalid; dm_pool_destroy(mem); return r; } static int _ignore_frozen_raid(struct device *dev, const char *params) { struct dm_status_raid *s; struct dm_pool *mem; int r = 0; if (!(mem = dm_pool_create("frozen raid", 128))) return_0; if (!dm_get_status_raid(mem, params, &s)) stack; else if (s->sync_action && !strcmp(s->sync_action, "frozen")) { log_warn("WARNING: %s frozen raid device (%d:%d) needs inspection.", dev_name(dev), (int)MAJOR(dev->dev), (int)MINOR(dev->dev)); r = 1; } dm_pool_destroy(mem); return r; } static int _is_usable_uuid(const struct device *dev, const char *name, const char *uuid, int check_reserved, int check_lv, int *is_lv) { char *vgname, *lvname, *layer; char vg_name[NAME_LEN]; if (!check_reserved && !check_lv) return 1; if (!strncmp(uuid, UUID_PREFIX, sizeof(UUID_PREFIX) - 1)) { /* with LVM- prefix */ if (check_reserved) { /* Check internal lvm devices */ if (strlen(uuid) > (sizeof(UUID_PREFIX) + 2 * ID_LEN)) { /* 68 with suffix */ log_debug_activation("%s: Reserved uuid %s on internal LV device %s not usable.", dev_name(dev), uuid, name); return 0; } /* Recognize some older reserved LVs just from the LV name (snapshot, pvmove...) */ vgname = vg_name; if (!dm_strncpy(vg_name, name, sizeof(vg_name)) || !dm_split_lvm_name(NULL, NULL, &vgname, &lvname, &layer)) return_0; /* FIXME: fails to handle dev aliases i.e. /dev/dm-5, replace with UUID suffix */ if (lvname && (is_reserved_lvname(lvname) || *layer)) { log_debug_activation("%s: Reserved internal LV device %s/%s%s%s not usable.", dev_name(dev), vgname, lvname, *layer ? "-" : "", layer); return 0; } } if (check_lv) { /* Skip LVs */ if (is_lv) *is_lv = 1; return 0; } } if (check_reserved && (!strncmp(uuid, CRYPT_TEMP, sizeof(CRYPT_TEMP) - 1) || !strncmp(uuid, CRYPT_SUBDEV, sizeof(CRYPT_SUBDEV) - 1) || !strncmp(uuid, STRATIS, sizeof(STRATIS) - 1))) { /* Skip private crypto devices */ log_debug_activation("%s: Reserved uuid %s on %s device %s not usable.", dev_name(dev), uuid, uuid[0] == 'C' ? "crypto" : "stratis", name); return 0; } return 1; } /* * device_is_usable * @dev * @check_lv_names * * A device is considered not usable if it is: * 1) An empty device (no targets) * 2) A blocked mirror (i.e. a mirror with a failure and block_on_error set) * 3) ignore_suspended_devices is set and * a) the device is suspended * b) it is a snapshot origin * 4) an error target * 5) the LV name is a reserved name. * * Returns: 1 if usable, 0 otherwise */ int device_is_usable(struct cmd_context *cmd, struct device *dev, struct dev_usable_check_params check, int *is_lv) { struct dm_task *dmt; struct dm_info info; const char *name, *uuid; uint64_t start, length; char *target_type = NULL; char *params; void *next = NULL; int only_error_or_zero_target = 1; int r = 0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, NULL, NULL, MAJOR(dev->dev), MINOR(dev->dev), 0, 0, 0))) return_0; if (!info.exists) goto out; name = dm_task_get_name(dmt); uuid = dm_task_get_uuid(dmt); if (check.check_empty && !info.target_count) { log_debug_activation("%s: Empty device %s not usable.", dev_name(dev), name); goto out; } if (check.check_suspended && info.suspended) { log_debug_activation("%s: Suspended device %s not usable.", dev_name(dev), name); goto out; } if (uuid && !_is_usable_uuid(dev, name, uuid, check.check_reserved, check.check_lv, is_lv)) goto out; /* FIXME Also check for mpath no paths */ do { next = dm_get_next_target(dmt, next, &start, &length, &target_type, ¶ms); if (!target_type) continue; if (check.check_blocked && !strcmp(target_type, TARGET_NAME_MIRROR)) { if (ignore_lvm_mirrors()) { log_debug_activation("%s: Scanning mirror devices is disabled.", dev_name(dev)); goto out; } if (!_ignore_blocked_mirror_devices(cmd, dev, start, length, params)) { log_debug_activation("%s: Mirror device %s not usable.", dev_name(dev), name); goto out; } } /* * FIXME: Snapshot origin could be sitting on top of a mirror * which could be blocking I/O. We should add a check for the * stack here and see if there's blocked mirror underneath. * Currently, mirrors used as origin or snapshot is not * supported anymore and in general using mirrors in a stack * is disabled by default (with a warning that if enabled, * it could cause various deadlocks). * Similar situation can happen with RAID devices where * a RAID device can be snapshotted. * If one of the RAID legs are down and we're doing * lvconvert --repair, there's a time period in which * snapshot components are (besides other devs) suspended. * See also https://bugzilla.redhat.com/show_bug.cgi?id=1219222 * for an example where this causes problems. * * This is a quick check for now, but replace it with more * robust and better check that would check the stack * correctly, not just snapshots but any cobimnation possible * in a stack - use proper dm tree to check this instead. */ if (check.check_suspended && (!strcmp(target_type, TARGET_NAME_SNAPSHOT) || !strcmp(target_type, TARGET_NAME_SNAPSHOT_ORIGIN)) && _ignore_suspended_snapshot_component(dev)) { log_debug_activation("%s: %s device %s not usable.", dev_name(dev), target_type, name); goto out; } if (!strcmp(target_type, TARGET_NAME_SNAPSHOT) && _ignore_invalid_snapshot(params)) { log_debug_activation("%s: Invalid %s device %s not usable.", dev_name(dev), target_type, name); goto out; } if (!strncmp(target_type, TARGET_NAME_RAID, 4) && _ignore_frozen_raid(dev, params)) { log_debug_activation("%s: Frozen %s device %s not usable.", dev_name(dev), target_type, name); goto out; } /* TODO: extend check struct ? */ if (!strcmp(target_type, TARGET_NAME_THIN) && !_ignore_unusable_thins(dev)) { log_debug_activation("%s: %s device %s not usable.", dev_name(dev), target_type, name); goto out; } if (only_error_or_zero_target && strcmp(target_type, TARGET_NAME_ERROR) && strcmp(target_type, TARGET_NAME_ZERO)) only_error_or_zero_target = 0; } while (next); /* Skip devices consisting entirely of error or zero targets. */ /* FIXME Deal with device stacked above error targets? */ if (check.check_error_target && only_error_or_zero_target) { log_debug_activation("%s: Error device %s not usable.", dev_name(dev), name); goto out; } /* FIXME Also check dependencies? */ r = 1; out: dm_task_destroy(dmt); return r; } /* * If active LVs were activated by a version of LVM2 before 2.02.00 we must * perform additional checks to find them because they do not have the LVM- * prefix on their dm uuids. * As of 2.02.150, we've chosen to disable this compatibility arbitrarily if * we're running kernel version 3 or above. */ #define MIN_KERNEL_MAJOR 3 static int _original_uuid_format_check_required(struct cmd_context *cmd) { static int _kernel_major = 0; if (!_kernel_major) { if ((sscanf(cmd->kernel_vsn, "%d", &_kernel_major) == 1) && (_kernel_major >= MIN_KERNEL_MAJOR)) log_debug_activation("Skipping checks for old devices without " UUID_PREFIX " dm uuid prefix (kernel vsn %d >= %d).", _kernel_major, MIN_KERNEL_MAJOR); else _kernel_major = -1; } return (_kernel_major == -1); } static int _info(struct cmd_context *cmd, const char *name, const char *dlid, int with_open_count, int with_read_ahead, int with_name_check, struct dm_info *dminfo, uint32_t *read_ahead, struct lv_seg_status *seg_status) { char old_style_dlid[sizeof(UUID_PREFIX) + 2 * ID_LEN]; const char *suffix, *suffix_position; const char *name_check = (with_name_check) ? name : NULL; unsigned i = 0; log_debug_activation("Getting device info for %s [%s].", name, dlid); /* Check for dlid */ if (!_info_run(dlid, dminfo, read_ahead, seg_status, name_check, with_open_count, with_read_ahead, 0, 0)) return_0; if (dminfo->exists) return 1; /* Check for original version of dlid before the suffixes got added in 2.02.106 */ if ((suffix_position = strrchr(dlid, '-'))) { while ((suffix = uuid_suffix_list[i++])) { if (strcmp(suffix_position + 1, suffix)) continue; (void) dm_strncpy(old_style_dlid, dlid, sizeof(old_style_dlid)); if (!_info_run(old_style_dlid, dminfo, read_ahead, seg_status, name_check, with_open_count, with_read_ahead, 0, 0)) return_0; if (dminfo->exists) return 1; } } /* Must we still check for the pre-2.02.00 dm uuid format? */ if (!_original_uuid_format_check_required(cmd)) return 1; /* Check for dlid before UUID_PREFIX was added */ if (!_info_run(dlid + sizeof(UUID_PREFIX) - 1, dminfo, read_ahead, seg_status, name_check, with_open_count, with_read_ahead, 0, 0)) return_0; return 1; } int dev_manager_remove_dm_major_minor(uint32_t major, uint32_t minor) { struct dm_task *dmt; int r = 0; log_verbose("Removing dm dev %u:%u", major, minor); if (!(dmt = dm_task_create(DM_DEVICE_REMOVE))) return_0; if (!dm_task_set_major(dmt, major) || !dm_task_set_minor(dmt, minor)) { log_error("Failed to set device number for remove %u:%u", major, minor); goto out; } r = dm_task_run(dmt); out: dm_task_destroy(dmt); return r; } static int _info_by_dev(uint32_t major, uint32_t minor, struct dm_info *info) { return _info_run(NULL, info, NULL, NULL, NULL, 0, 0, major, minor); } int dev_manager_check_prefix_dm_major_minor(uint32_t major, uint32_t minor, const char *prefix) { struct dm_task *dmt; const char *uuid; int r = 1; if (!(dmt = _setup_task_run(DM_DEVICE_INFO, NULL, NULL, NULL, 0, major, minor, 0, 0, 0))) return_0; if (!(uuid = dm_task_get_uuid(dmt)) || strncasecmp(uuid, prefix, strlen(prefix))) r = 0; dm_task_destroy(dmt); return r; } int dev_manager_get_device_list(const char *prefix, struct dm_list **devs, unsigned *devs_features) { struct dm_task *dmt; int r = 1; if (!(dmt = _setup_task_run(DM_DEVICE_LIST, NULL, NULL, NULL, 0, 0, 0, 0, 0, 0))) return_0; if (!dm_task_get_device_list(dmt, devs, devs_features)) { r = 0; goto_out; } out: dm_task_destroy(dmt); return r; } int dev_manager_info(struct cmd_context *cmd, const struct logical_volume *lv, const char *layer, int with_open_count, int with_read_ahead, int with_name_check, struct dm_info *dminfo, uint32_t *read_ahead, struct lv_seg_status *seg_status) { char *dlid, *name; int r = 0; if (!(name = dm_build_dm_name(cmd->mem, lv->vg->name, lv->name, layer))) return_0; if (!(dlid = build_dm_uuid(cmd->mem, lv, layer))) goto_out; if (!cmd->disable_dm_devs && cmd->cache_dm_devs && !dm_device_list_find_by_uuid(cmd->cache_dm_devs, dlid, NULL)) { log_debug("Cached as inactive %s.", name); if (dminfo) memset(dminfo, 0, sizeof(*dminfo)); r = 1; goto out; } if (!(r = _info(cmd, name, dlid, with_open_count, with_read_ahead, with_name_check, dminfo, read_ahead, seg_status))) stack; out: dm_pool_free(cmd->mem, name); return r; } static const struct dm_info *_cached_dm_info(struct dm_pool *mem, struct dm_tree *dtree, const struct logical_volume *lv, const char *layer) { char *dlid; const struct dm_tree_node *dnode; const struct dm_info *dinfo = NULL; if (!(dlid = build_dm_uuid(mem, lv, layer))) return_NULL; if (!(dnode = dm_tree_find_node_by_uuid(dtree, dlid))) goto out; if (!(dinfo = dm_tree_node_get_info(dnode))) { log_warn("WARNING: Cannot get info from tree node for %s.", display_lvname(lv)); goto out; } if (!dinfo->exists) dinfo = NULL; out: dm_pool_free(mem, dlid); return dinfo; } int lv_has_target_type(struct dm_pool *mem, const struct logical_volume *lv, const char *layer, const char *target_type) { int r = 0; char *dlid; struct dm_task *dmt; struct dm_info info; void *next = NULL; uint64_t start, length; char *type = NULL; char *params = NULL; if (!(dlid = build_dm_uuid(mem, lv, layer))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, 0, 0, 0, 0, 0, 0))) goto_bad; if (!info.exists) goto_out; /* If there is a preloaded table, use that in preference. */ if (info.inactive_table) { dm_task_destroy(dmt); if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, 0, 0, 0, 0, 0, 1))) goto_bad; if (!info.exists || !info.inactive_table) goto_out; } do { next = dm_get_next_target(dmt, next, &start, &length, &type, ¶ms); if (type && !strncmp(type, target_type, strlen(target_type))) { r = 1; break; } } while (next); out: dm_task_destroy(dmt); bad: dm_pool_free(mem, dlid); return r; } static int _lv_has_thin_device_id(struct dm_pool *mem, const struct logical_volume *lv, const char *layer, unsigned device_id) { char *dlid; struct dm_task *dmt; struct dm_info info; void *next = NULL; uint64_t start, length; char *type = NULL; char *params = NULL; unsigned id = ~0; if (!(dlid = build_dm_uuid(mem, lv, layer))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_TABLE, &info, NULL, dlid, 0, 0, 0, 0, 1, 0))) goto_bad; if (!info.exists) goto_out; /* If there is a preloaded table, use that in preference. */ if (info.inactive_table) { dm_task_destroy(dmt); if (!(dmt = _setup_task_run(DM_DEVICE_TABLE, &info, NULL, dlid, 0, 0, 0, 0, 1, 1))) goto_bad; if (!info.exists || !info.inactive_table) goto_out; } (void) dm_get_next_target(dmt, next, &start, &length, &type, ¶ms); if (!type || strcmp(type, TARGET_NAME_THIN)) goto_out; if (!params || sscanf(params, "%*u:%*u %u", &id) != 1) goto_out; log_debug_activation("%soaded thin volume %s with id %u is %smatching id %u.", info.inactive_table ? "Prel" : "L", display_lvname(lv), id, (device_id != id) ? "not " : "", device_id); out: dm_task_destroy(dmt); bad: dm_pool_free(mem, dlid); return (device_id == id); } int add_linear_area_to_dtree(struct dm_tree_node *node, uint64_t size, uint32_t extent_size, int use_linear_target, const char *vgname, const char *lvname) { uint32_t page_size; /* * Use striped or linear target? */ if (!use_linear_target) { page_size = lvm_getpagesize() >> SECTOR_SHIFT; /* * We'll use the extent size as the stripe size. * Extent size and page size are always powers of 2. * The striped target requires that the stripe size is * divisible by the page size. */ if (extent_size >= page_size) { /* Use striped target */ if (!dm_tree_node_add_striped_target(node, size, extent_size)) return_0; return 1; } /* Some exotic cases are unsupported by striped. */ log_warn("WARNING: Using linear target for %s/%s: Striped requires extent size " "(" FMTu32 " sectors) >= page size (" FMTu32 ").", vgname, lvname, extent_size, page_size); } /* * Use linear target. */ if (!dm_tree_node_add_linear_target(node, size)) return_0; return 1; } static dm_percent_range_t _combine_percent(dm_percent_t a, dm_percent_t b, uint32_t numerator, uint32_t denominator) { if (a == LVM_PERCENT_MERGE_FAILED || b == LVM_PERCENT_MERGE_FAILED) return LVM_PERCENT_MERGE_FAILED; if (a == DM_PERCENT_INVALID || b == DM_PERCENT_INVALID) return DM_PERCENT_INVALID; if (a == DM_PERCENT_100 && b == DM_PERCENT_100) return DM_PERCENT_100; if (a == DM_PERCENT_0 && b == DM_PERCENT_0) return DM_PERCENT_0; return (dm_percent_range_t) dm_make_percent(numerator, denominator); } static int _percent_run(struct dev_manager *dm, const char *name, const char *dlid, const char *target_type, int wait, const struct logical_volume *lv, dm_percent_t *overall_percent, uint32_t *event_nr, int fail_if_percent_unsupported, int *interrupted) { int r = 0; struct dm_task *dmt; struct dm_info info; void *next = NULL; uint64_t start, length; char *type = NULL; char *params = NULL; const struct dm_list *segh = lv ? &lv->segments : NULL; struct lv_segment *seg = NULL; int first_time = 1; dm_percent_t percent = DM_PERCENT_INVALID; uint64_t total_numerator = 0, total_denominator = 0; struct segment_type *segtype; *overall_percent = percent; if (!(segtype = get_segtype_from_string(dm->cmd, target_type))) return_0; if (wait) sigint_allow(); if (!(dmt = _setup_task_run(wait ? DM_DEVICE_WAITEVENT : DM_DEVICE_STATUS, &info, name, dlid, event_nr, 0, 0, 0, 0, 0))) goto_bad; if (!info.exists) goto_out; if (event_nr) *event_nr = info.event_nr; do { next = dm_get_next_target(dmt, next, &start, &length, &type, ¶ms); if (lv) { if (!(segh = dm_list_next(&lv->segments, segh))) { log_error("Number of segments in active LV %s " "does not match metadata.", display_lvname(lv)); goto out; } seg = dm_list_item(segh, struct lv_segment); } if (!type || !params) continue; if (strcmp(type, target_type)) { /* If kernel's type isn't an exact match is it compatible? */ if (!segtype->ops->target_status_compatible || !segtype->ops->target_status_compatible(type)) continue; } if (!segtype->ops->target_percent) continue; if (!segtype->ops->target_percent(&dm->target_state, &percent, dm->mem, dm->cmd, seg, params, &total_numerator, &total_denominator)) goto_out; if (first_time) { *overall_percent = percent; first_time = 0; } else *overall_percent = _combine_percent(*overall_percent, percent, total_numerator, total_denominator); } while (next); if (lv && dm_list_next(&lv->segments, segh)) { log_error("Number of segments in active LV %s does not " "match metadata.", display_lvname(lv)); goto out; } if (first_time) { /* above ->target_percent() was not executed! */ /* FIXME why return PERCENT_100 et. al. in this case? */ *overall_percent = DM_PERCENT_100; if (fail_if_percent_unsupported) goto_out; } log_debug_activation("LV percent: %s", display_percent(dm->cmd, *overall_percent)); r = 1; out: dm_task_destroy(dmt); bad: if (wait) { sigint_restore(); if (sigint_caught()) { *interrupted = 1; return_0; } } return r; } static int _percent(struct dev_manager *dm, const char *name, const char *dlid, const char *target_type, int wait, const struct logical_volume *lv, dm_percent_t *percent, uint32_t *event_nr, int fail_if_percent_unsupported) { int interrupted = 0; if (dlid && *dlid) { if (_percent_run(dm, NULL, dlid, target_type, wait, lv, percent, event_nr, fail_if_percent_unsupported, &interrupted)) return 1; if (!interrupted && _original_uuid_format_check_required(dm->cmd) && _percent_run(dm, NULL, dlid + sizeof(UUID_PREFIX) - 1, target_type, wait, lv, percent, event_nr, fail_if_percent_unsupported, &interrupted)) return 1; } if (!interrupted && name && _percent_run(dm, name, NULL, target_type, wait, lv, percent, event_nr, fail_if_percent_unsupported, &interrupted)) return 1; return_0; } /* FIXME Merge with the percent function */ int dev_manager_transient(struct dev_manager *dm, const struct logical_volume *lv) { int r = 0; struct dm_task *dmt; struct dm_info info; void *next = NULL; uint64_t start, length; char *type = NULL; char *params = NULL; char *dlid = NULL; const char *layer = lv_layer(lv); const struct dm_list *segh = &lv->segments; struct lv_segment *seg = NULL; if (!(dlid = build_dm_uuid(dm->mem, lv, layer))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, NULL, 0, 0, 0, 0, 0))) return_0; if (!info.exists) goto_out; do { next = dm_get_next_target(dmt, next, &start, &length, &type, ¶ms); if (!(segh = dm_list_next(&lv->segments, segh))) { log_error("Number of segments in active LV %s " "does not match metadata.", display_lvname(lv)); goto out; } seg = dm_list_item(segh, struct lv_segment); if (!type || !params) continue; if (!seg) { log_error(INTERNAL_ERROR "Segment is not selected."); goto out; } if (seg->segtype->ops->check_transient_status && !seg->segtype->ops->check_transient_status(dm->mem, seg, params)) goto_out; } while (next); if (dm_list_next(&lv->segments, segh)) { log_error("Number of segments in active LV %s does not " "match metadata.", display_lvname(lv)); goto out; } r = 1; out: dm_task_destroy(dmt); return r; } /* * dev_manager implementation. */ struct dev_manager *dev_manager_create(struct cmd_context *cmd, const char *vg_name, unsigned track_pvmove_deps) { struct dm_pool *mem; struct dev_manager *dm; if (!(mem = dm_pool_create("dev_manager", 16 * 1024))) return_NULL; if (!(dm = dm_pool_zalloc(mem, sizeof(*dm)))) goto_bad; dm->cmd = cmd; dm->mem = mem; dm->vg_name = vg_name; /* * When we manipulate (normally suspend/resume) the PVMOVE * device directly, there's no need to touch the LVs above. */ dm->track_pvmove_deps = track_pvmove_deps; dm->target_state = NULL; dm_udev_set_sync_support(cmd->current_settings.udev_sync); return dm; bad: dm_pool_destroy(mem); return NULL; } void dev_manager_destroy(struct dev_manager *dm) { dm_pool_destroy(dm->mem); } void dev_manager_release(void) { dm_lib_release(); } void dev_manager_exit(void) { dm_lib_exit(); } int dev_manager_snapshot_percent(struct dev_manager *dm, const struct logical_volume *lv, dm_percent_t *percent) { const struct logical_volume *snap_lv; char *name; const char *dlid; int fail_if_percent_unsupported = 0; if (lv_is_merging_origin(lv)) { /* * Set 'fail_if_percent_unsupported', otherwise passing * unsupported LV types to _percent will lead to a default * successful return with percent_range as PERCENT_100. * - For a merging origin, this will result in a polldaemon * that runs infinitely (because completion is PERCENT_0) * - We unfortunately don't yet _know_ if a snapshot-merge * target is active (activation is deferred if dev is open); * so we can't short-circuit origin devices based purely on * existing LVM LV attributes. */ fail_if_percent_unsupported = 1; } if (lv_is_merging_cow(lv)) { /* must check percent of origin for a merging snapshot */ snap_lv = origin_from_cow(lv); } else snap_lv = lv; /* * Build a name for the top layer. */ if (!(name = dm_build_dm_name(dm->mem, snap_lv->vg->name, snap_lv->name, NULL))) return_0; if (!(dlid = build_dm_uuid(dm->mem, snap_lv, NULL))) return_0; /* * Try and get some info on this device. */ if (!_percent(dm, name, dlid, TARGET_NAME_SNAPSHOT, 0, NULL, percent, NULL, fail_if_percent_unsupported)) return_0; /* If the snapshot isn't available, percent will be -1 */ return 1; } /* FIXME Merge with snapshot_percent, auto-detecting target type */ /* FIXME Cope with more than one target */ int dev_manager_mirror_percent(struct dev_manager *dm, const struct logical_volume *lv, int wait, dm_percent_t *percent, uint32_t *event_nr) { char *name; const char *dlid; const char *target_type = first_seg(lv)->segtype->name; const char *layer = lv_layer(lv); /* * Build a name for the top layer. */ if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name, layer))) return_0; if (!(dlid = build_dm_uuid(dm->mem, lv, layer))) return_0; log_debug_activation("Getting device %s status percentage for %s.", target_type, name); if (!_percent(dm, name, dlid, target_type, wait, lv, percent, event_nr, 0)) return_0; return 1; } int dev_manager_raid_status(struct dev_manager *dm, const struct logical_volume *lv, struct lv_status_raid **status, int *exists) { int r = 0; const char *dlid; struct dm_task *dmt; struct dm_info info; uint64_t start, length; char *type = NULL; char *params = NULL; const char *layer = lv_layer(lv); struct dm_status_raid *sr; *exists = -1; if (!(*status = dm_pool_zalloc(dm->mem, sizeof(struct lv_status_cache)))) return_0; if (!(dlid = build_dm_uuid(dm->mem, lv, layer))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, 0, 0, 0, 0, 0, 0))) return_0; if (!(*exists = info.exists)) goto out; log_debug_activation("Checking raid status for volume %s.", display_lvname(lv)); dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms); if (!type || strcmp(type, TARGET_NAME_RAID)) { log_error("Expected %s segment type but got %s instead.", TARGET_NAME_RAID, type ? type : "NULL"); goto out; } /* FIXME Check there's only one target */ if (!dm_get_status_raid(dm->mem, params, &sr)) goto_out; (*status)->mem = dm->mem; /* User has to destroy this mem pool later */ (*status)->raid = sr; (*status)->in_sync = dm_make_percent(sr->insync_regions, sr->total_regions); r = 1; out: dm_task_destroy(dmt); return r; } int dev_manager_raid_message(struct dev_manager *dm, const struct logical_volume *lv, const char *msg) { int r = 0; const char *dlid; struct dm_task *dmt; const char *layer = lv_layer(lv); if (!lv_is_raid(lv)) { log_error(INTERNAL_ERROR "%s is not a RAID logical volume.", display_lvname(lv)); return 0; } /* These are the supported RAID messages for dm-raid v1.9.0 */ if (strcmp(msg, "idle") && strcmp(msg, "frozen") && strcmp(msg, "resync") && strcmp(msg, "recover") && strcmp(msg, "check") && strcmp(msg, "repair")) { log_error(INTERNAL_ERROR "Unknown RAID message: %s.", msg); return 0; } if (!(dlid = build_dm_uuid(dm->mem, lv, layer))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_TARGET_MSG, NULL, NULL, dlid, 0, 0, 0, 0, 1, 0))) return_0; if (!dm_task_set_message(dmt, msg)) goto_out; if (!dm_task_run(dmt)) goto_out; r = 1; out: dm_task_destroy(dmt); return r; } int dev_manager_writecache_message(struct dev_manager *dm, const struct logical_volume *lv, const char *msg) { int r = 0; const char *dlid; struct dm_task *dmt; const char *layer = lv_layer(lv); if (!lv_is_writecache(lv)) { log_error(INTERNAL_ERROR "%s is not a writecache logical volume.", display_lvname(lv)); return 0; } if (!(dlid = build_dm_uuid(dm->mem, lv, layer))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_TARGET_MSG, NULL, NULL, dlid, 0, 0, 0, 0, 1, 0))) return_0; if (!dm_task_set_message(dmt, msg)) goto_out; if (!dm_task_run(dmt)) goto_out; r = 1; out: dm_task_destroy(dmt); return r; } int dev_manager_cache_status(struct dev_manager *dm, const struct logical_volume *lv, struct lv_status_cache **status, int *exists) { int r = 0; const char *dlid; struct dm_task *dmt; struct dm_info info; uint64_t start, length; char *type = NULL; char *params = NULL; struct dm_status_cache *c; *exists = -1; if (!(dlid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, 0, 0, 0, 0, 0, 0))) return_0; if (!(*exists = info.exists)) goto out; log_debug_activation("Checking status for cache volume %s.", display_lvname(lv)); dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms); if (!type || strcmp(type, TARGET_NAME_CACHE)) { log_error("Expected %s segment type but got %s instead.", TARGET_NAME_CACHE, type ? type : "NULL"); goto out; } /* * FIXME: * ->target_percent() API is able to transfer only a single value. * Needs to be able to pass whole structure. */ if (!dm_get_status_cache(dm->mem, params, &c)) goto_out; if (!(*status = dm_pool_zalloc(dm->mem, sizeof(struct lv_status_cache)))) goto_out; (*status)->mem = dm->mem; /* User has to destroy this mem pool later */ (*status)->cache = c; if (c->fail || c->error) { (*status)->data_usage = (*status)->metadata_usage = (*status)->dirty_usage = DM_PERCENT_INVALID; } else { (*status)->data_usage = dm_make_percent(c->used_blocks, c->total_blocks); (*status)->metadata_usage = dm_make_percent(c->metadata_used_blocks, c->metadata_total_blocks); (*status)->dirty_usage = (c->used_blocks) ? dm_make_percent(c->dirty_blocks, c->used_blocks) : DM_PERCENT_0; } r = 1; out: dm_task_destroy(dmt); return r; } int dev_manager_thin_pool_status(struct dev_manager *dm, const struct logical_volume *lv, int flush, struct lv_status_thin_pool **status, int *exists) { struct dm_status_thin_pool *dm_status; const char *dlid; struct dm_task *dmt; struct dm_info info; uint64_t start, length; char *type = NULL; char *params = NULL; int r = 0; *exists = -1; if (!(*status = dm_pool_zalloc(dm->mem, sizeof(struct lv_status_thin_pool)))) return_0; /* Build dlid for the thin pool layer */ if (!(dlid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, 0, 0, 0, 0, flush, 0))) return_0; if (!(*exists = info.exists)) goto out; log_debug_activation("Checking thin pool status for LV %s.", display_lvname(lv)); dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms); if (!type || strcmp(type, TARGET_NAME_THIN_POOL)) { log_error("Expected %s segment type but got %s instead.", TARGET_NAME_THIN_POOL, type ? type : "NULL"); goto out; } if (!dm_get_status_thin_pool(dm->mem, params, &dm_status)) goto_out; (*status)->mem = dm->mem; (*status)->thin_pool = dm_status; if (dm_status->fail || dm_status->error) { (*status)->data_usage = (*status)->metadata_usage = DM_PERCENT_INVALID; } else { (*status)->data_usage = dm_make_percent(dm_status->used_data_blocks, dm_status->total_data_blocks); (*status)->metadata_usage = dm_make_percent(dm_status->used_metadata_blocks, dm_status->total_metadata_blocks); } r = 1; out: dm_task_destroy(dmt); return r; } int dev_manager_thin_status(struct dev_manager *dm, const struct logical_volume *lv, int flush, struct lv_status_thin **status, int *exists) { struct dm_status_thin *dm_status; const char *dlid; struct dm_task *dmt; struct dm_info info; uint64_t start, length; char *type = NULL; char *params = NULL; uint64_t csize; int r = 0; *exists = -1; if (!(*status = dm_pool_zalloc(dm->mem, sizeof(struct lv_status_thin)))) return_0; if (!(dlid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, 0, 0, 0, 0, flush, 0))) return_0; if (!(*exists = info.exists)) goto out; log_debug_activation("Checking thin status for LV %s.", display_lvname(lv)); dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms); if (!type || strcmp(type, TARGET_NAME_THIN)) { log_error("Expected %s segment type but got %s instead.", TARGET_NAME_THIN, type ? type : "NULL"); goto out; } if (!dm_get_status_thin(dm->mem, params, &dm_status)) goto_out; (*status)->mem = dm->mem; (*status)->thin = dm_status; if (dm_status->fail) (*status)->usage = DM_PERCENT_INVALID; else { /* Pool allocates whole chunk so round-up to nearest one */ csize = first_seg(first_seg(lv)->pool_lv)->chunk_size; csize = ((lv->size + csize - 1) / csize) * csize; if (dm_status->mapped_sectors > csize) { log_warn("WARNING: LV %s maps %s while the size is only %s.", display_lvname(lv), display_size(dm->cmd, dm_status->mapped_sectors), display_size(dm->cmd, csize)); /* Don't show nonsense numbers like i.e. 1000% full */ dm_status->mapped_sectors = csize; } (*status)->usage = dm_make_percent(dm_status->mapped_sectors, csize); } r = 1; out: dm_task_destroy(dmt); return r; } /* * Explore state of running DM table to obtain currently used deviceId */ int dev_manager_thin_device_id(struct dev_manager *dm, const struct logical_volume *lv, uint32_t *device_id, int *exists) { const char *dlid; struct dm_task *dmt; struct dm_info info; uint64_t start, length; char *params, *target_type = NULL; const char *layer = lv_layer(lv); int r = 0; *exists = -1; if (lv_is_merging_origin(lv) && !lv_info(lv->vg->cmd, lv, 1, NULL, 0, 0)) /* If the merge has already happened, that table * can already be using correct LV without -real layer */ layer = NULL; /* Build dlid for the thin layer */ if (!(dlid = build_dm_uuid(dm->mem, lv, layer))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_TABLE, &info, NULL, dlid, 0, 0, 0, 0, 1, 0))) return_0; if (!(*exists = info.exists)) goto out; log_debug_activation("Checking device id for LV %s.", display_lvname(lv)); if (dm_get_next_target(dmt, NULL, &start, &length, &target_type, ¶ms)) { log_error("More then one table line found for %s.", display_lvname(lv)); goto out; } if (!target_type || strcmp(target_type, TARGET_NAME_THIN)) { log_error("Unexpected target type %s found for thin %s.", target_type, display_lvname(lv)); goto out; } if (!params || sscanf(params, "%*u:%*u %u", device_id) != 1) { log_error("Cannot parse table like parameters %s for %s.", params, display_lvname(lv)); goto out; } r = 1; out: dm_task_destroy(dmt); return r; } int dev_manager_vdo_pool_status(struct dev_manager *dm, const struct logical_volume *lv, int flush, struct lv_status_vdo **status, int *exists) { const char *dlid; struct dm_info info; uint64_t start, length; struct dm_task *dmt = NULL; char *type = NULL; char *params = NULL; int r = 0; *exists = -1; if (!(*status = dm_pool_zalloc(dm->mem, sizeof(struct lv_status_vdo)))) return_0; if (!(dlid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_STATUS, &info, NULL, dlid, 0, 0, 0, 0, flush, 0))) return_0; if (!(*exists = info.exists)) goto out; log_debug_activation("Checking VDO pool status for LV %s.", display_lvname(lv)); if (dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms)) { log_error("More then one table line found for %s.", display_lvname(lv)); goto out; } if (!type || strcmp(type, TARGET_NAME_VDO)) { log_error("Expected %s segment type but got %s instead.", TARGET_NAME_VDO, type ? type : "NULL"); goto out; } if (!parse_vdo_pool_status(dm->mem, lv, params, &info, *status)) goto_out; (*status)->mem = dm->mem; r = 1; out: dm_task_destroy(dmt); return r; } int dev_manager_vdo_pool_size_config(struct dev_manager *dm, const struct logical_volume *lv, struct vdo_pool_size_config *cfg) { const char *dlid; struct dm_info info; uint64_t start, length; struct dm_task *dmt = NULL; char *type = NULL; char *params = NULL; int r = 0; unsigned version = 0; memset(cfg, 0, sizeof(*cfg)); if (!(dlid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_TABLE, &info, NULL, dlid, 0, 0, 0, 0, 0, 0))) return_0; if (!info.exists) goto inactive; /* VDO device is not active, should not happen here... */ log_debug_activation("Checking VDO pool table line for LV %s.", display_lvname(lv)); if (dm_get_next_target(dmt, NULL, &start, &length, &type, ¶ms)) { log_error("More then one table line found for %s.", display_lvname(lv)); goto out; } if (!type || strcmp(type, TARGET_NAME_VDO)) { log_error("Expected %s segment type but got %s instead.", TARGET_NAME_VDO, type ? type : "NULL"); goto out; } if (sscanf(params, "V%u %*s " FMTu64 " %*u " FMTu32, &version, &cfg->physical_size, &cfg->block_map_cache_size_mb) != 3) { log_error("Failed to parse VDO parameters %s for LV %s.", params, display_lvname(lv)); goto out; } switch (version) { case 2: break; case 4: break; default: log_warn("WARNING: Unknown VDO table line version %u.", version); } cfg->virtual_size = length; cfg->physical_size *= 8; // From 4K unit to 512B cfg->block_map_cache_size_mb /= 256; // From 4K unit to MiB cfg->index_memory_size_mb = first_seg(lv)->vdo_params.index_memory_size_mb; // Preserved inactive: r = 1; out: dm_task_destroy(dmt); return r; } /*************************/ /* NEW CODE STARTS HERE */ /*************************/ static int _dev_manager_lv_mknodes(const struct logical_volume *lv) { char *name; if (!(name = dm_build_dm_name(lv->vg->cmd->mem, lv->vg->name, lv->name, NULL))) return_0; return fs_add_lv(lv, name); } static int _dev_manager_lv_rmnodes(const struct logical_volume *lv) { return fs_del_lv(lv); } static int _lv_has_mknode(const struct logical_volume *lv) { return (lv_is_visible(lv) && (!lv_is_thin_pool(lv) || lv_is_new_thin_pool(lv))); } int dev_manager_mknodes(const struct logical_volume *lv) { struct dm_info dminfo; struct dm_task *dmt; char *name; int r = 0; if (!(name = dm_build_dm_name(lv->vg->cmd->mem, lv->vg->name, lv->name, NULL))) return_0; if (!(dmt = _setup_task_run(DM_DEVICE_MKNODES, &dminfo, name, NULL, 0, 0, 0, 0, 0, 0))) return_0; if (dminfo.exists) { /* read-only component LV is also made visible */ if (_lv_has_mknode(lv) || (dminfo.read_only && lv_is_component(lv))) r = _dev_manager_lv_mknodes(lv); } else r = _dev_manager_lv_rmnodes(lv); dm_task_destroy(dmt); return r; } #ifdef UDEV_SYNC_SUPPORT /* * Until the DM_UEVENT_GENERATED_FLAG was introduced in kernel patch * 856a6f1dbd8940e72755af145ebcd806408ecedd * some operations could not be performed by udev, requiring our fallback code. */ static int _dm_driver_has_stable_udev_support(void) { char vsn[80]; unsigned maj, min, patchlevel; return driver_version(vsn, sizeof(vsn)) && (sscanf(vsn, "%u.%u.%u", &maj, &min, &patchlevel) == 3) && (maj == 4 ? min >= 18 : maj > 4); } static int _check_udev_fallback(struct cmd_context *cmd) { struct config_info *settings = &cmd->current_settings; if (settings->udev_fallback != -1) goto out; /* * Use udev fallback automatically in case udev * is disabled via DM_DISABLE_UDEV environment * variable or udev rules are switched off. */ settings->udev_fallback = !settings->udev_rules ? 1 : find_config_tree_bool(cmd, activation_verify_udev_operations_CFG, NULL); /* Do not rely fully on udev if the udev support is known to be incomplete. */ if (!settings->udev_fallback && !_dm_driver_has_stable_udev_support()) { log_very_verbose("Kernel driver has incomplete udev support so " "LVM will check and perform some operations itself."); settings->udev_fallback = 1; } out: return settings->udev_fallback; } #else /* UDEV_SYNC_SUPPORT */ static int _check_udev_fallback(struct cmd_context *cmd) { /* We must use old node/symlink creation code if not compiled with udev support at all! */ return cmd->current_settings.udev_fallback = 1; } #endif /* UDEV_SYNC_SUPPORT */ static uint16_t _get_udev_flags(struct dev_manager *dm, const struct logical_volume *lv, const char *layer, int noscan, int temporary, int visible_component) { uint16_t udev_flags = 0; /* * Instruct also libdevmapper to disable udev * fallback in accordance to LVM2 settings. */ if (!_check_udev_fallback(dm->cmd)) udev_flags |= DM_UDEV_DISABLE_LIBRARY_FALLBACK; /* * Is this top-level and visible device? * If not, create just the /dev/mapper content. */ /* FIXME: add target's method for this */ if (lv_is_new_thin_pool(lv) || visible_component) /* New thin-pool is regular LV with -tpool UUID suffix. */ udev_flags |= DM_UDEV_DISABLE_DISK_RULES_FLAG | DM_UDEV_DISABLE_OTHER_RULES_FLAG; else if (layer || !lv_is_visible(lv) || lv_is_thin_pool(lv) || lv_is_vdo_pool(lv)) udev_flags |= DM_UDEV_DISABLE_SUBSYSTEM_RULES_FLAG | DM_UDEV_DISABLE_DISK_RULES_FLAG | DM_UDEV_DISABLE_OTHER_RULES_FLAG; /* * There's no need for other udev rules to touch special LVs with * reserved names. We don't need to populate /dev/disk here either. * Even if they happen to be visible and top-level. */ else if (is_reserved_lvname(lv->name)) udev_flags |= DM_UDEV_DISABLE_DISK_RULES_FLAG | DM_UDEV_DISABLE_OTHER_RULES_FLAG; /* * Snapshots and origins could have the same rule applied that will * give symlinks exactly the same name (e.g. a name based on * filesystem UUID). We give preference to origins to make such * naming deterministic (e.g. symlinks in /dev/disk/by-uuid). */ if (lv_is_cow(lv)) udev_flags |= DM_UDEV_LOW_PRIORITY_FLAG; /* * Finally, add flags to disable /dev/mapper and /dev/ content * to be created by udev if it is requested by user's configuration. * This is basically an explicit fallback to old node/symlink creation * without udev. */ if (!dm->cmd->current_settings.udev_rules) udev_flags |= DM_UDEV_DISABLE_DM_RULES_FLAG | DM_UDEV_DISABLE_SUBSYSTEM_RULES_FLAG; /* * LVM subsystem specific flags. */ if (noscan) udev_flags |= DM_SUBSYSTEM_UDEV_FLAG0; if (temporary) udev_flags |= DM_UDEV_DISABLE_DISK_RULES_FLAG | DM_UDEV_DISABLE_OTHER_RULES_FLAG; return udev_flags; } static int _add_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, int origin_only); static int _add_new_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, struct lv_activate_opts *laopts, const char *layer); /* * Check for device holders (ATM used only for removed pvmove targets) * and add them into dtree structures. * When 'laopts != NULL' add them as new nodes - which also corrects READ_AHEAD. * Note: correct table are already explicitelly PRELOADED. */ static int _check_holder(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, struct lv_activate_opts *laopts, uint32_t major, const char *d_name) { const char *default_uuid_prefix = dm_uuid_prefix(); const size_t default_uuid_prefix_len = strlen(default_uuid_prefix); const char *name; const char *uuid; struct dm_info info; struct dm_task *dmt; struct logical_volume *lv_det; union lvid id; int dev, r = 0; errno = 0; dev = strtoll(d_name + 3, NULL, 10); if (errno) { log_error("Failed to parse dm device minor number from %s.", d_name); return 0; } if (!(dmt = _setup_task_run(DM_DEVICE_INFO, &info, NULL, NULL, NULL, major, dev, 0, 0, 0))) return_0; if (info.exists) { uuid = dm_task_get_uuid(dmt); name = dm_task_get_name(dmt); log_debug_activation("Checking holder of %s %s (" FMTu32 ":" FMTu32 ") %s.", display_lvname(lv), uuid, info.major, info.minor, name); /* Skip common uuid prefix */ if (!strncmp(default_uuid_prefix, uuid, default_uuid_prefix_len)) uuid += default_uuid_prefix_len; if (!memcmp(uuid, &lv->vg->id, ID_LEN) && !dm_tree_find_node_by_uuid(dtree, uuid)) { /* trims any UUID suffix (i.e. -cow) */ (void) dm_strncpy((char*)&id, uuid, 2 * sizeof(struct id) + 1); /* If UUID is not yet in dtree, look for matching LV */ if (!(lv_det = find_lv_in_vg_by_lvid(lv->vg, &id))) { log_error("Cannot find holder with device name %s in VG %s.", name, lv->vg->name); goto out; } if (lv_is_cow(lv_det)) lv_det = origin_from_cow(lv_det); log_debug_activation("Found holder %s of %s.", display_lvname(lv_det), display_lvname(lv)); if (!laopts) { if (!_add_lv_to_dtree(dm, dtree, lv_det, 0)) goto_out; } else if (!_add_new_lv_to_dtree(dm, dtree, lv_det, laopts, 0)) goto_out; } } r = 1; out: dm_task_destroy(dmt); return r; } /* * Add exiting devices which holds given LV device open. * This is used in case when metadata already do not contain information * i.e. PVMOVE is being finished and final table is going to be resumed. */ static int _add_holders_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, struct lv_activate_opts *laopts, const struct dm_info *info) { const char *sysfs_dir = dm_sysfs_dir(); char sysfs_path[PATH_MAX]; struct dirent *dirent; DIR *d; int r = 0; /* Sysfs path of holders */ if (dm_snprintf(sysfs_path, sizeof(sysfs_path), "%sblock/dm-" FMTu32 "/holders", sysfs_dir, info->minor) < 0) { log_error("sysfs_path dm_snprintf failed."); return 0; } if (!(d = opendir(sysfs_path))) { log_sys_error("opendir", sysfs_path); return 0; } while ((dirent = readdir(d))) /* Expects minor is added to 'dm-' prefix */ if (!strncmp(dirent->d_name, "dm-", 3) && !_check_holder(dm, dtree, lv, laopts, info->major, dirent->d_name)) goto_out; r = 1; out: if (closedir(d)) log_sys_debug("closedir", "holders"); return r; } static int _add_dev_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, const char *layer) { char *dlid, *name; struct dm_info info, info2; const struct dm_active_device *dev; if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name, layer))) return_0; if (!(dlid = build_dm_uuid(dm->track_pending_delete ? dm->cmd->pending_delete_mem : dm->mem, lv, layer))) return_0; if (!dm->cmd->disable_dm_devs && dm->cmd->cache_dm_devs) { if (!dm_device_list_find_by_uuid(dm->cmd->cache_dm_devs, dlid, &dev)) { log_debug("Cached as not present %s.", name); return 1; } info = (struct dm_info) { .exists = 1, .major = dev->major, .minor = dev->minor, }; log_debug("Cached as present %s %s (%d:%d).", name, dlid, info.major, info.minor); } else if (!_info(dm->cmd, name, dlid, 0, 0, 0, &info, NULL, NULL)) return_0; /* * For top level volumes verify that existing device match * requested major/minor and that major/minor pair is available for use */ if (!layer && lv->major != -1 && lv->minor != -1) { /* * FIXME compare info.major with lv->major if multiple major support */ if (info.exists && ((int) info.minor != lv->minor)) { log_error("Volume %s (%" PRIu32 ":%" PRIu32")" " differs from already active device " "(%" PRIu32 ":%" PRIu32").", display_lvname(lv), lv->major, lv->minor, info.major, info.minor); return 0; } if (!info.exists && _info_by_dev(lv->major, lv->minor, &info2) && info2.exists) { log_error("The requested major:minor pair " "(%" PRIu32 ":%" PRIu32") is already used.", lv->major, lv->minor); return 0; } } if (info.exists && !dm_tree_add_dev_with_udev_flags(dtree, info.major, info.minor, _get_udev_flags(dm, lv, layer, 0, 0, 0))) { log_error("Failed to add device (%" PRIu32 ":%" PRIu32") to dtree.", info.major, info.minor); return 0; } if (info.exists && dm->track_pending_delete) { log_debug_activation("Tracking pending delete for %s (%s).", display_lvname(lv), dlid); if (!str_list_add(dm->cmd->pending_delete_mem, &dm->cmd->pending_delete, dlid)) return_0; } /* * Find holders of existing active LV where name starts with 'pvmove', * but it's not anymore PVMOVE LV and also it's not PVMOVE _mimage */ if (info.exists && !lv_is_pvmove(lv) && !strchr(lv->name, '_') && !strncmp(lv->name, "pvmove", 6)) if (!_add_holders_to_dtree(dm, dtree, lv, NULL, &info)) return_0; return 1; } struct pool_cb_data { struct dev_manager *dm; const struct logical_volume *pool_lv; int skip_zero; /* to skip zeroed device header (check first 64B) */ int exec; /* which binary to call */ int opts; struct { unsigned maj; unsigned min; unsigned patch; } version; const char *global; }; /* * Simple version of check function calling 'tool -V' * * Returns 1 if the tool's version is equal or better to given. * Otherwise it returns 0. */ static int _check_tool_version(struct cmd_context *cmd, const char *tool, unsigned maj, unsigned min, unsigned patch) { const char *argv[] = { tool, "-V", NULL }; struct pipe_data pdata; FILE *f; char buf[128] = { 0 }; char *nl; unsigned v_maj, v_min, v_patch; int ret = 0; if (!(f = pipe_open(cmd, argv, 0, &pdata))) { log_warn("WARNING: Cannot read output from %s.", argv[0]); } else { if (fgets(buf, sizeof(buf) - 1, f) && (sscanf(buf, "%u.%u.%u", &v_maj, &v_min, &v_patch) == 3)) { if ((v_maj > maj) || ((v_maj == maj) && ((v_min > min) || (v_min == min && v_patch >= patch)))) ret = 1; if ((nl = strchr(buf, '\n'))) nl[0] = 0; /* cut newline away */ log_verbose("Found version of %s %s is %s then requested %u.%u.%u.", argv[0], buf, ret ? "better" : "older", maj, min, patch); } else log_warn("WARNING: Cannot parse output '%s' from %s.", buf, argv[0]); (void) pipe_close(&pdata); } return ret; } static int _pool_callback(struct dm_tree_node *node, dm_node_callback_t type, void *cb_data) { int ret, status = 0, fd; const struct dm_config_node *cn; const struct dm_config_value *cv; const struct pool_cb_data *data = cb_data; const struct logical_volume *pool_lv = data->pool_lv; const struct logical_volume *mlv = first_seg(pool_lv)->metadata_lv; struct cmd_context *cmd = pool_lv->vg->cmd; long buf[64 / sizeof(long)]; /* buffer for short disk header (64B) */ int args = 0; char *mpath; const char *argv[19] = { /* Max supported 15 args */ find_config_tree_str_allow_empty(cmd, data->exec, NULL) }; if (!*argv[0]) /* *_check tool is unconfigured/disabled with "" setting */ return 1; if (lv_is_cache_vol(pool_lv)) { if (!(mpath = lv_dmpath_suffix_dup(data->dm->mem, pool_lv, "-cmeta"))) { log_error("Failed to build device path for checking cachevol metadata %s.", display_lvname(pool_lv)); return 0; } } else { if (!(mpath = lv_dmpath_dup(data->dm->mem, mlv))) { log_error("Failed to build device path for checking pool metadata %s.", display_lvname(mlv)); return 0; } } dm_device_list_destroy(&cmd->cache_dm_devs); /* Cache no longer valid */ log_debug("Running check command on %s", mpath); if (data->skip_zero) { if ((fd = open(mpath, O_RDONLY)) < 0) { log_sys_error("open", mpath); return 0; } /* let's assume there is no problem to read 64 bytes */ if (read(fd, buf, sizeof(buf)) < (int)sizeof(buf)) { log_sys_error("read", mpath); if (close(fd)) log_sys_error("close", mpath); return 0; } for (ret = 0; ret < (int) DM_ARRAY_SIZE(buf); ++ret) if (buf[ret]) break; if (close(fd)) log_sys_error("close", mpath); if (ret == (int) DM_ARRAY_SIZE(buf)) { log_debug_activation("Metadata checking skipped, detected empty disk header on %s.", mpath); return 1; } } if (!(cn = find_config_tree_array(cmd, data->opts, NULL))) { log_error(INTERNAL_ERROR "Unable to find configuration for pool check options."); return 0; } for (cv = cn->v; cv && args < 16; cv = cv->next) { if (cv->type != DM_CFG_STRING) { log_error("Invalid string in config file: " "global/%s_check_options.", data->global); return 0; } argv[++args] = cv->v.str; } if (args == 16) { log_error("Too many options for %s command.", argv[0]); return 0; } argv[++args] = mpath; if (!(ret = exec_cmd(cmd, (const char * const *)argv, &status, 0))) { if (status == ENOENT) { log_warn("WARNING: Check is skipped, please install recommended missing binary %s!", argv[0]); return 1; } if ((data->version.maj || data->version.min || data->version.patch) && !_check_tool_version(cmd, argv[0], data->version.maj, data->version.min, data->version.patch)) { log_warn("WARNING: Check is skipped, please upgrade installed version of %s!", argv[0]); return 1; } switch (type) { case DM_NODE_CALLBACK_PRELOADED: log_err_once("Check of pool %s failed (status:%d). " "Manual repair required!", display_lvname(pool_lv), status); break; default: log_warn("WARNING: Integrity check of metadata for pool " "%s failed.", display_lvname(pool_lv)); } /* * FIXME: What should we do here?? * * Maybe mark the node, so it's not activating * as pool but as error/linear and let the * dm tree resolve the issue. */ } return ret; } static int _pool_register_callback(struct dev_manager *dm, struct dm_tree_node *node, const struct logical_volume *lv) { struct pool_cb_data *data; /* Do not skip metadata of testing even for unused thin pools */ #if 0 /* Skip metadata testing for unused thin pool. */ if (lv_is_thin_pool(lv) && (!first_seg(lv)->transaction_id || ((first_seg(lv)->transaction_id == 1) && pool_has_message(first_seg(lv), NULL, 0)))) return 1; #endif if (!(data = dm_pool_zalloc(dm->mem, sizeof(*data)))) { log_error("Failed to allocated path for callback."); return 0; } data->dm = dm; if (lv_is_thin_pool(lv)) { data->pool_lv = lv; data->skip_zero = 1; data->exec = global_thin_check_executable_CFG; data->opts = global_thin_check_options_CFG; data->global = "thin"; } else if (lv_is_cache(lv)) { /* cache pool */ data->pool_lv = first_seg(lv)->pool_lv; data->skip_zero = 1; /* cheap read-error detection */ data->exec = global_cache_check_executable_CFG; data->opts = global_cache_check_options_CFG; data->global = "cache"; if (first_seg(first_seg(lv)->pool_lv)->cache_metadata_format > 1) { data->version.maj = 0; data->version.min = 7; } } else { log_error(INTERNAL_ERROR "Registering unsupported pool callback."); return 0; } dm_tree_node_set_callback(node, _pool_callback, data); return 1; } /* Add special devices _cmeta & _cdata on top of CacheVol to dm tree */ static int _add_cvol_subdev_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, int meta_or_data) { const char *layer = meta_or_data ? "cmeta" : "cdata"; struct dm_pool *mem = dm->track_pending_delete ? dm->cmd->pending_delete_mem : dm->mem; const struct logical_volume *pool_lv = first_seg(lv)->pool_lv; struct lv_segment *lvseg = first_seg(lv); struct dm_info info; char *name ,*dlid; union lvid lvid = { { { "" } } }; memcpy(&lvid.id[0], &lv->vg->id, sizeof(struct id)); /* When ID is provided in form of metadata_id or data_id, otherwise use CVOL ID */ memcpy(&lvid.id[1], (meta_or_data && lvseg->metadata_id) ? lvseg->metadata_id : (lvseg->data_id) ? lvseg->data_id : &pool_lv->lvid.id[1], sizeof(struct id)); if (!(dlid = dm_build_dm_uuid(mem, UUID_PREFIX, (const char *)&lvid.s, layer))) return_0; /* Name is actually not really needed here, but aids debugging... */ if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, pool_lv->name, layer))) return_0; if (!_info(dm->cmd, name, dlid, 1, 0, 0, &info, NULL, NULL)) return_0; if (info.exists) { if (!dm_tree_add_dev_with_udev_flags(dtree, info.major, info.minor, _get_udev_flags(dm, lv, layer, 0, 0, 0))) { log_error("Failed to add device (%" PRIu32 ":%" PRIu32") to dtree.", info.major, info.minor); return 0; } if (dm->track_pending_delete) { log_debug_activation("Tracking pending delete for %s %s (%s).", layer, display_lvname(lv), dlid); if (!str_list_add(mem, &dm->cmd->pending_delete, dlid)) return_0; } } return 1; } /* Declaration to resolve suspend tree and message passing for thin-pool */ static int _add_target_to_dtree(struct dev_manager *dm, struct dm_tree_node *dnode, struct lv_segment *seg, struct lv_activate_opts *laopts); /* * Add LV and any known dependencies */ static int _add_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, int origin_only) { uint32_t s; struct seg_list *sl; struct dm_list *snh; struct lv_segment *seg; struct dm_tree_node *node; const char *uuid; const struct logical_volume *plv; if (lv_is_pvmove(lv) && (dm->track_pvmove_deps == 2)) return 1; /* Avoid rechecking of already seen pvmove LV */ if (lv_is_cache_pool(lv)) { if (!dm_list_empty(&lv->segs_using_this_lv)) { if (!_add_lv_to_dtree(dm, dtree, seg_lv(first_seg(lv), 0), 0)) return_0; if (!_add_lv_to_dtree(dm, dtree, first_seg(lv)->metadata_lv, 0)) return_0; /* Cache pool does not have a real device node */ return 1; } /* Unused cache pool is activated as metadata */ } if (lv_is_cache(lv) && (plv = (first_seg(lv)->pool_lv)) && lv_is_cache_vol(plv)) { if (!_add_cvol_subdev_to_dtree(dm, dtree, lv, 0) || !_add_cvol_subdev_to_dtree(dm, dtree, lv, 1) || !_add_dev_to_dtree(dm, dtree, plv, lv_layer(plv))) return_0; } if (!origin_only && !_add_dev_to_dtree(dm, dtree, lv, NULL)) return_0; /* FIXME Can we avoid doing this every time? */ /* Reused also for lv_is_external_origin(lv) */ if (!_add_dev_to_dtree(dm, dtree, lv, "real")) return_0; if (!origin_only && !_add_dev_to_dtree(dm, dtree, lv, "cow")) return_0; if (origin_only && lv_is_thin_volume(lv)) { if (!_add_dev_to_dtree(dm, dtree, lv, lv_layer(lv))) return_0; #if 0 /* ? Use origin_only to avoid 'deep' thin pool suspend ? */ /* FIXME Implement dm_tree_node_skip_childrens optimisation */ if (!(uuid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if ((node = dm_tree_find_node_by_uuid(dtree, uuid))) dm_tree_node_skip_childrens(node, 1); #endif } if (origin_only && dm->activation && dm->track_external_lv_deps && lv_is_external_origin(lv)) { /* Find possible users of external origin lv */ dm->track_external_lv_deps = 0; /* avoid recursion */ dm_list_iterate_items(sl, &lv->segs_using_this_lv) /* Match only external_lv users */ if ((sl->seg->external_lv == lv) && !_add_lv_to_dtree(dm, dtree, sl->seg->lv, 1)) return_0; dm->track_external_lv_deps = 1; } if (lv_is_thin_pool(lv)) { /* * For both origin_only and !origin_only * skips test for -tpool-real and tpool-cow */ if (!_add_dev_to_dtree(dm, dtree, lv, lv_layer(lv))) return_0; /* * TODO: change API and move this code * Could be easier to handle this in _add_dev_to_dtree() * and base this according to info.exists ? */ if (!dm->activation) { if (!(uuid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if ((node = dm_tree_find_node_by_uuid(dtree, uuid))) { if (origin_only) { struct lv_activate_opts laopts = { .origin_only = 1, .send_messages = 1 /* Node with messages */ }; /* * Add some messsages if right node exist in the table only * when building SUSPEND tree for origin-only thin-pool. * * TODO: Fix call of '_add_target_to_dtree()' to add message * to thin-pool node as we already know the pool node exists * in the table. Any better/cleaner API way ? * * Probably some 'new' target method to add messages for any node? */ if (dm->suspend && !dm_list_empty(&(first_seg(lv)->thin_messages)) && !_add_target_to_dtree(dm, node, first_seg(lv), &laopts)) return_0; } else { /* Setup callback for non-activation partial tree */ /* Activation gets own callback when needed */ /* TODO: extend _cached_dm_info() to return dnode */ if (!_pool_register_callback(dm, node, lv)) return_0; } } } } if (lv_is_vdo_pool(lv)) { /* * For both origin_only and !origin_only * skips test for -vpool-real and vpool-cow */ if (!_add_dev_to_dtree(dm, dtree, lv, lv_layer(lv))) return_0; } if (lv_is_cache(lv)) { if (!origin_only && !dm->activation && !dm->track_pending_delete) { /* Setup callback for non-activation partial tree */ /* Activation gets own callback when needed */ /* TODO: extend _cached_dm_info() to return dnode */ if (!(uuid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; if ((node = dm_tree_find_node_by_uuid(dtree, uuid)) && !_pool_register_callback(dm, node, lv)) return_0; } } /* Add any snapshots of this LV */ if (!origin_only && lv_is_origin(lv)) dm_list_iterate(snh, &lv->snapshot_segs) if (!_add_lv_to_dtree(dm, dtree, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, 0)) return_0; if (dm->activation && !origin_only && lv_is_merging_origin(lv) && !_add_lv_to_dtree(dm, dtree, find_snapshot(lv)->lv, 1)) return_0; /* Add any LVs referencing a PVMOVE LV unless told not to. */ if ((dm->track_pvmove_deps == 1) && lv_is_pvmove(lv)) { dm->track_pvmove_deps = 2; /* Mark as already seen */ dm_list_iterate_items(sl, &lv->segs_using_this_lv) { /* If LV is snapshot COW - whole snapshot needs reload */ plv = lv_is_cow(sl->seg->lv) ? origin_from_cow(sl->seg->lv) : sl->seg->lv; if (!_add_lv_to_dtree(dm, dtree, plv, 0)) return_0; } dm->track_pvmove_deps = 1; } if (!dm->track_pending_delete) dm_list_iterate_items(sl, &lv->segs_using_this_lv) { if (lv_is_pending_delete(sl->seg->lv)) { /* LV is referenced by 'cache pending delete LV */ dm->track_pending_delete = 1; if (!_add_lv_to_dtree(dm, dtree, sl->seg->lv, origin_only)) return_0; dm->track_pending_delete = 0; } } /* Add any LVs used by segments in this LV */ dm_list_iterate_items(seg, &lv->segments) { if (seg->external_lv && dm->track_external_lv_deps && !_add_lv_to_dtree(dm, dtree, seg->external_lv, 1)) /* stack */ return_0; if (seg->log_lv && !_add_lv_to_dtree(dm, dtree, seg->log_lv, 0)) return_0; if (seg->metadata_lv && !_add_lv_to_dtree(dm, dtree, seg->metadata_lv, 0)) return_0; if (seg->writecache && seg_is_writecache(seg)) { if (!_add_lv_to_dtree(dm, dtree, seg->writecache, dm->activation ? origin_only : 1)) return_0; } if (seg->integrity_meta_dev && seg_is_integrity(seg)) { if (!_add_lv_to_dtree(dm, dtree, seg->integrity_meta_dev, dm->activation ? origin_only : 1)) return_0; } if (seg->pool_lv && (lv_is_cache_pool(seg->pool_lv) || lv_is_cache_vol(seg->pool_lv) || dm->track_external_lv_deps) && /* When activating and not origin_only detect linear 'overlay' over pool */ !_add_lv_to_dtree(dm, dtree, seg->pool_lv, dm->activation ? origin_only : 1)) return_0; for (s = 0; s < seg->area_count; s++) { if (seg_type(seg, s) == AREA_LV && seg_lv(seg, s) && /* origin only for cache without pending delete */ (!dm->track_pending_delete || !lv_is_cache(lv)) && !_add_lv_to_dtree(dm, dtree, seg_lv(seg, s), lv_is_vdo_pool(seg_lv(seg, s)) ? 1 : 0)) return_0; if (seg_is_raid_with_meta(seg) && seg->meta_areas && seg_metalv(seg, s) && !_add_lv_to_dtree(dm, dtree, seg_metalv(seg, s), 0)) return_0; } /* When activating, detect merging LV presence */ if (dm->activation && seg->merge_lv && !_add_lv_to_dtree(dm, dtree, seg->merge_lv, 1)) return_0; } return 1; } static struct dm_tree *_create_partial_dtree(struct dev_manager *dm, const struct logical_volume *lv, int origin_only) { struct dm_tree *dtree; if (!(dtree = dm_tree_create())) { log_debug_activation("Partial dtree creation failed for %s.", display_lvname(lv)); return NULL; } dm_tree_set_optional_uuid_suffixes(dtree, &uuid_suffix_list[0]); if (!_add_lv_to_dtree(dm, dtree, lv, (lv_is_origin(lv) || lv_is_thin_volume(lv) || lv_is_thin_pool(lv)) ? origin_only : 0)) goto_bad; return dtree; bad: dm_tree_free(dtree); return NULL; } static char *_add_error_or_zero_device(struct dev_manager *dm, struct dm_tree *dtree, struct lv_segment *seg, int s, int use_zero) { char *dlid, *name; char errid[32]; struct dm_tree_node *node; struct lv_segment *seg_i; struct dm_info info; int segno = -1, i = 0; uint64_t size = (uint64_t) _seg_len(seg) * seg->lv->vg->extent_size; dm_list_iterate_items(seg_i, &seg->lv->segments) { if (seg == seg_i) { segno = i; break; } ++i; } if (segno < 0) { log_error(INTERNAL_ERROR "_add_error_or_zero_device called with bad segment."); return NULL; } sprintf(errid, "missing_%d_%d", segno, s); if (!(dlid = build_dm_uuid(dm->mem, seg->lv, errid))) return_NULL; if (!(name = dm_build_dm_name(dm->mem, seg->lv->vg->name, seg->lv->name, errid))) return_NULL; if (!_info(dm->cmd, name, dlid, 1, 0, 0, &info, NULL, NULL)) return_NULL; if (!info.exists) { /* Create new node */ if (!(node = dm_tree_add_new_dev(dtree, name, dlid, 0, 0, 0, 0, 0))) return_NULL; if (use_zero) { if (!dm_tree_node_add_zero_target(node, size)) return_NULL; } else if (!dm_tree_node_add_error_target(node, size)) return_NULL; } else { /* Already exists */ if (!dm_tree_add_dev(dtree, info.major, info.minor)) { log_error("Failed to add device (%" PRIu32 ":%" PRIu32") to dtree.", info.major, info.minor); return NULL; } } return dlid; } static int _add_error_area(struct dev_manager *dm, struct dm_tree_node *node, struct lv_segment *seg, int s) { char *dlid; uint64_t extent_size = seg->lv->vg->extent_size; int use_zero = !strcmp(dm->cmd->stripe_filler, TARGET_NAME_ZERO) ? 1 : 0; if (!strcmp(dm->cmd->stripe_filler, TARGET_NAME_ERROR) || use_zero) { /* * FIXME, the tree pointer is first field of dm_tree_node, but * we don't have the struct definition available. */ struct dm_tree **tree = (struct dm_tree **) node; if (!(dlid = _add_error_or_zero_device(dm, *tree, seg, s, use_zero))) return_0; if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s))) return_0; } else if (!dm_tree_node_add_target_area(node, dm->cmd->stripe_filler, NULL, UINT64_C(0))) return_0; return 1; } int add_areas_line(struct dev_manager *dm, struct lv_segment *seg, struct dm_tree_node *node, uint32_t start_area, uint32_t areas) { uint64_t extent_size = seg->lv->vg->extent_size; uint32_t s; char *dlid; struct stat info; const char *name; unsigned num_error_areas = 0; unsigned num_existing_areas = 0; /* FIXME Avoid repeating identical stat in dm_tree_node_add_target_area */ for (s = start_area; s < areas; s++) { /* FIXME: dev_name() does not return NULL! It needs to check if dm_list_empty(&dev->aliases) but this knot of logic is too complex to pull apart without careful deconstruction. */ if ((seg_type(seg, s) == AREA_PV && (!seg_pvseg(seg, s) || !seg_pv(seg, s) || !seg_dev(seg, s) || !(name = dev_name(seg_dev(seg, s))) || !*name || stat(name, &info) < 0 || !S_ISBLK(info.st_mode))) || (seg_type(seg, s) == AREA_LV && !seg_lv(seg, s))) { if (!seg->lv->vg->cmd->partial_activation) { if (!seg->lv->vg->cmd->degraded_activation || !lv_is_raid_type(seg->lv)) { log_error("Aborting. LV %s is now incomplete " "and '--activationmode partial' was not specified.", display_lvname(seg->lv)); return 0; } } if (!_add_error_area(dm, node, seg, s)) return_0; num_error_areas++; } else if (seg_type(seg, s) == AREA_PV) { struct device *dev = seg_dev(seg, s); name = dm_list_empty(&dev->aliases) ? NULL : dev_name(dev); if (!dm_tree_node_add_target_area(node, name, NULL, (seg_pv(seg, s)->pe_start + (extent_size * seg_pe(seg, s))))) return_0; num_existing_areas++; } else if (seg_is_raid(seg)) { /* * RAID can handle unassigned areas. It simple puts * '- -' in for the metadata/data device pair. This * is a valid way to indicate to the RAID target that * the device is missing. * * If an image is marked as VISIBLE_LV and !LVM_WRITE, * it means the device has temporarily been extracted * from the array. It may come back at a future date, * so the bitmap must track differences. Again, '- -' * is used in the CTR table. */ if ((seg_type(seg, s) == AREA_UNASSIGNED) || (lv_is_visible(seg_lv(seg, s)) && !(seg_lv(seg, s)->status & LVM_WRITE))) { /* One each for metadata area and data area */ if (!dm_tree_node_add_null_area(node, 0) || !dm_tree_node_add_null_area(node, 0)) return_0; continue; } if (seg->meta_areas && seg_metalv(seg, s)) { if (!(dlid = build_dm_uuid(dm->mem, seg_metalv(seg, s), NULL))) return_0; if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_metale(seg, s))) return_0; } else if (!dm_tree_node_add_null_area(node, 0)) return_0; if (!(dlid = build_dm_uuid(dm->mem, seg_lv(seg, s), NULL))) return_0; if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s))) return_0; } else if (seg_type(seg, s) == AREA_LV) { if (!(dlid = build_dm_uuid(dm->mem, seg_lv(seg, s), NULL))) return_0; if (!dm_tree_node_add_target_area(node, NULL, dlid, extent_size * seg_le(seg, s))) return_0; } else { log_error(INTERNAL_ERROR "Unassigned area found in LV %s.", display_lvname(seg->lv)); return 0; } } if (num_error_areas) { /* Thins currently do not support partial activation */ if (lv_is_thin_type(seg->lv)) { log_error("Cannot activate %s: pool incomplete.", display_lvname(seg->lv)); return 0; } } return 1; } static int _add_layer_target_to_dtree(struct dev_manager *dm, struct dm_tree_node *dnode, const struct logical_volume *lv) { const char *layer_dlid; if (!(layer_dlid = build_dm_uuid(dm->mem, lv, lv_layer(lv)))) return_0; /* Add linear mapping over layered LV */ /* From VDO layer expose ONLY vdo pool header, we would need to use virtual size otherwise */ if (!add_linear_area_to_dtree(dnode, lv_is_vdo_pool(lv) ? 8 : lv->size, lv->vg->extent_size, lv->vg->cmd->use_linear_target, lv->vg->name, lv->name) || !dm_tree_node_add_target_area(dnode, NULL, layer_dlid, 0)) return_0; return 1; } static int _add_origin_target_to_dtree(struct dev_manager *dm, struct dm_tree_node *dnode, const struct logical_volume *lv) { const char *real_dlid; if (!(real_dlid = build_dm_uuid(dm->mem, lv, "real"))) return_0; if (!dm_tree_node_add_snapshot_origin_target(dnode, lv->size, real_dlid)) return_0; return 1; } static int _add_snapshot_merge_target_to_dtree(struct dev_manager *dm, struct dm_tree_node *dnode, const struct logical_volume *lv) { const char *origin_dlid, *cow_dlid, *merge_dlid; struct lv_segment *merging_snap_seg = find_snapshot(lv); if (!lv_is_merging_origin(lv)) { log_error(INTERNAL_ERROR "LV %s is not merging snapshot.", display_lvname(lv)); return 0; } if (!(origin_dlid = build_dm_uuid(dm->mem, lv, "real"))) return_0; if (!(cow_dlid = build_dm_uuid(dm->mem, merging_snap_seg->cow, "cow"))) return_0; if (!(merge_dlid = build_dm_uuid(dm->mem, merging_snap_seg->cow, NULL))) return_0; if (!dm_tree_node_add_snapshot_merge_target(dnode, lv->size, origin_dlid, cow_dlid, merge_dlid, merging_snap_seg->chunk_size)) return_0; return 1; } static int _add_snapshot_target_to_dtree(struct dev_manager *dm, struct dm_tree_node *dnode, const struct logical_volume *lv, struct lv_activate_opts *laopts) { const char *origin_dlid; const char *cow_dlid; struct lv_segment *snap_seg; uint64_t size; if (!(snap_seg = find_snapshot(lv))) { log_error("Couldn't find snapshot for '%s'.", display_lvname(lv)); return 0; } if (!(origin_dlid = build_dm_uuid(dm->mem, snap_seg->origin, "real"))) return_0; if (!(cow_dlid = build_dm_uuid(dm->mem, snap_seg->cow, "cow"))) return_0; size = (uint64_t) snap_seg->len * snap_seg->origin->vg->extent_size; if (!laopts->no_merging && lv_is_merging_cow(lv)) { /* cow is to be merged so load the error target */ if (!dm_tree_node_add_error_target(dnode, size)) return_0; } else if (!dm_tree_node_add_snapshot_target(dnode, size, origin_dlid, cow_dlid, 1, snap_seg->chunk_size)) return_0; return 1; } static int _add_target_to_dtree(struct dev_manager *dm, struct dm_tree_node *dnode, struct lv_segment *seg, struct lv_activate_opts *laopts) { uint64_t extent_size = seg->lv->vg->extent_size; if (!seg->segtype->ops->add_target_line) { log_error(INTERNAL_ERROR "_emit_target cannot handle " "segment type %s.", lvseg_name(seg)); return 0; } return seg->segtype->ops->add_target_line(dm, dm->mem, dm->cmd, &dm->target_state, seg, laopts, dnode, extent_size * _seg_len(seg), &dm->pvmove_mirror_count); } static int _add_new_external_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, struct logical_volume *external_lv, struct lv_activate_opts *laopts) { struct seg_list *sl; /* Do not want to recursively add externals again */ if (!dm->track_external_lv_deps) return 1; /* * Any LV can have only 1 external origin, so we will * process all LVs related to this LV, and we want to * skip repeated invocation of external lv processing */ dm->track_external_lv_deps = 0; log_debug_activation("Adding external origin LV %s and all active users.", display_lvname(external_lv)); if (!_add_new_lv_to_dtree(dm, dtree, external_lv, laopts, lv_layer(external_lv))) return_0; /* * Add all ACTIVE LVs using this external origin LV. This is * needed because of conversion of thin which could have been * also an old-snapshot to external origin. */ dm_list_iterate_items(sl, &external_lv->segs_using_this_lv) if ((sl->seg->external_lv == external_lv) && /* Add only active layered devices (also avoids loop) */ _cached_dm_info(dm->mem, dtree, sl->seg->lv, lv_layer(sl->seg->lv)) && !_add_new_lv_to_dtree(dm, dtree, sl->seg->lv, laopts, lv_layer(sl->seg->lv))) return_0; log_debug_activation("Finished adding external origin LV %s and all active users.", display_lvname(external_lv)); dm->track_external_lv_deps = 1; return 1; } static int _add_segment_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, struct dm_tree_node *dnode, struct lv_segment *seg, struct lv_activate_opts *laopts, const char *layer) { uint32_t s; struct lv_segment *seg_present; const struct segment_type *segtype; const char *target_name; /* Ensure required device-mapper targets are loaded */ seg_present = find_snapshot(seg->lv) ? : seg; segtype = seg_present->segtype; target_name = (segtype->ops->target_name ? segtype->ops->target_name(seg_present, laopts) : segtype->name); log_debug_activation("Checking kernel supports %s segment type for %s%s%s", target_name, display_lvname(seg->lv), layer ? "-" : "", layer ? : ""); if (segtype->ops->target_present && !segtype->ops->target_present(seg_present->lv->vg->cmd, seg_present, NULL)) { log_error("Can't process LV %s: %s target support missing " "from kernel?", display_lvname(seg->lv), target_name); return 0; } /* Add external origin layer */ if (seg->external_lv && !_add_new_external_lv_to_dtree(dm, dtree, seg->external_lv, laopts)) return_0; /* Add mirror log */ if (seg->log_lv && !_add_new_lv_to_dtree(dm, dtree, seg->log_lv, laopts, NULL)) return_0; /* Add pool metadata */ if (seg->metadata_lv && !_add_new_lv_to_dtree(dm, dtree, seg->metadata_lv, laopts, NULL)) return_0; /* Add pool layer */ if (seg->pool_lv && !laopts->origin_only && !_add_new_lv_to_dtree(dm, dtree, seg->pool_lv, laopts, lv_layer(seg->pool_lv))) return_0; if (seg->writecache && !laopts->origin_only && !_add_new_lv_to_dtree(dm, dtree, seg->writecache, laopts, lv_layer(seg->writecache))) return_0; if (seg->integrity_meta_dev && !laopts->origin_only && !_add_new_lv_to_dtree(dm, dtree, seg->integrity_meta_dev, laopts, lv_layer(seg->integrity_meta_dev))) return_0; /* Add any LVs used by this segment */ for (s = 0; s < seg->area_count; ++s) { if ((seg_type(seg, s) == AREA_LV) && /* do not bring up tracked image */ !lv_is_raid_image_with_tracking(seg_lv(seg, s)) && /* origin only for cache without pending delete */ (!dm->track_pending_delete || !seg_is_cache(seg)) && !_add_new_lv_to_dtree(dm, dtree, seg_lv(seg, s), laopts, lv_is_vdo_pool(seg_lv(seg, s)) ? lv_layer(seg_lv(seg, s)) : NULL)) return_0; if (seg_is_raid_with_meta(seg) && seg->meta_areas && seg_metalv(seg, s) && !lv_is_raid_image_with_tracking(seg_lv(seg, s)) && !_add_new_lv_to_dtree(dm, dtree, seg_metalv(seg, s), laopts, NULL)) return_0; } if (dm->track_pending_delete) { /* Replace target and all its used devs with error mapping */ log_debug_activation("Using error for pending delete %s.", display_lvname(seg->lv)); if (!dm_tree_node_add_error_target(dnode, (uint64_t)seg->lv->vg->extent_size * _seg_len(seg))) return_0; } else if (!_add_target_to_dtree(dm, dnode, seg, laopts)) return_0; return 1; } static int _add_new_lv_to_dtree(struct dev_manager *dm, struct dm_tree *dtree, const struct logical_volume *lv, struct lv_activate_opts *laopts, const char *layer) { struct lv_segment *seg; struct lv_layer *lvlayer; struct seg_list *sl; struct dm_list *snh; struct dm_tree_node *dnode; const struct dm_info *dinfo; char *name, *dlid; uint32_t max_stripe_size = UINT32_C(0); uint32_t read_ahead = lv->read_ahead; uint32_t read_ahead_flags = UINT32_C(0); int save_pending_delete = dm->track_pending_delete; int merge_in_progress = 0; if (!(lvlayer = dm_pool_alloc(dm->mem, sizeof(*lvlayer)))) { log_error("_add_new_lv_to_dtree: pool alloc failed for %s %s.", display_lvname(lv), layer); return 0; } lvlayer->lv = lv; lvlayer->visible_component = (laopts->component_lv == lv) ? 1 : 0; log_debug_activation("Adding new LV %s%s%s to dtree", display_lvname(lv), layer ? "-" : "", layer ? : ""); /* LV with pending delete is never put new into a table */ if (lv_is_pending_delete(lv) && !_cached_dm_info(dm->mem, dtree, lv, NULL)) return 1; /* Replace with error only when already exists */ if (lv_is_cache_pool(lv) && !dm_list_empty(&lv->segs_using_this_lv)) { /* cache pool is 'meta' LV and does not have a real device node */ if (!_add_new_lv_to_dtree(dm, dtree, seg_lv(first_seg(lv), 0), laopts, NULL)) return_0; if (!_add_new_lv_to_dtree(dm, dtree, first_seg(lv)->metadata_lv, laopts, NULL)) return_0; return 1; } if (lv_is_cache(lv) && lv_is_cache_vol(first_seg(lv)->pool_lv)) { struct logical_volume *pool_lv = first_seg(lv)->pool_lv; struct lv_segment *lvseg = first_seg(lv); struct volume_group *vg = lv->vg; struct dm_tree_node *dnode_meta; struct dm_tree_node *dnode_data; union lvid lvid_meta; union lvid lvid_data; char *name_meta; char *name_data; char *dlid_meta; char *dlid_data; char *dlid_pool; uint64_t meta_size = first_seg(lv)->metadata_len; uint64_t data_size = first_seg(lv)->data_len; uint16_t udev_flags = _get_udev_flags(dm, lv, layer, laopts->noscan, laopts->temporary, 0); if (lv_is_pending_delete(lvseg->lv)) dm->track_pending_delete = 1; log_debug("Add cachevol %s to dtree before cache %s.", pool_lv->name, lv->name); if (!_add_new_lv_to_dtree(dm, dtree, pool_lv, laopts, lv_layer(pool_lv))) { log_error("Failed to add cachevol to dtree before cache."); return 0; } memset(&lvid_meta, 0, sizeof(lvid_meta)); memset(&lvid_data, 0, sizeof(lvid_meta)); memcpy(&lvid_meta.id[0], &vg->id, sizeof(struct id)); memcpy(&lvid_meta.id[1], lvseg->metadata_id ? : &pool_lv->lvid.id[1], sizeof(struct id)); memcpy(&lvid_data.id[0], &vg->id, sizeof(struct id)); memcpy(&lvid_data.id[1], lvseg->data_id ? : &pool_lv->lvid.id[1], sizeof(struct id)); if (!(dlid_meta = dm_build_dm_uuid(dm->mem, UUID_PREFIX, (const char *)&lvid_meta.s, "cmeta"))) return_0; if (!(dlid_data = dm_build_dm_uuid(dm->mem, UUID_PREFIX, (const char *)&lvid_data.s, "cdata"))) return_0; if (!(name_meta = dm_build_dm_name(dm->mem, vg->name, pool_lv->name, "cmeta"))) return_0; if (!(name_data = dm_build_dm_name(dm->mem, vg->name, pool_lv->name, "cdata"))) return_0; if (!(dlid_pool = build_dm_uuid(dm->mem, pool_lv, NULL))) return_0; /* add meta dnode */ if (!(dnode_meta = dm_tree_add_new_dev_with_udev_flags(dtree, name_meta, dlid_meta, -1, -1, read_only_lv(lv, laopts, layer), ((lv->vg->status & PRECOMMITTED) | laopts->revert) ? 1 : 0, lvlayer, udev_flags))) return_0; if (dm->track_pending_delete) { log_debug_activation("Using error for pending meta delete %s.", display_lvname(lv)); if (!dm_tree_node_add_error_target(dnode_meta, meta_size)) return_0; } else { /* add load_segment to meta dnode: linear, size of meta area */ if (!add_linear_area_to_dtree(dnode_meta, meta_size, lv->vg->extent_size, lv->vg->cmd->use_linear_target, lv->vg->name, lv->name)) return_0; /* add seg_area to prev load_seg: offset 0 maps to cachepool lv offset 0 */ if (!dm_tree_node_add_target_area(dnode_meta, NULL, dlid_pool, 0)) return_0; } /* add data dnode */ if (!(dnode_data = dm_tree_add_new_dev_with_udev_flags(dtree, name_data, dlid_data, -1, -1, read_only_lv(lv, laopts, layer), ((lv->vg->status & PRECOMMITTED) | laopts->revert) ? 1 : 0, lvlayer, udev_flags))) return_0; if (dm->track_pending_delete) { log_debug_activation("Using error for pending data delete %s.", display_lvname(lv)); if (!dm_tree_node_add_error_target(dnode_data, data_size)) return_0; } else { /* add load_segment to data dnode: linear, size of data area */ if (!add_linear_area_to_dtree(dnode_data, data_size, lv->vg->extent_size, lv->vg->cmd->use_linear_target, lv->vg->name, lv->name)) return_0; /* add seg_area to prev load_seg: offset 0 maps to cachepool lv after meta */ if (!dm_tree_node_add_target_area(dnode_data, NULL, dlid_pool, meta_size)) return_0; } } /* FIXME Seek a simpler way to lay out the snapshot-merge tree. */ if (!layer && lv_is_merging_origin(lv)) { seg = find_snapshot(lv); /* * Prevent merge if merge isn't currently possible: * either origin or merging snapshot are open * - for old snaps use "snapshot-merge" if it is already in use * - open_count is always retrieved (as of dm-ioctl 4.7.0) * so just use the tree's existing nodes' info */ if ((dinfo = _cached_dm_info(dm->mem, dtree, lv, NULL))) { /* Merging origin LV is present, check if mergins is already running. */ if ((seg_is_thin_volume(seg) && _lv_has_thin_device_id(dm->mem, lv, NULL, seg->device_id)) || (!seg_is_thin_volume(seg) && lv_has_target_type(dm->mem, lv, NULL, TARGET_NAME_SNAPSHOT_MERGE))) { log_debug_activation("Merging of snapshot volume %s to origin %s is in progress.", display_lvname(seg->lv), display_lvname(seg->lv)); merge_in_progress = 1; /* Merge is already running */ } /* Merge is not yet running, so check if it can be started */ else if (laopts->resuming) { log_debug_activation("Postponing pending snapshot merge for origin %s, " "merge was not started before suspend.", display_lvname(lv)); laopts->no_merging = 1; /* Cannot be reloaded in suspend */ } /* Non-resuming merge requires origin to be unused */ else if (dinfo->open_count) { log_debug_activation("Postponing pending snapshot merge for origin %s, " "origin volume is opened.", display_lvname(lv)); laopts->no_merging = 1; } } /* If merge would be still undecided, look as snapshot */ if (!merge_in_progress && !laopts->no_merging && (dinfo = _cached_dm_info(dm->mem, dtree, seg_is_thin_volume(seg) ? seg->lv : seg->cow, NULL))) { if (seg_is_thin_volume(seg)) { /* Active thin snapshot prevents merge */ log_debug_activation("Postponing pending snapshot merge for origin volume %s, " "merging thin snapshot volume %s is active.", display_lvname(lv), display_lvname(seg->lv)); laopts->no_merging = 1; } else if (dinfo->open_count) { log_debug_activation("Postponing pending snapshot merge for origin volume %s, " "merging snapshot volume %s is opened.", display_lvname(lv), display_lvname(seg->lv)); laopts->no_merging = 1; } } } if (!(name = dm_build_dm_name(dm->mem, lv->vg->name, lv->name, layer))) return_0; /* Even unused thin-pool still needs to get layered UUID -suffix */ if (!layer && lv_is_new_thin_pool(lv)) layer = lv_layer(lv); /* Adds -real to the dm uuid of wcorig LV. */ if (!layer && lv_is_writecache_origin(lv)) layer = lv_layer(lv); /* "real" */ if (!(dlid = build_dm_uuid(dm->mem, lv, layer))) return_0; /* We've already processed this node if it already has a context ptr */ if ((dnode = dm_tree_find_node_by_uuid(dtree, dlid)) && dm_tree_node_get_context(dnode)) return 1; lvlayer->lv = lv; lvlayer->visible_component = (laopts->component_lv == lv) ? 1 : 0; /* * Add LV to dtree. * If we're working with precommitted metadata, clear any * existing inactive table left behind. * Major/minor settings only apply to the visible layer. */ /* FIXME Move the clear from here until later, so we can leave * identical inactive tables untouched. (For pvmove.) */ if (!(dnode = dm_tree_add_new_dev_with_udev_flags(dtree, name, dlid, layer ? UINT32_C(0) : (uint32_t) lv->major, layer ? UINT32_C(0) : (uint32_t) lv->minor, read_only_lv(lv, laopts, layer), ((lv->vg->status & PRECOMMITTED) | laopts->revert) ? 1 : 0, lvlayer, _get_udev_flags(dm, lv, layer, laopts->noscan, laopts->temporary, lvlayer->visible_component)))) return_0; /* Store existing name so we can do rename later */ lvlayer->old_name = dm_tree_node_get_name(dnode); /* Create table */ dm->pvmove_mirror_count = 0u; if (lv_is_pending_delete(lv)) /* Handle LVs with pending delete */ /* Fow now used only by cache segtype, TODO snapshots */ dm->track_pending_delete = 1; /* This is unused cache-pool - make metadata accessible */ if (lv_is_cache_pool(lv)) lv = first_seg(lv)->metadata_lv; /* If this is a snapshot origin, add real LV */ /* If this is a snapshot origin + merging snapshot, add cow + real LV */ /* Snapshot origin could be also external origin */ if (lv_is_origin(lv) && !layer) { if (!_add_new_lv_to_dtree(dm, dtree, lv, laopts, "real")) return_0; if (!laopts->no_merging && lv_is_merging_origin(lv)) { if (!_add_new_lv_to_dtree(dm, dtree, find_snapshot(lv)->cow, laopts, "cow")) return_0; /* * Must also add "real" LV for use when * snapshot-merge target is added */ if (!_add_snapshot_merge_target_to_dtree(dm, dnode, lv)) return_0; } else if (!_add_origin_target_to_dtree(dm, dnode, lv)) return_0; /* Add any snapshots of this LV */ dm_list_iterate(snh, &lv->snapshot_segs) if (!_add_new_lv_to_dtree(dm, dtree, dm_list_struct_base(snh, struct lv_segment, origin_list)->cow, laopts, NULL)) return_0; } else if (lv_is_cow(lv) && !layer) { if (!_add_new_lv_to_dtree(dm, dtree, lv, laopts, "cow")) return_0; if (!_add_snapshot_target_to_dtree(dm, dnode, lv, laopts)) return_0; } else if (!layer && ((lv_is_thin_pool(lv) && !lv_is_new_thin_pool(lv)) || lv_is_vdo_pool(lv) || lv_is_external_origin(lv))) { /* External origin or 'used' Thin pool or VDO pool is using layer */ if (!_add_new_lv_to_dtree(dm, dtree, lv, laopts, lv_layer(lv))) return_0; if (!_add_layer_target_to_dtree(dm, dnode, lv)) return_0; } else { /* Add 'real' segments for LVs */ dm_list_iterate_items(seg, &lv->segments) { if (!_add_segment_to_dtree(dm, dtree, dnode, seg, laopts, layer)) return_0; if (max_stripe_size < seg->stripe_size * seg->area_count) max_stripe_size = seg->stripe_size * seg->area_count; } if (!layer && lv_is_vdo_pool(lv) && !_add_layer_target_to_dtree(dm, dnode, lv)) return_0; } /* Setup thin pool callback */ if (lv_is_thin_pool(lv) && layer && !_pool_register_callback(dm, dnode, lv)) return_0; if (lv_is_cache(lv) && !lv_is_cache_vol(first_seg(lv)->pool_lv) && /* Register callback only for layer activation or non-layered cache LV */ (layer || !lv_layer(lv)) && /* Register callback when metadata LV is NOT already active */ !_cached_dm_info(dm->mem, dtree, first_seg(first_seg(lv)->pool_lv)->metadata_lv, NULL) && !_pool_register_callback(dm, dnode, lv)) return_0; if (lv_is_cache(lv) && lv_is_cache_vol(first_seg(lv)->pool_lv) && /* Register callback only for layer activation or non-layered cache LV */ (layer || !lv_layer(lv)) && /* Register callback when cachevol LV is NOT already active */ !_cached_dm_info(dm->mem, dtree, first_seg(lv)->pool_lv, NULL) && !_pool_register_callback(dm, dnode, lv)) return_0; /* * Update tables for ANY PVMOVE holders for active LV where the name starts with 'pvmove', * but it's not anymore PVMOVE LV and also it's not a PVMOVE _mimage LV. * When resume happens, tables MUST be already preloaded with correct entries! * (since we can't preload different table while devices are suspended) */ if (!lv_is_pvmove(lv) && !strncmp(lv->name, "pvmove", 6) && !strchr(lv->name, '_') && (dinfo = _cached_dm_info(dm->mem, dtree, lv, NULL))) if (!_add_holders_to_dtree(dm, dtree, lv, laopts, dinfo)) return_0; if (read_ahead == DM_READ_AHEAD_AUTO) { /* we need RA at least twice a whole stripe - see the comment in md/raid0.c */ read_ahead = max_stripe_size * 2; /* FIXME: layered device read-ahead */ if (!read_ahead) lv_calculate_readahead(lv, &read_ahead); read_ahead_flags = DM_READ_AHEAD_MINIMUM_FLAG; } dm_tree_node_set_read_ahead(dnode, read_ahead, read_ahead_flags); /* Add any LVs referencing a PVMOVE LV unless told not to */ if (dm->track_pvmove_deps && lv_is_pvmove(lv)) dm_list_iterate_items(sl, &lv->segs_using_this_lv) if (!_add_new_lv_to_dtree(dm, dtree, sl->seg->lv, laopts, NULL)) return_0; dm->track_pending_delete = save_pending_delete; /* restore */ return 1; } /* FIXME: symlinks should be created/destroyed at the same time * as the kernel devices but we can't do that from within libdevmapper * at present so we must walk the tree twice instead. */ /* * Create LV symlinks for children of supplied root node. */ static int _create_lv_symlinks(struct dev_manager *dm, struct dm_tree_node *root) { void *handle = NULL; struct dm_tree_node *child; struct lv_layer *lvlayer; char *old_vgname, *old_lvname, *old_layer; char *new_vgname, *new_lvname, *new_layer; const char *name; int r = 1; /* Nothing to do if udev fallback is disabled. */ if (!_check_udev_fallback(dm->cmd)) { fs_set_create(); return 1; } while ((child = dm_tree_next_child(&handle, root, 0))) { if (!(lvlayer = dm_tree_node_get_context(child))) continue; /* Detect rename */ name = dm_tree_node_get_name(child); if (name && lvlayer->old_name && *lvlayer->old_name && strcmp(name, lvlayer->old_name)) { if (!dm_split_lvm_name(dm->mem, lvlayer->old_name, &old_vgname, &old_lvname, &old_layer)) { log_error("_create_lv_symlinks: Couldn't split up old device name %s.", lvlayer->old_name); return 0; } if (!dm_split_lvm_name(dm->mem, name, &new_vgname, &new_lvname, &new_layer)) { log_error("_create_lv_symlinks: Couldn't split up new device name %s.", name); return 0; } if (!fs_rename_lv(lvlayer->lv, name, old_vgname, old_lvname)) r = 0; continue; } if (_lv_has_mknode(lvlayer->lv) || lvlayer->visible_component) { if (!_dev_manager_lv_mknodes(lvlayer->lv)) r = 0; continue; } if (!_dev_manager_lv_rmnodes(lvlayer->lv)) r = 0; } return r; } /* * Remove LV symlinks for children of supplied root node. */ static int _remove_lv_symlinks(struct dev_manager *dm, struct dm_tree_node *root) { void *handle = NULL; struct dm_tree_node *child; char *vgname, *lvname, *layer; int r = 1; /* Nothing to do if udev fallback is disabled. */ if (!_check_udev_fallback(dm->cmd)) return 1; while ((child = dm_tree_next_child(&handle, root, 0))) { if (!dm_split_lvm_name(dm->mem, dm_tree_node_get_name(child), &vgname, &lvname, &layer)) { r = 0; continue; } if (!*vgname) continue; /* only top level layer has symlinks */ if (*layer) continue; fs_del_lv_byname(dm->cmd->dev_dir, vgname, lvname, dm->cmd->current_settings.udev_rules); } return r; } static int _clean_tree(struct dev_manager *dm, struct dm_tree_node *root, const char *non_toplevel_tree_dlid) { void *handle = NULL; struct dm_tree_node *child; char *vgname, *lvname, *layer; const char *name, *uuid; struct dm_str_list *dl; while ((child = dm_tree_next_child(&handle, root, 0))) { if (!(name = dm_tree_node_get_name(child))) continue; if (!(uuid = dm_tree_node_get_uuid(child))) continue; if (!dm_split_lvm_name(dm->mem, name, &vgname, &lvname, &layer)) { log_error("_clean_tree: Couldn't split up device name %s.", name); return 0; } /* Not meant to be top level? */ if (!*layer) continue; /* If operation was performed on a partial tree, don't remove it */ if (non_toplevel_tree_dlid && !strcmp(non_toplevel_tree_dlid, uuid)) continue; if (!(uuid = dm_pool_strdup(dm->cmd->pending_delete_mem, uuid))) { log_error("_clean_tree: Failed to duplicate uuid."); return 0; } if (!str_list_add(dm->cmd->pending_delete_mem, &dm->cmd->pending_delete, uuid)) return_0; } /* Deactivate any tracked pending delete nodes */ if (!dm_list_empty(&dm->cmd->pending_delete) && !dm_get_suspended_counter()) { fs_unlock(); dm_tree_set_cookie(root, fs_get_cookie()); dm_list_iterate_items(dl, &dm->cmd->pending_delete) { log_debug_activation("Deleting tracked UUID %s.", dl->str); if (!dm_tree_deactivate_children(root, dl->str, strlen(dl->str))) return_0; } dm_list_init(&dm->cmd->pending_delete); dm_pool_empty(dm->cmd->pending_delete_mem); } return 1; } static int _tree_action(struct dev_manager *dm, const struct logical_volume *lv, struct lv_activate_opts *laopts, action_t action) { static const char _action_names[][24] = { "PRELOAD", "ACTIVATE", "DEACTIVATE", "SUSPEND", "SUSPEND_WITH_LOCKFS", "CLEAN" }; const size_t DLID_SIZE = ID_LEN + sizeof(UUID_PREFIX) - 1; struct dm_tree *dtree; struct dm_tree_node *root; char *dlid; int r = 0; unsigned tmp_state; if (action < DM_ARRAY_SIZE(_action_names)) log_debug_activation("Creating %s%s tree for %s.", _action_names[action], (laopts->origin_only) ? " origin-only" : "", display_lvname(lv)); /* Some LV cannot be used for top level tree */ /* TODO: add more.... */ if (lv_is_cache_pool(lv) && !dm_list_empty(&lv->segs_using_this_lv)) { log_error(INTERNAL_ERROR "Cannot create tree for %s.", display_lvname(lv)); return 0; } /* Some targets may build bigger tree for activation */ dm->activation = ((action == PRELOAD) || (action == ACTIVATE)); dm->suspend = (action == SUSPEND_WITH_LOCKFS) || (action == SUSPEND); dm->track_external_lv_deps = 1; /* ATM do not use caching for anything else then striped target. * And also skip for CLEAN action */ tmp_state = dm->cmd->disable_dm_devs; if (!seg_is_striped_target(first_seg(lv)) || (action == CLEAN)) dm->cmd->disable_dm_devs = 1; dtree = _create_partial_dtree(dm, lv, laopts->origin_only); dm->cmd->disable_dm_devs = tmp_state; if (!dtree) return_0; if (!(root = dm_tree_find_node(dtree, 0, 0))) { log_error("Lost dependency tree root node."); goto out_no_root; } /* Restore fs cookie */ dm_tree_set_cookie(root, fs_get_cookie()); if (!(dlid = build_dm_uuid(dm->mem, lv, laopts->origin_only ? lv_layer(lv) : NULL))) goto_out; /* Only process nodes with uuid of "LVM-" plus VG id. */ switch(action) { case CLEAN: if (retry_deactivation()) dm_tree_retry_remove(root); /* Deactivate any unused non-toplevel nodes */ if (!_clean_tree(dm, root, laopts->origin_only ? dlid : NULL)) goto_out; break; case DEACTIVATE: if (retry_deactivation()) dm_tree_retry_remove(root); /* Deactivate LV and all devices it references that nothing else has open. */ if (!dm_tree_deactivate_children(root, dlid, DLID_SIZE)) goto_out; if (!_remove_lv_symlinks(dm, root)) log_warn("Failed to remove all device symlinks associated with %s.", display_lvname(lv)); break; case SUSPEND: dm_tree_skip_lockfs(root); if (!dm->flush_required) dm_tree_use_no_flush_suspend(root); /* Fall through */ case SUSPEND_WITH_LOCKFS: if (!dm_tree_suspend_children(root, dlid, DLID_SIZE)) goto_out; break; case PRELOAD: case ACTIVATE: /* Add all required new devices to tree */ if (!_add_new_lv_to_dtree(dm, dtree, lv, laopts, (lv_is_origin(lv) && laopts->origin_only) ? "real" : (laopts->origin_only && (lv_is_thin_pool(lv) || lv_is_vdo_pool(lv))) ? lv_layer(lv) : NULL)) goto_out; /* Preload any devices required before any suspensions */ if (!dm_tree_preload_children(root, dlid, DLID_SIZE)) goto_out; if ((dm_tree_node_size_changed(root) < 0)) dm->flush_required = 1; /* Currently keep the code require flush for any * non 'thin pool/volume' and size increase */ else if (!lv_is_thin_volume(lv) && !lv_is_thin_pool(lv) && !lv_is_vdo(lv) && !lv_is_vdo_pool(lv) && dm_tree_node_size_changed(root)) dm->flush_required = 1; if (action == ACTIVATE) { if (!dm_tree_activate_children(root, dlid, DLID_SIZE)) goto_out; if (!_create_lv_symlinks(dm, root)) log_warn("Failed to create symlinks for %s.", display_lvname(lv)); } break; default: log_error(INTERNAL_ERROR "_tree_action: Action %u not supported.", action); goto out; } r = 1; out: /* Save fs cookie for udev settle, do not wait here */ fs_set_cookie(dm_tree_get_cookie(root)); out_no_root: dm_tree_free(dtree); return r; } /* origin_only may only be set if we are resuming (not activating) an origin LV */ int dev_manager_activate(struct dev_manager *dm, const struct logical_volume *lv, struct lv_activate_opts *laopts) { if (!_tree_action(dm, lv, laopts, ACTIVATE)) return_0; if (!_tree_action(dm, lv, laopts, CLEAN)) return_0; return 1; } /* origin_only may only be set if we are resuming (not activating) an origin LV */ int dev_manager_preload(struct dev_manager *dm, const struct logical_volume *lv, struct lv_activate_opts *laopts, int *flush_required) { dm->flush_required = *flush_required; if (!_tree_action(dm, lv, laopts, PRELOAD)) return_0; *flush_required = dm->flush_required; return 1; } int dev_manager_deactivate(struct dev_manager *dm, const struct logical_volume *lv) { struct lv_activate_opts laopts = { 0 }; if (!_tree_action(dm, lv, &laopts, DEACTIVATE)) return_0; return 1; } int dev_manager_suspend(struct dev_manager *dm, const struct logical_volume *lv, struct lv_activate_opts *laopts, int lockfs, int flush_required) { dm->flush_required = flush_required; if (!_tree_action(dm, lv, laopts, lockfs ? SUSPEND_WITH_LOCKFS : SUSPEND)) return_0; return 1; } /* * Does device use VG somewhere in its construction? * Returns 1 if uncertain. */ int dev_manager_device_uses_vg(struct device *dev, struct volume_group *vg) { struct dm_tree *dtree; struct dm_tree_node *root; char dlid[sizeof(UUID_PREFIX) + sizeof(struct id) - 1] __attribute__((aligned(8))); int r = 1; if (!(dtree = dm_tree_create())) { log_error("Failed to create partial dtree."); return r; } dm_tree_set_optional_uuid_suffixes(dtree, &uuid_suffix_list[0]); if (!dm_tree_add_dev(dtree, (uint32_t) MAJOR(dev->dev), (uint32_t) MINOR(dev->dev))) { log_error("Failed to add device %s (%" PRIu32 ":%" PRIu32") to dtree.", dev_name(dev), (uint32_t) MAJOR(dev->dev), (uint32_t) MINOR(dev->dev)); goto out; } memcpy(dlid, UUID_PREFIX, sizeof(UUID_PREFIX) - 1); memcpy(dlid + sizeof(UUID_PREFIX) - 1, &vg->id.uuid[0], sizeof(vg->id)); if (!(root = dm_tree_find_node(dtree, 0, 0))) { log_error("Lost dependency tree root node."); goto out; } if (dm_tree_children_use_uuid(root, dlid, sizeof(UUID_PREFIX) + sizeof(vg->id) - 1)) goto_out; r = 0; out: dm_tree_free(dtree); return r; } /* * crypt offset is usually the LUKS header size but can be larger. * The LUKS header is usually 2MB for LUKS1 and 16MB for LUKS2. * The offset needs to be subtracted from the LV size to get the * size used to resize the crypt device. */ int get_crypt_table_offset(dev_t crypt_devt, uint32_t *offset_bytes) { struct dm_task *dmt = dm_task_create(DM_DEVICE_TABLE); uint64_t start, length; char *target_type = NULL; void *next = NULL; char *params = NULL; char offset_str[32] = { 0 }; int copy_offset = 0; int spaces = 0; int i, i_off = 0; if (!dmt) return_0; if (!dm_task_set_major_minor(dmt, (int)MAJOR(crypt_devt), (int)MINOR(crypt_devt), 0)) { dm_task_destroy(dmt); return_0; } /* Non-blocking status read */ if (!dm_task_no_flush(dmt)) log_warn("WARNING: Can't set no_flush for dm status."); if (!dm_task_run(dmt)) { dm_task_destroy(dmt); return_0; } next = dm_get_next_target(dmt, next, &start, &length, &target_type, ¶ms); if (!target_type || !params || strcmp(target_type, "crypt")) { dm_task_destroy(dmt); return_0; } /* * get offset from params string: * [<#opt_params> ] * is reported in 512 byte sectors. */ for (i = 0; i < strlen(params); i++) { if (params[i] == ' ') { spaces++; if (spaces == 4) copy_offset = 1; if (spaces == 5) break; continue; } if (!copy_offset) continue; offset_str[i_off++] = params[i]; if (i_off == sizeof(offset_str)) { offset_str[0] = '\0'; break; } } dm_task_destroy(dmt); if (!offset_str[0]) return_0; *offset_bytes = ((uint32_t)strtoul(offset_str, NULL, 0) * 512); return 1; }