| # |
| # |
| |
| # Copyright (C) 2006, 2007, 2010, 2011, 2012, 2013, 2014 Google Inc. |
| # All rights reserved. |
| # |
| # Redistribution and use in source and binary forms, with or without |
| # modification, are permitted provided that the following conditions are |
| # met: |
| # |
| # 1. Redistributions of source code must retain the above copyright notice, |
| # this list of conditions and the following disclaimer. |
| # |
| # 2. Redistributions in binary form must reproduce the above copyright |
| # notice, this list of conditions and the following disclaimer in the |
| # documentation and/or other materials provided with the distribution. |
| # |
| # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS |
| # IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
| # TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| # PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR |
| # CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| # EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| # PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| # PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF |
| # LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING |
| # NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS |
| # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| """Cluster related commands""" |
| |
| # pylint: disable=W0401,W0613,W0614,C0103 |
| # W0401: Wildcard import ganeti.cli |
| # W0613: Unused argument, since all functions follow the same API |
| # W0614: Unused import %s from wildcard import (since we need cli) |
| # C0103: Invalid name gnt-cluster |
| |
| from cStringIO import StringIO |
| import os |
| import time |
| import OpenSSL |
| import tempfile |
| import itertools |
| |
| from ganeti.cli import * |
| from ganeti import bootstrap |
| from ganeti import compat |
| from ganeti import constants |
| from ganeti import config |
| from ganeti import errors |
| from ganeti import netutils |
| from ganeti import objects |
| from ganeti import opcodes |
| from ganeti import pathutils |
| from ganeti import qlang |
| from ganeti.rpc.node import RunWithRPC |
| from ganeti import serializer |
| from ganeti import ssconf |
| from ganeti import ssh |
| from ganeti import uidpool |
| from ganeti import utils |
| from ganeti import wconfd |
| from ganeti.client import base |
| |
| |
| ON_OPT = cli_option("--on", default=False, |
| action="store_true", dest="on", |
| help="Recover from an EPO") |
| |
| GROUPS_OPT = cli_option("--groups", default=False, |
| action="store_true", dest="groups", |
| help="Arguments are node groups instead of nodes") |
| |
| FORCE_FAILOVER = cli_option("--yes-do-it", dest="yes_do_it", |
| help="Override interactive check for --no-voting", |
| default=False, action="store_true") |
| |
| IGNORE_OFFLINE_NODES_FAILOVER = cli_option( |
| "--ignore-offline-nodes", dest="ignore_offline_nodes", |
| help="Ignores offline nodes for master failover voting", default=True) |
| |
| FORCE_DISTRIBUTION = cli_option("--yes-do-it", dest="yes_do_it", |
| help="Unconditionally distribute the" |
| " configuration, even if the queue" |
| " is drained", |
| default=False, action="store_true") |
| |
| TO_OPT = cli_option("--to", default=None, type="string", |
| help="The Ganeti version to upgrade to") |
| |
| RESUME_OPT = cli_option("--resume", default=False, action="store_true", |
| help="Resume any pending Ganeti upgrades") |
| |
| DATA_COLLECTOR_INTERVAL_OPT = cli_option( |
| "--data-collector-interval", default={}, type="keyval", |
| help="Set collection intervals in seconds of data collectors.") |
| |
| STRICT_OPT = cli_option("--no-strict", default=False, |
| dest="no_strict", action="store_true", |
| help="Do not run group verify in strict mode") |
| |
| _EPO_PING_INTERVAL = 30 # 30 seconds between pings |
| _EPO_PING_TIMEOUT = 1 # 1 second |
| _EPO_REACHABLE_TIMEOUT = 15 * 60 # 15 minutes |
| |
| |
| def _InitEnabledDiskTemplates(opts): |
| """Initialize the list of enabled disk templates. |
| |
| """ |
| if opts.enabled_disk_templates: |
| return opts.enabled_disk_templates.split(",") |
| else: |
| return constants.DEFAULT_ENABLED_DISK_TEMPLATES |
| |
| |
| def _InitVgName(opts, enabled_disk_templates): |
| """Initialize the volume group name. |
| |
| @type enabled_disk_templates: list of strings |
| @param enabled_disk_templates: cluster-wide enabled disk templates |
| |
| """ |
| vg_name = None |
| if opts.vg_name is not None: |
| vg_name = opts.vg_name |
| if vg_name: |
| if not utils.IsLvmEnabled(enabled_disk_templates): |
| ToStdout("You specified a volume group with --vg-name, but you did not" |
| " enable any disk template that uses lvm.") |
| elif utils.IsLvmEnabled(enabled_disk_templates): |
| raise errors.OpPrereqError( |
| "LVM disk templates are enabled, but vg name not set.") |
| elif utils.IsLvmEnabled(enabled_disk_templates): |
| vg_name = constants.DEFAULT_VG |
| return vg_name |
| |
| |
| def _InitDrbdHelper(opts, enabled_disk_templates, feedback_fn=ToStdout): |
| """Initialize the DRBD usermode helper. |
| |
| """ |
| drbd_enabled = constants.DT_DRBD8 in enabled_disk_templates |
| |
| if not drbd_enabled and opts.drbd_helper is not None: |
| feedback_fn("Note: You specified a DRBD usermode helper, while DRBD storage" |
| " is not enabled.") |
| |
| if drbd_enabled: |
| if opts.drbd_helper is None: |
| return constants.DEFAULT_DRBD_HELPER |
| if opts.drbd_helper == '': |
| raise errors.OpPrereqError( |
| "Unsetting the drbd usermode helper while enabling DRBD is not" |
| " allowed.") |
| |
| return opts.drbd_helper |
| |
| |
| @RunWithRPC |
| def InitCluster(opts, args): |
| """Initialize the cluster. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain only one element, the desired |
| cluster name |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| enabled_disk_templates = _InitEnabledDiskTemplates(opts) |
| |
| try: |
| vg_name = _InitVgName(opts, enabled_disk_templates) |
| drbd_helper = _InitDrbdHelper(opts, enabled_disk_templates) |
| except errors.OpPrereqError, e: |
| ToStderr(str(e)) |
| return 1 |
| |
| master_netdev = opts.master_netdev |
| if master_netdev is None: |
| nic_mode = opts.nicparams.get(constants.NIC_MODE, None) |
| if not nic_mode: |
| # default case, use bridging |
| master_netdev = constants.DEFAULT_BRIDGE |
| elif nic_mode == constants.NIC_MODE_OVS: |
| # default ovs is different from default bridge |
| master_netdev = constants.DEFAULT_OVS |
| opts.nicparams[constants.NIC_LINK] = constants.DEFAULT_OVS |
| |
| hvlist = opts.enabled_hypervisors |
| if hvlist is None: |
| hvlist = constants.DEFAULT_ENABLED_HYPERVISOR |
| hvlist = hvlist.split(",") |
| |
| hvparams = dict(opts.hvparams) |
| beparams = opts.beparams |
| nicparams = opts.nicparams |
| |
| diskparams = dict(opts.diskparams) |
| |
| # check the disk template types here, as we cannot rely on the type check done |
| # by the opcode parameter types |
| diskparams_keys = set(diskparams.keys()) |
| if not (diskparams_keys <= constants.DISK_TEMPLATES): |
| unknown = utils.NiceSort(diskparams_keys - constants.DISK_TEMPLATES) |
| ToStderr("Disk templates unknown: %s" % utils.CommaJoin(unknown)) |
| return 1 |
| |
| # prepare beparams dict |
| beparams = objects.FillDict(constants.BEC_DEFAULTS, beparams) |
| utils.ForceDictType(beparams, constants.BES_PARAMETER_COMPAT) |
| |
| # prepare nicparams dict |
| nicparams = objects.FillDict(constants.NICC_DEFAULTS, nicparams) |
| utils.ForceDictType(nicparams, constants.NICS_PARAMETER_TYPES) |
| |
| # prepare ndparams dict |
| if opts.ndparams is None: |
| ndparams = dict(constants.NDC_DEFAULTS) |
| else: |
| ndparams = objects.FillDict(constants.NDC_DEFAULTS, opts.ndparams) |
| utils.ForceDictType(ndparams, constants.NDS_PARAMETER_TYPES) |
| |
| # prepare hvparams dict |
| for hv in constants.HYPER_TYPES: |
| if hv not in hvparams: |
| hvparams[hv] = {} |
| hvparams[hv] = objects.FillDict(constants.HVC_DEFAULTS[hv], hvparams[hv]) |
| utils.ForceDictType(hvparams[hv], constants.HVS_PARAMETER_TYPES) |
| |
| # prepare diskparams dict |
| for templ in constants.DISK_TEMPLATES: |
| if templ not in diskparams: |
| diskparams[templ] = {} |
| diskparams[templ] = objects.FillDict(constants.DISK_DT_DEFAULTS[templ], |
| diskparams[templ]) |
| utils.ForceDictType(diskparams[templ], constants.DISK_DT_TYPES) |
| |
| # prepare ipolicy dict |
| ipolicy = CreateIPolicyFromOpts( |
| ispecs_mem_size=opts.ispecs_mem_size, |
| ispecs_cpu_count=opts.ispecs_cpu_count, |
| ispecs_disk_count=opts.ispecs_disk_count, |
| ispecs_disk_size=opts.ispecs_disk_size, |
| ispecs_nic_count=opts.ispecs_nic_count, |
| minmax_ispecs=opts.ipolicy_bounds_specs, |
| std_ispecs=opts.ipolicy_std_specs, |
| ipolicy_disk_templates=opts.ipolicy_disk_templates, |
| ipolicy_vcpu_ratio=opts.ipolicy_vcpu_ratio, |
| ipolicy_spindle_ratio=opts.ipolicy_spindle_ratio, |
| ipolicy_memory_ratio=opts.ipolicy_memory_ratio, |
| fill_all=True) |
| |
| if opts.candidate_pool_size is None: |
| opts.candidate_pool_size = constants.MASTER_POOL_SIZE_DEFAULT |
| |
| if opts.mac_prefix is None: |
| opts.mac_prefix = constants.DEFAULT_MAC_PREFIX |
| |
| uid_pool = opts.uid_pool |
| if uid_pool is not None: |
| uid_pool = uidpool.ParseUidPool(uid_pool) |
| |
| if opts.prealloc_wipe_disks is None: |
| opts.prealloc_wipe_disks = False |
| |
| external_ip_setup_script = opts.use_external_mip_script |
| if external_ip_setup_script is None: |
| external_ip_setup_script = False |
| |
| try: |
| primary_ip_version = int(opts.primary_ip_version) |
| except (ValueError, TypeError), err: |
| ToStderr("Invalid primary ip version value: %s" % str(err)) |
| return 1 |
| |
| master_netmask = opts.master_netmask |
| try: |
| if master_netmask is not None: |
| master_netmask = int(master_netmask) |
| except (ValueError, TypeError), err: |
| ToStderr("Invalid master netmask value: %s" % str(err)) |
| return 1 |
| |
| if opts.disk_state: |
| disk_state = utils.FlatToDict(opts.disk_state) |
| else: |
| disk_state = {} |
| |
| hv_state = dict(opts.hv_state) |
| |
| if opts.install_image: |
| install_image = opts.install_image |
| else: |
| install_image = "" |
| |
| if opts.zeroing_image: |
| zeroing_image = opts.zeroing_image |
| else: |
| zeroing_image = "" |
| |
| compression_tools = _GetCompressionTools(opts) |
| |
| default_ialloc_params = opts.default_iallocator_params |
| |
| if opts.enabled_user_shutdown: |
| enabled_user_shutdown = True |
| else: |
| enabled_user_shutdown = False |
| |
| if opts.ssh_key_type: |
| ssh_key_type = opts.ssh_key_type |
| else: |
| ssh_key_type = constants.SSH_DEFAULT_KEY_TYPE |
| |
| ssh_key_bits = ssh.DetermineKeyBits(ssh_key_type, opts.ssh_key_bits, None, |
| None) |
| |
| bootstrap.InitCluster(cluster_name=args[0], |
| secondary_ip=opts.secondary_ip, |
| vg_name=vg_name, |
| mac_prefix=opts.mac_prefix, |
| master_netmask=master_netmask, |
| master_netdev=master_netdev, |
| file_storage_dir=opts.file_storage_dir, |
| shared_file_storage_dir=opts.shared_file_storage_dir, |
| gluster_storage_dir=opts.gluster_storage_dir, |
| enabled_hypervisors=hvlist, |
| hvparams=hvparams, |
| beparams=beparams, |
| nicparams=nicparams, |
| ndparams=ndparams, |
| diskparams=diskparams, |
| ipolicy=ipolicy, |
| candidate_pool_size=opts.candidate_pool_size, |
| modify_etc_hosts=opts.modify_etc_hosts, |
| modify_ssh_setup=opts.modify_ssh_setup, |
| maintain_node_health=opts.maintain_node_health, |
| drbd_helper=drbd_helper, |
| uid_pool=uid_pool, |
| default_iallocator=opts.default_iallocator, |
| default_iallocator_params=default_ialloc_params, |
| primary_ip_version=primary_ip_version, |
| prealloc_wipe_disks=opts.prealloc_wipe_disks, |
| use_external_mip_script=external_ip_setup_script, |
| hv_state=hv_state, |
| disk_state=disk_state, |
| enabled_disk_templates=enabled_disk_templates, |
| install_image=install_image, |
| zeroing_image=zeroing_image, |
| compression_tools=compression_tools, |
| enabled_user_shutdown=enabled_user_shutdown, |
| ssh_key_type=ssh_key_type, |
| ssh_key_bits=ssh_key_bits, |
| ) |
| op = opcodes.OpClusterPostInit() |
| SubmitOpCode(op, opts=opts) |
| return 0 |
| |
| |
| @RunWithRPC |
| def DestroyCluster(opts, args): |
| """Destroy the cluster. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| if not opts.yes_do_it: |
| ToStderr("Destroying a cluster is irreversible. If you really want" |
| " destroy this cluster, supply the --yes-do-it option.") |
| return 1 |
| |
| op = opcodes.OpClusterDestroy() |
| master_uuid = SubmitOpCode(op, opts=opts) |
| # if we reached this, the opcode didn't fail; we can proceed to |
| # shutdown all the daemons |
| bootstrap.FinalizeClusterDestroy(master_uuid) |
| return 0 |
| |
| |
| def RenameCluster(opts, args): |
| """Rename the cluster. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain only one element, the new cluster name |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| cl = GetClient() |
| |
| (cluster_name, ) = cl.QueryConfigValues(["cluster_name"]) |
| |
| new_name = args[0] |
| if not opts.force: |
| usertext = ("This will rename the cluster from '%s' to '%s'. If you are" |
| " connected over the network to the cluster name, the" |
| " operation is very dangerous as the IP address will be" |
| " removed from the node and the change may not go through." |
| " Continue?") % (cluster_name, new_name) |
| if not AskUser(usertext): |
| return 1 |
| |
| op = opcodes.OpClusterRename(name=new_name) |
| result = SubmitOpCode(op, opts=opts, cl=cl) |
| |
| if result: |
| ToStdout("Cluster renamed from '%s' to '%s'", cluster_name, result) |
| |
| return 0 |
| |
| |
| def ActivateMasterIp(opts, args): |
| """Activates the master IP. |
| |
| """ |
| op = opcodes.OpClusterActivateMasterIp() |
| SubmitOpCode(op) |
| return 0 |
| |
| |
| def DeactivateMasterIp(opts, args): |
| """Deactivates the master IP. |
| |
| """ |
| if not opts.confirm: |
| usertext = ("This will disable the master IP. All the open connections to" |
| " the master IP will be closed. To reach the master you will" |
| " need to use its node IP." |
| " Continue?") |
| if not AskUser(usertext): |
| return 1 |
| |
| op = opcodes.OpClusterDeactivateMasterIp() |
| SubmitOpCode(op) |
| return 0 |
| |
| |
| def RedistributeConfig(opts, args): |
| """Forces push of the cluster configuration. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| op = opcodes.OpClusterRedistConf() |
| if opts.yes_do_it: |
| SubmitOpCodeToDrainedQueue(op) |
| else: |
| SubmitOrSend(op, opts) |
| return 0 |
| |
| |
| def ShowClusterVersion(opts, args): |
| """Write version of ganeti software to the standard output. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| cl = GetClient() |
| result = cl.QueryClusterInfo() |
| ToStdout("Software version: %s", result["software_version"]) |
| ToStdout("Internode protocol: %s", result["protocol_version"]) |
| ToStdout("Configuration format: %s", result["config_version"]) |
| ToStdout("OS api version: %s", result["os_api_version"]) |
| ToStdout("Export interface: %s", result["export_version"]) |
| ToStdout("VCS version: %s", result["vcs_version"]) |
| return 0 |
| |
| |
| def ShowClusterMaster(opts, args): |
| """Write name of master node to the standard output. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| master = bootstrap.GetMaster() |
| ToStdout(master) |
| return 0 |
| |
| |
| def _FormatGroupedParams(paramsdict, roman=False): |
| """Format Grouped parameters (be, nic, disk) by group. |
| |
| @type paramsdict: dict of dicts |
| @param paramsdict: {group: {param: value, ...}, ...} |
| @rtype: dict of dicts |
| @return: copy of the input dictionaries with strings as values |
| |
| """ |
| ret = {} |
| for (item, val) in paramsdict.items(): |
| if isinstance(val, dict): |
| ret[item] = _FormatGroupedParams(val, roman=roman) |
| elif roman and isinstance(val, int): |
| ret[item] = compat.TryToRoman(val) |
| else: |
| ret[item] = str(val) |
| return ret |
| |
| |
| def _FormatDataCollectors(paramsdict): |
| """Format Grouped parameters (be, nic, disk) by group. |
| |
| @type paramsdict: dict of dicts |
| @param paramsdict: response of QueryClusterInfo |
| @rtype: dict of dicts |
| @return: parameter grouped by data collector |
| |
| """ |
| |
| enabled = paramsdict[constants.DATA_COLLECTORS_ENABLED_NAME] |
| interval = paramsdict[constants.DATA_COLLECTORS_INTERVAL_NAME] |
| |
| ret = {} |
| for key in enabled: |
| ret[key] = dict(active=enabled[key], |
| interval="%.3fs" % (interval[key] / 1e6)) |
| return ret |
| |
| |
| def ShowClusterConfig(opts, args): |
| """Shows cluster information. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| cl = GetClient() |
| result = cl.QueryClusterInfo() |
| |
| if result["tags"]: |
| tags = utils.CommaJoin(utils.NiceSort(result["tags"])) |
| else: |
| tags = "(none)" |
| if result["reserved_lvs"]: |
| reserved_lvs = utils.CommaJoin(result["reserved_lvs"]) |
| else: |
| reserved_lvs = "(none)" |
| |
| enabled_hv = result["enabled_hypervisors"] |
| hvparams = dict((k, v) for k, v in result["hvparams"].iteritems() |
| if k in enabled_hv) |
| |
| info = [ |
| ("Cluster name", result["name"]), |
| ("Cluster UUID", result["uuid"]), |
| |
| ("Creation time", utils.FormatTime(result["ctime"])), |
| ("Modification time", utils.FormatTime(result["mtime"])), |
| |
| ("Master node", result["master"]), |
| |
| ("Architecture (this node)", |
| "%s (%s)" % (result["architecture"][0], result["architecture"][1])), |
| |
| ("Tags", tags), |
| |
| ("Default hypervisor", result["default_hypervisor"]), |
| ("Enabled hypervisors", utils.CommaJoin(enabled_hv)), |
| |
| ("Hypervisor parameters", _FormatGroupedParams(hvparams, |
| opts.roman_integers)), |
| |
| ("OS-specific hypervisor parameters", |
| _FormatGroupedParams(result["os_hvp"], opts.roman_integers)), |
| |
| ("OS parameters", _FormatGroupedParams(result["osparams"], |
| opts.roman_integers)), |
| |
| ("Hidden OSes", utils.CommaJoin(result["hidden_os"])), |
| ("Blacklisted OSes", utils.CommaJoin(result["blacklisted_os"])), |
| |
| ("Cluster parameters", [ |
| ("candidate pool size", |
| compat.TryToRoman(result["candidate_pool_size"], |
| convert=opts.roman_integers)), |
| ("maximal number of jobs running simultaneously", |
| compat.TryToRoman(result["max_running_jobs"], |
| convert=opts.roman_integers)), |
| ("maximal number of jobs simultaneously tracked by the scheduler", |
| compat.TryToRoman(result["max_tracked_jobs"], |
| convert=opts.roman_integers)), |
| ("mac prefix", result["mac_prefix"]), |
| ("master netdev", result["master_netdev"]), |
| ("master netmask", compat.TryToRoman(result["master_netmask"], |
| opts.roman_integers)), |
| ("use external master IP address setup script", |
| result["use_external_mip_script"]), |
| ("lvm volume group", result["volume_group_name"]), |
| ("lvm reserved volumes", reserved_lvs), |
| ("drbd usermode helper", result["drbd_usermode_helper"]), |
| ("file storage path", result["file_storage_dir"]), |
| ("shared file storage path", result["shared_file_storage_dir"]), |
| ("gluster storage path", result["gluster_storage_dir"]), |
| ("maintenance of node health", result["maintain_node_health"]), |
| ("uid pool", uidpool.FormatUidPool(result["uid_pool"])), |
| ("default instance allocator", result["default_iallocator"]), |
| ("default instance allocator parameters", |
| result["default_iallocator_params"]), |
| ("primary ip version", compat.TryToRoman(result["primary_ip_version"], |
| opts.roman_integers)), |
| ("preallocation wipe disks", result["prealloc_wipe_disks"]), |
| ("OS search path", utils.CommaJoin(pathutils.OS_SEARCH_PATH)), |
| ("ExtStorage Providers search path", |
| utils.CommaJoin(pathutils.ES_SEARCH_PATH)), |
| ("enabled disk templates", |
| utils.CommaJoin(result["enabled_disk_templates"])), |
| ("install image", result["install_image"]), |
| ("instance communication network", |
| result["instance_communication_network"]), |
| ("zeroing image", result["zeroing_image"]), |
| ("compression tools", result["compression_tools"]), |
| ("enabled user shutdown", result["enabled_user_shutdown"]), |
| ("modify ssh setup", result["modify_ssh_setup"]), |
| ("ssh_key_type", result["ssh_key_type"]), |
| ("ssh_key_bits", result["ssh_key_bits"]), |
| ]), |
| |
| ("Default node parameters", |
| _FormatGroupedParams(result["ndparams"], roman=opts.roman_integers)), |
| |
| ("Default instance parameters", |
| _FormatGroupedParams(result["beparams"], roman=opts.roman_integers)), |
| |
| ("Default nic parameters", |
| _FormatGroupedParams(result["nicparams"], roman=opts.roman_integers)), |
| |
| ("Default disk parameters", |
| _FormatGroupedParams(result["diskparams"], roman=opts.roman_integers)), |
| |
| ("Instance policy - limits for instances", |
| FormatPolicyInfo(result["ipolicy"], None, True, opts.roman_integers)), |
| ("Data collectors", _FormatDataCollectors(result)), |
| ] |
| |
| PrintGenericInfo(info) |
| return 0 |
| |
| |
| def ClusterCopyFile(opts, args): |
| """Copy a file from master to some nodes. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain only one element, the path of |
| the file to be copied |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| filename = args[0] |
| filename = os.path.abspath(filename) |
| |
| if not os.path.exists(filename): |
| raise errors.OpPrereqError("No such filename '%s'" % filename, |
| errors.ECODE_INVAL) |
| |
| cl = GetClient() |
| qcl = GetClient() |
| try: |
| cluster_name = cl.QueryConfigValues(["cluster_name"])[0] |
| |
| results = GetOnlineNodes(nodes=opts.nodes, cl=qcl, filter_master=True, |
| secondary_ips=opts.use_replication_network, |
| nodegroup=opts.nodegroup) |
| ports = GetNodesSshPorts(opts.nodes, qcl) |
| finally: |
| cl.Close() |
| qcl.Close() |
| |
| srun = ssh.SshRunner(cluster_name) |
| for (node, port) in zip(results, ports): |
| if not srun.CopyFileToNode(node, port, filename): |
| ToStderr("Copy of file %s to node %s:%d failed", filename, node, port) |
| |
| return 0 |
| |
| |
| def RunClusterCommand(opts, args): |
| """Run a command on some nodes. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain the command to be run and its arguments |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| cl = GetClient() |
| qcl = GetClient() |
| |
| command = " ".join(args) |
| |
| nodes = GetOnlineNodes(nodes=opts.nodes, cl=qcl, nodegroup=opts.nodegroup) |
| ports = GetNodesSshPorts(nodes, qcl) |
| |
| cluster_name, master_node = cl.QueryConfigValues(["cluster_name", |
| "master_node"]) |
| |
| srun = ssh.SshRunner(cluster_name=cluster_name) |
| |
| # Make sure master node is at list end |
| if master_node in nodes: |
| nodes.remove(master_node) |
| nodes.append(master_node) |
| |
| for (name, port) in zip(nodes, ports): |
| result = srun.Run(name, constants.SSH_LOGIN_USER, command, port=port) |
| |
| if opts.failure_only and result.exit_code == constants.EXIT_SUCCESS: |
| # Do not output anything for successful commands |
| continue |
| |
| ToStdout("------------------------------------------------") |
| if opts.show_machine_names: |
| for line in result.output.splitlines(): |
| ToStdout("%s: %s", name, line) |
| else: |
| ToStdout("node: %s", name) |
| ToStdout("%s", result.output) |
| ToStdout("return code = %s", result.exit_code) |
| |
| return 0 |
| |
| |
| def VerifyCluster(opts, args): |
| """Verify integrity of cluster, performing various test on nodes. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| skip_checks = [] |
| |
| if opts.skip_nplusone_mem: |
| skip_checks.append(constants.VERIFY_NPLUSONE_MEM) |
| |
| cl = GetClient() |
| |
| op = opcodes.OpClusterVerify(verbose=opts.verbose, |
| error_codes=opts.error_codes, |
| debug_simulate_errors=opts.simulate_errors, |
| skip_checks=skip_checks, |
| ignore_errors=opts.ignore_errors, |
| group_name=opts.nodegroup, |
| verify_clutter=opts.verify_clutter) |
| result = SubmitOpCode(op, cl=cl, opts=opts) |
| |
| # Keep track of submitted jobs |
| jex = JobExecutor(cl=cl, opts=opts) |
| |
| for (status, job_id) in result[constants.JOB_IDS_KEY]: |
| jex.AddJobId(None, status, job_id) |
| |
| results = jex.GetResults() |
| |
| bad_jobs = sum(1 for (job_success, _) in results if not job_success) |
| bad_results = sum(1 for (_, op_res) in results if not (op_res and op_res[0])) |
| |
| if bad_jobs == 0 and bad_results == 0: |
| rcode = constants.EXIT_SUCCESS |
| else: |
| rcode = constants.EXIT_FAILURE |
| if bad_jobs > 0: |
| ToStdout("%s job(s) failed while verifying the cluster.", bad_jobs) |
| |
| return rcode |
| |
| |
| def VerifyDisks(opts, args): |
| """Verify integrity of cluster disks. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| cl = GetClient() |
| |
| op = opcodes.OpClusterVerifyDisks(group_name=opts.nodegroup, |
| is_strict=not opts.no_strict) |
| |
| result = SubmitOpCode(op, cl=cl, opts=opts) |
| |
| # Keep track of submitted jobs |
| jex = JobExecutor(cl=cl, opts=opts) |
| |
| for (status, job_id) in result[constants.JOB_IDS_KEY]: |
| jex.AddJobId(None, status, job_id) |
| |
| retcode = constants.EXIT_SUCCESS |
| |
| for (status, result) in jex.GetResults(): |
| if not status: |
| ToStdout("Job failed: %s", result) |
| continue |
| |
| ((bad_nodes, instances, missing), ) = result |
| |
| for node, text in bad_nodes.items(): |
| ToStdout("Error gathering data on node %s: %s", |
| node, utils.SafeEncode(text[-400:])) |
| retcode = constants.EXIT_FAILURE |
| ToStdout("You need to fix these nodes first before fixing instances") |
| |
| for iname in instances: |
| if iname in missing: |
| continue |
| op = opcodes.OpInstanceActivateDisks(instance_name=iname) |
| try: |
| ToStdout("Activating disks for instance '%s'", iname) |
| SubmitOpCode(op, opts=opts, cl=cl) |
| except errors.GenericError, err: |
| nret, msg = FormatError(err) |
| retcode |= nret |
| ToStderr("Error activating disks for instance %s: %s", iname, msg) |
| |
| if missing: |
| for iname, ival in missing.iteritems(): |
| all_missing = compat.all(x[0] in bad_nodes for x in ival) |
| if all_missing: |
| ToStdout("Instance %s cannot be verified as it lives on" |
| " broken nodes", iname) |
| else: |
| ToStdout("Instance %s has missing logical volumes:", iname) |
| ival.sort() |
| for node, vol in ival: |
| if node in bad_nodes: |
| ToStdout("\tbroken node %s /dev/%s", node, vol) |
| else: |
| ToStdout("\t%s /dev/%s", node, vol) |
| |
| ToStdout("You need to replace or recreate disks for all the above" |
| " instances if this message persists after fixing broken nodes.") |
| retcode = constants.EXIT_FAILURE |
| elif not instances: |
| ToStdout("No disks need to be activated.") |
| |
| return retcode |
| |
| |
| def RepairDiskSizes(opts, args): |
| """Verify sizes of cluster disks. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: optional list of instances to restrict check to |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| op = opcodes.OpClusterRepairDiskSizes(instances=args) |
| SubmitOpCode(op, opts=opts) |
| |
| |
| @RunWithRPC |
| def MasterFailover(opts, args): |
| """Failover the master node. |
| |
| This command, when run on a non-master node, will cause the current |
| master to cease being master, and the non-master to become new |
| master. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| if opts.no_voting: |
| # Don't ask for confirmation if the user provides the confirmation flag. |
| if not opts.yes_do_it: |
| usertext = ("This will perform the failover even if most other nodes" |
| " are down, or if this node is outdated. This is dangerous" |
| " as it can lead to a non-consistent cluster. Check the" |
| " gnt-cluster(8) man page before proceeding. Continue?") |
| if not AskUser(usertext): |
| return 1 |
| else: |
| # Verify that a majority of nodes are still healthy |
| (majority_healthy, unhealthy_nodes) = bootstrap.MajorityHealthy( |
| opts.ignore_offline_nodes) |
| if not majority_healthy: |
| ToStderr("Master-failover with voting is only possible if the majority" |
| " of nodes are still healthy; use the --no-voting option after" |
| " ensuring by other means that you won't end up in a dual-master" |
| " scenario. Unhealthy nodes: %s" % unhealthy_nodes) |
| return 1 |
| |
| rvalue, msgs = bootstrap.MasterFailover(no_voting=opts.no_voting) |
| for msg in msgs: |
| ToStderr(msg) |
| |
| return rvalue |
| |
| |
| def MasterPing(opts, args): |
| """Checks if the master is alive. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| try: |
| cl = GetClient() |
| cl.QueryClusterInfo() |
| return 0 |
| except Exception: # pylint: disable=W0703 |
| return 1 |
| |
| |
| def SearchTags(opts, args): |
| """Searches the tags on all the cluster. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain only one element, the tag pattern |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| op = opcodes.OpTagsSearch(pattern=args[0]) |
| result = SubmitOpCode(op, opts=opts) |
| if not result: |
| return 1 |
| result = list(result) |
| result.sort() |
| for path, tag in result: |
| ToStdout("%s %s", path, tag) |
| |
| |
| def _ReadAndVerifyCert(cert_filename, verify_private_key=False): |
| """Reads and verifies an X509 certificate. |
| |
| @type cert_filename: string |
| @param cert_filename: the path of the file containing the certificate to |
| verify encoded in PEM format |
| @type verify_private_key: bool |
| @param verify_private_key: whether to verify the private key in addition to |
| the public certificate |
| @rtype: string |
| @return: a string containing the PEM-encoded certificate. |
| |
| """ |
| try: |
| pem = utils.ReadFile(cert_filename) |
| except IOError, err: |
| raise errors.X509CertError(cert_filename, |
| "Unable to read certificate: %s" % str(err)) |
| |
| try: |
| OpenSSL.crypto.load_certificate(OpenSSL.crypto.FILETYPE_PEM, pem) |
| except Exception, err: |
| raise errors.X509CertError(cert_filename, |
| "Unable to load certificate: %s" % str(err)) |
| |
| if verify_private_key: |
| try: |
| OpenSSL.crypto.load_privatekey(OpenSSL.crypto.FILETYPE_PEM, pem) |
| except Exception, err: |
| raise errors.X509CertError(cert_filename, |
| "Unable to load private key: %s" % str(err)) |
| |
| return pem |
| |
| |
| # pylint: disable=R0913 |
| def _RenewCrypto(new_cluster_cert, new_rapi_cert, # pylint: disable=R0911 |
| rapi_cert_filename, new_spice_cert, spice_cert_filename, |
| spice_cacert_filename, new_confd_hmac_key, new_cds, |
| cds_filename, force, new_node_cert, new_ssh_keys, |
| ssh_key_type, ssh_key_bits, verbose, debug): |
| """Renews cluster certificates, keys and secrets. |
| |
| @type new_cluster_cert: bool |
| @param new_cluster_cert: Whether to generate a new cluster certificate |
| @type new_rapi_cert: bool |
| @param new_rapi_cert: Whether to generate a new RAPI certificate |
| @type rapi_cert_filename: string |
| @param rapi_cert_filename: Path to file containing new RAPI certificate |
| @type new_spice_cert: bool |
| @param new_spice_cert: Whether to generate a new SPICE certificate |
| @type spice_cert_filename: string |
| @param spice_cert_filename: Path to file containing new SPICE certificate |
| @type spice_cacert_filename: string |
| @param spice_cacert_filename: Path to file containing the certificate of the |
| CA that signed the SPICE certificate |
| @type new_confd_hmac_key: bool |
| @param new_confd_hmac_key: Whether to generate a new HMAC key |
| @type new_cds: bool |
| @param new_cds: Whether to generate a new cluster domain secret |
| @type cds_filename: string |
| @param cds_filename: Path to file containing new cluster domain secret |
| @type force: bool |
| @param force: Whether to ask user for confirmation |
| @type new_node_cert: bool |
| @param new_node_cert: Whether to generate new node certificates |
| @type new_ssh_keys: bool |
| @param new_ssh_keys: Whether to generate new node SSH keys |
| @type ssh_key_type: One of L{constants.SSHK_ALL} |
| @param ssh_key_type: The type of SSH key to be generated |
| @type ssh_key_bits: int |
| @param ssh_key_bits: The length of the key to be generated |
| @type verbose: boolean |
| @param verbose: Show verbose output |
| @type debug: boolean |
| @param debug: Show debug output |
| |
| """ |
| ToStdout("Updating certificates now. Running \"gnt-cluster verify\" " |
| " is recommended after this operation.") |
| |
| if new_rapi_cert and rapi_cert_filename: |
| ToStderr("Only one of the --new-rapi-certificate and --rapi-certificate" |
| " options can be specified at the same time.") |
| return 1 |
| |
| if new_cds and cds_filename: |
| ToStderr("Only one of the --new-cluster-domain-secret and" |
| " --cluster-domain-secret options can be specified at" |
| " the same time.") |
| return 1 |
| |
| if new_spice_cert and (spice_cert_filename or spice_cacert_filename): |
| ToStderr("When using --new-spice-certificate, the --spice-certificate" |
| " and --spice-ca-certificate must not be used.") |
| return 1 |
| |
| if bool(spice_cacert_filename) ^ bool(spice_cert_filename): |
| ToStderr("Both --spice-certificate and --spice-ca-certificate must be" |
| " specified.") |
| return 1 |
| |
| rapi_cert_pem, spice_cert_pem, spice_cacert_pem = (None, None, None) |
| try: |
| if rapi_cert_filename: |
| rapi_cert_pem = _ReadAndVerifyCert(rapi_cert_filename, True) |
| if spice_cert_filename: |
| spice_cert_pem = _ReadAndVerifyCert(spice_cert_filename, True) |
| spice_cacert_pem = _ReadAndVerifyCert(spice_cacert_filename) |
| except errors.X509CertError, err: |
| ToStderr("Unable to load X509 certificate from %s: %s", err[0], err[1]) |
| return 1 |
| |
| if cds_filename: |
| try: |
| cds = utils.ReadFile(cds_filename) |
| except Exception, err: # pylint: disable=W0703 |
| ToStderr("Can't load new cluster domain secret from %s: %s" % |
| (cds_filename, str(err))) |
| return 1 |
| else: |
| cds = None |
| |
| if not force: |
| usertext = ("This requires all daemons on all nodes to be restarted and" |
| " may take some time. Continue?") |
| if not AskUser(usertext): |
| return 1 |
| |
| def _RenewCryptoInner(ctx): |
| ctx.feedback_fn("Updating certificates and keys") |
| |
| bootstrap.GenerateClusterCrypto(False, |
| new_rapi_cert, |
| new_spice_cert, |
| new_confd_hmac_key, |
| new_cds, |
| False, |
| None, |
| rapi_cert_pem=rapi_cert_pem, |
| spice_cert_pem=spice_cert_pem, |
| spice_cacert_pem=spice_cacert_pem, |
| cds=cds) |
| |
| files_to_copy = [] |
| |
| if new_rapi_cert or rapi_cert_pem: |
| files_to_copy.append(pathutils.RAPI_CERT_FILE) |
| |
| if new_spice_cert or spice_cert_pem: |
| files_to_copy.append(pathutils.SPICE_CERT_FILE) |
| files_to_copy.append(pathutils.SPICE_CACERT_FILE) |
| |
| if new_confd_hmac_key: |
| files_to_copy.append(pathutils.CONFD_HMAC_KEY) |
| |
| if new_cds or cds: |
| files_to_copy.append(pathutils.CLUSTER_DOMAIN_SECRET_FILE) |
| |
| if files_to_copy: |
| for node_name in ctx.nonmaster_nodes: |
| port = ctx.ssh_ports[node_name] |
| ctx.feedback_fn("Copying %s to %s:%d" % |
| (", ".join(files_to_copy), node_name, port)) |
| for file_name in files_to_copy: |
| ctx.ssh.CopyFileToNode(node_name, port, file_name) |
| |
| def _RenewClientCerts(ctx): |
| ctx.feedback_fn("Updating client SSL certificates.") |
| |
| cluster_name = ssconf.SimpleStore().GetClusterName() |
| |
| for node_name in ctx.nonmaster_nodes + [ctx.master_node]: |
| ssh_port = ctx.ssh_ports[node_name] |
| data = { |
| constants.NDS_CLUSTER_NAME: cluster_name, |
| constants.NDS_NODE_DAEMON_CERTIFICATE: |
| utils.ReadFile(pathutils.NODED_CERT_FILE), |
| constants.NDS_NODE_NAME: node_name, |
| constants.NDS_ACTION: constants.CRYPTO_ACTION_CREATE, |
| } |
| |
| ssh.RunSshCmdWithStdin( |
| cluster_name, |
| node_name, |
| pathutils.SSL_UPDATE, |
| ssh_port, |
| data, |
| debug=ctx.debug, |
| verbose=ctx.verbose, |
| use_cluster_key=True, |
| ask_key=False, |
| strict_host_check=True) |
| |
| # Create a temporary ssconf file using the master's client cert digest |
| # and the 'bootstrap' keyword to enable distribution of all nodes' digests. |
| master_digest = utils.GetCertificateDigest() |
| ssconf_master_candidate_certs_filename = os.path.join( |
| pathutils.DATA_DIR, "%s%s" % |
| (constants.SSCONF_FILEPREFIX, constants.SS_MASTER_CANDIDATES_CERTS)) |
| utils.WriteFile( |
| ssconf_master_candidate_certs_filename, |
| data="%s=%s" % (constants.CRYPTO_BOOTSTRAP, master_digest)) |
| for node_name in ctx.nonmaster_nodes: |
| port = ctx.ssh_ports[node_name] |
| ctx.feedback_fn("Copying %s to %s:%d" % |
| (ssconf_master_candidate_certs_filename, node_name, port)) |
| ctx.ssh.CopyFileToNode(node_name, port, |
| ssconf_master_candidate_certs_filename) |
| |
| # Write the boostrap entry to the config using wconfd. |
| config_live_lock = utils.livelock.LiveLock("renew_crypto") |
| cfg = config.GetConfig(None, config_live_lock) |
| cfg.AddNodeToCandidateCerts(constants.CRYPTO_BOOTSTRAP, master_digest) |
| cfg.Update(cfg.GetClusterInfo(), ctx.feedback_fn) |
| |
| def _RenewServerAndClientCerts(ctx): |
| ctx.feedback_fn("Updating the cluster SSL certificate.") |
| |
| master_name = ssconf.SimpleStore().GetMasterNode() |
| bootstrap.GenerateClusterCrypto(True, # cluster cert |
| False, # rapi cert |
| False, # spice cert |
| False, # confd hmac key |
| False, # cds |
| True, # client cert |
| master_name) |
| |
| for node_name in ctx.nonmaster_nodes: |
| port = ctx.ssh_ports[node_name] |
| server_cert = pathutils.NODED_CERT_FILE |
| ctx.feedback_fn("Copying %s to %s:%d" % |
| (server_cert, node_name, port)) |
| ctx.ssh.CopyFileToNode(node_name, port, server_cert) |
| |
| _RenewClientCerts(ctx) |
| |
| if new_rapi_cert or new_spice_cert or new_confd_hmac_key or new_cds: |
| RunWhileClusterStopped(ToStdout, _RenewCryptoInner) |
| |
| # If only node certficates are recreated, call _RenewClientCerts only. |
| if new_node_cert and not new_cluster_cert: |
| RunWhileDaemonsStopped(ToStdout, [constants.NODED, constants.WCONFD], |
| _RenewClientCerts, verbose=verbose, debug=debug) |
| |
| # If the cluster certificate are renewed, the client certificates need |
| # to be renewed too. |
| if new_cluster_cert: |
| RunWhileDaemonsStopped(ToStdout, [constants.NODED, constants.WCONFD], |
| _RenewServerAndClientCerts, verbose=verbose, |
| debug=debug) |
| |
| if new_node_cert or new_cluster_cert or new_ssh_keys: |
| cl = GetClient() |
| renew_op = opcodes.OpClusterRenewCrypto( |
| node_certificates=new_node_cert or new_cluster_cert, |
| renew_ssh_keys=new_ssh_keys, |
| ssh_key_type=ssh_key_type, |
| ssh_key_bits=ssh_key_bits, |
| verbose=verbose, |
| debug=debug) |
| SubmitOpCode(renew_op, cl=cl) |
| |
| ToStdout("All requested certificates and keys have been replaced." |
| " Running \"gnt-cluster verify\" now is recommended.") |
| |
| return 0 |
| |
| |
| def _BuildGanetiPubKeys(options, pub_key_file=pathutils.SSH_PUB_KEYS, cl=None, |
| get_online_nodes_fn=GetOnlineNodes, |
| get_nodes_ssh_ports_fn=GetNodesSshPorts, |
| get_node_uuids_fn=GetNodeUUIDs, |
| homedir_fn=None): |
| """Recreates the 'ganeti_pub_key' file by polling all nodes. |
| |
| """ |
| |
| if not cl: |
| cl = GetClient() |
| |
| (cluster_name, master_node, modify_ssh_setup, ssh_key_type) = \ |
| cl.QueryConfigValues(["cluster_name", "master_node", "modify_ssh_setup", |
| "ssh_key_type"]) |
| |
| # In case Ganeti is not supposed to modify the SSH setup, simply exit and do |
| # not update this file. |
| if not modify_ssh_setup: |
| return |
| |
| if os.path.exists(pub_key_file): |
| utils.CreateBackup(pub_key_file) |
| utils.RemoveFile(pub_key_file) |
| |
| ssh.ClearPubKeyFile(pub_key_file) |
| |
| online_nodes = get_online_nodes_fn([], cl=cl) |
| ssh_ports = get_nodes_ssh_ports_fn(online_nodes + [master_node], cl) |
| ssh_port_map = dict(zip(online_nodes + [master_node], ssh_ports)) |
| |
| node_uuids = get_node_uuids_fn(online_nodes + [master_node], cl) |
| node_uuid_map = dict(zip(online_nodes + [master_node], node_uuids)) |
| |
| nonmaster_nodes = [name for name in online_nodes |
| if name != master_node] |
| |
| _, pub_key_filename, _ = \ |
| ssh.GetUserFiles(constants.SSH_LOGIN_USER, mkdir=False, dircheck=False, |
| kind=ssh_key_type, _homedir_fn=homedir_fn) |
| |
| # get the key file of the master node |
| pub_key = utils.ReadFile(pub_key_filename) |
| ssh.AddPublicKey(node_uuid_map[master_node], pub_key, |
| key_file=pub_key_file) |
| |
| # get the key files of all non-master nodes |
| for node in nonmaster_nodes: |
| pub_key = ssh.ReadRemoteSshPubKey(pub_key_filename, node, cluster_name, |
| ssh_port_map[node], |
| options.ssh_key_check, |
| options.ssh_key_check) |
| ssh.AddPublicKey(node_uuid_map[node], pub_key, key_file=pub_key_file) |
| |
| |
| def RenewCrypto(opts, args): |
| """Renews cluster certificates, keys and secrets. |
| |
| """ |
| if opts.new_ssh_keys: |
| _BuildGanetiPubKeys(opts) |
| return _RenewCrypto(opts.new_cluster_cert, |
| opts.new_rapi_cert, |
| opts.rapi_cert, |
| opts.new_spice_cert, |
| opts.spice_cert, |
| opts.spice_cacert, |
| opts.new_confd_hmac_key, |
| opts.new_cluster_domain_secret, |
| opts.cluster_domain_secret, |
| opts.force, |
| opts.new_node_cert, |
| opts.new_ssh_keys, |
| opts.ssh_key_type, |
| opts.ssh_key_bits, |
| opts.verbose, |
| opts.debug > 0) |
| |
| |
| def _GetEnabledDiskTemplates(opts): |
| """Determine the list of enabled disk templates. |
| |
| """ |
| if opts.enabled_disk_templates: |
| return opts.enabled_disk_templates.split(",") |
| else: |
| return None |
| |
| |
| def _GetVgName(opts, enabled_disk_templates): |
| """Determine the volume group name. |
| |
| @type enabled_disk_templates: list of strings |
| @param enabled_disk_templates: cluster-wide enabled disk-templates |
| |
| """ |
| # consistency between vg name and enabled disk templates |
| vg_name = None |
| if opts.vg_name is not None: |
| vg_name = opts.vg_name |
| if enabled_disk_templates: |
| if vg_name and not utils.IsLvmEnabled(enabled_disk_templates): |
| ToStdout("You specified a volume group with --vg-name, but you did not" |
| " enable any of the following lvm-based disk templates: %s" % |
| utils.CommaJoin(constants.DTS_LVM)) |
| return vg_name |
| |
| |
| def _GetDrbdHelper(opts, enabled_disk_templates): |
| """Determine the DRBD usermode helper. |
| |
| """ |
| drbd_helper = opts.drbd_helper |
| if enabled_disk_templates: |
| drbd_enabled = constants.DT_DRBD8 in enabled_disk_templates |
| if not drbd_enabled and opts.drbd_helper: |
| ToStdout("You specified a DRBD usermode helper with " |
| " --drbd-usermode-helper while DRBD is not enabled.") |
| return drbd_helper |
| |
| |
| def _GetCompressionTools(opts): |
| """Determine the list of custom compression tools. |
| |
| """ |
| if opts.compression_tools: |
| return opts.compression_tools.split(",") |
| elif opts.compression_tools is None: |
| return None # To note the parameter was not provided |
| else: |
| return constants.IEC_DEFAULT_TOOLS # Resetting to default |
| |
| |
| def SetClusterParams(opts, args): |
| """Modify the cluster. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| if not (opts.vg_name is not None or |
| opts.drbd_helper is not None or |
| opts.enabled_hypervisors or opts.hvparams or |
| opts.beparams or opts.nicparams or |
| opts.ndparams or opts.diskparams or |
| opts.candidate_pool_size is not None or |
| opts.max_running_jobs is not None or |
| opts.max_tracked_jobs is not None or |
| opts.uid_pool is not None or |
| opts.maintain_node_health is not None or |
| opts.add_uids is not None or |
| opts.remove_uids is not None or |
| opts.default_iallocator is not None or |
| opts.default_iallocator_params is not None or |
| opts.reserved_lvs is not None or |
| opts.mac_prefix is not None or |
| opts.master_netdev is not None or |
| opts.master_netmask is not None or |
| opts.use_external_mip_script is not None or |
| opts.prealloc_wipe_disks is not None or |
| opts.hv_state or |
| opts.enabled_disk_templates or |
| opts.disk_state or |
| opts.ipolicy_bounds_specs is not None or |
| opts.ipolicy_std_specs is not None or |
| opts.ipolicy_disk_templates is not None or |
| opts.ipolicy_vcpu_ratio is not None or |
| opts.ipolicy_spindle_ratio is not None or |
| opts.ipolicy_memory_ratio is not None or |
| opts.modify_etc_hosts is not None or |
| opts.file_storage_dir is not None or |
| opts.install_image is not None or |
| opts.instance_communication_network is not None or |
| opts.zeroing_image is not None or |
| opts.shared_file_storage_dir is not None or |
| opts.compression_tools is not None or |
| opts.shared_file_storage_dir is not None or |
| opts.enabled_user_shutdown is not None or |
| opts.maint_round_delay is not None or |
| opts.maint_balance is not None or |
| opts.maint_balance_threshold is not None or |
| opts.data_collector_interval or |
| opts.diagnose_data_collector_filename is not None or |
| opts.enabled_data_collectors): |
| ToStderr("Please give at least one of the parameters.") |
| return 1 |
| |
| enabled_disk_templates = _GetEnabledDiskTemplates(opts) |
| vg_name = _GetVgName(opts, enabled_disk_templates) |
| |
| try: |
| drbd_helper = _GetDrbdHelper(opts, enabled_disk_templates) |
| except errors.OpPrereqError, e: |
| ToStderr(str(e)) |
| return 1 |
| |
| hvlist = opts.enabled_hypervisors |
| if hvlist is not None: |
| hvlist = hvlist.split(",") |
| |
| # a list of (name, dict) we can pass directly to dict() (or []) |
| hvparams = dict(opts.hvparams) |
| for hv_params in hvparams.values(): |
| utils.ForceDictType(hv_params, constants.HVS_PARAMETER_TYPES) |
| |
| diskparams = dict(opts.diskparams) |
| |
| for dt_params in diskparams.values(): |
| utils.ForceDictType(dt_params, constants.DISK_DT_TYPES) |
| |
| beparams = opts.beparams |
| utils.ForceDictType(beparams, constants.BES_PARAMETER_COMPAT) |
| |
| nicparams = opts.nicparams |
| utils.ForceDictType(nicparams, constants.NICS_PARAMETER_TYPES) |
| |
| ndparams = opts.ndparams |
| if ndparams is not None: |
| utils.ForceDictType(ndparams, constants.NDS_PARAMETER_TYPES) |
| |
| ipolicy = CreateIPolicyFromOpts( |
| minmax_ispecs=opts.ipolicy_bounds_specs, |
| std_ispecs=opts.ipolicy_std_specs, |
| ipolicy_disk_templates=opts.ipolicy_disk_templates, |
| ipolicy_vcpu_ratio=opts.ipolicy_vcpu_ratio, |
| ipolicy_spindle_ratio=opts.ipolicy_spindle_ratio, |
| ipolicy_memory_ratio=opts.ipolicy_memory_ratio, |
| ) |
| |
| mnh = opts.maintain_node_health |
| |
| uid_pool = opts.uid_pool |
| if uid_pool is not None: |
| uid_pool = uidpool.ParseUidPool(uid_pool) |
| |
| add_uids = opts.add_uids |
| if add_uids is not None: |
| add_uids = uidpool.ParseUidPool(add_uids) |
| |
| remove_uids = opts.remove_uids |
| if remove_uids is not None: |
| remove_uids = uidpool.ParseUidPool(remove_uids) |
| |
| if opts.reserved_lvs is not None: |
| if opts.reserved_lvs == "": |
| opts.reserved_lvs = [] |
| else: |
| opts.reserved_lvs = utils.UnescapeAndSplit(opts.reserved_lvs, sep=",") |
| |
| if opts.master_netmask is not None: |
| try: |
| opts.master_netmask = int(opts.master_netmask) |
| except ValueError: |
| ToStderr("The --master-netmask option expects an int parameter.") |
| return 1 |
| |
| ext_ip_script = opts.use_external_mip_script |
| |
| if opts.disk_state: |
| disk_state = utils.FlatToDict(opts.disk_state) |
| else: |
| disk_state = {} |
| |
| hv_state = dict(opts.hv_state) |
| |
| compression_tools = _GetCompressionTools(opts) |
| |
| enabled_data_collectors = dict( |
| (k, v.lower().startswith("t")) |
| for k, v in opts.enabled_data_collectors.items()) |
| |
| unrecognized_data_collectors = [ |
| k for k in enabled_data_collectors.keys() |
| if k not in constants.DATA_COLLECTOR_NAMES] |
| if unrecognized_data_collectors: |
| ToStderr("Data collector names not recognized: %s" % |
| ", ".join(unrecognized_data_collectors)) |
| |
| try: |
| data_collector_interval = dict( |
| (k, long(1e6 * float(v))) |
| for (k, v) in opts.data_collector_interval.items()) |
| except ValueError: |
| ToStderr("Can't transform all values to integers: {}".format( |
| opts.data_collector_interval)) |
| return 1 |
| if any(v <= 0 for v in data_collector_interval): |
| ToStderr("Some interval times where not above zero.") |
| return 1 |
| |
| op = opcodes.OpClusterSetParams( |
| vg_name=vg_name, |
| drbd_helper=drbd_helper, |
| enabled_hypervisors=hvlist, |
| hvparams=hvparams, |
| os_hvp=None, |
| beparams=beparams, |
| nicparams=nicparams, |
| ndparams=ndparams, |
| diskparams=diskparams, |
| ipolicy=ipolicy, |
| candidate_pool_size=opts.candidate_pool_size, |
| max_running_jobs=opts.max_running_jobs, |
| max_tracked_jobs=opts.max_tracked_jobs, |
| maintain_node_health=mnh, |
| modify_etc_hosts=opts.modify_etc_hosts, |
| uid_pool=uid_pool, |
| add_uids=add_uids, |
| remove_uids=remove_uids, |
| default_iallocator=opts.default_iallocator, |
| default_iallocator_params=opts.default_iallocator_params, |
| prealloc_wipe_disks=opts.prealloc_wipe_disks, |
| mac_prefix=opts.mac_prefix, |
| master_netdev=opts.master_netdev, |
| master_netmask=opts.master_netmask, |
| reserved_lvs=opts.reserved_lvs, |
| use_external_mip_script=ext_ip_script, |
| hv_state=hv_state, |
| disk_state=disk_state, |
| enabled_disk_templates=enabled_disk_templates, |
| force=opts.force, |
| file_storage_dir=opts.file_storage_dir, |
| install_image=opts.install_image, |
| instance_communication_network=opts.instance_communication_network, |
| zeroing_image=opts.zeroing_image, |
| shared_file_storage_dir=opts.shared_file_storage_dir, |
| compression_tools=compression_tools, |
| enabled_user_shutdown=opts.enabled_user_shutdown, |
| maint_round_delay=opts.maint_round_delay, |
| maint_balance=opts.maint_balance, |
| maint_balance_threshold=opts.maint_balance_threshold, |
| enabled_data_collectors=enabled_data_collectors, |
| data_collector_interval=data_collector_interval, |
| diagnose_data_collector_filename=opts.diagnose_data_collector_filename |
| ) |
| return base.GetResult(None, opts, SubmitOrSend(op, opts)) |
| |
| |
| def QueueOps(opts, args): |
| """Queue operations. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain only one element, the subcommand |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| command = args[0] |
| client = GetClient() |
| if command in ("drain", "undrain"): |
| drain_flag = command == "drain" |
| client.SetQueueDrainFlag(drain_flag) |
| elif command == "info": |
| result = client.QueryConfigValues(["drain_flag"]) |
| if result[0]: |
| val = "set" |
| else: |
| val = "unset" |
| ToStdout("The drain flag is %s" % val) |
| else: |
| raise errors.OpPrereqError("Command '%s' is not valid." % command, |
| errors.ECODE_INVAL) |
| |
| return 0 |
| |
| |
| def _ShowWatcherPause(until): |
| if until is None or until < time.time(): |
| ToStdout("The watcher is not paused.") |
| else: |
| ToStdout("The watcher is paused until %s.", time.ctime(until)) |
| |
| |
| def WatcherOps(opts, args): |
| """Watcher operations. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain only one element, the subcommand |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| command = args[0] |
| client = GetClient() |
| |
| if command == "continue": |
| client.SetWatcherPause(None) |
| ToStdout("The watcher is no longer paused.") |
| |
| elif command == "pause": |
| if len(args) < 2: |
| raise errors.OpPrereqError("Missing pause duration", errors.ECODE_INVAL) |
| |
| result = client.SetWatcherPause(time.time() + ParseTimespec(args[1])) |
| _ShowWatcherPause(result) |
| |
| elif command == "info": |
| result = client.QueryConfigValues(["watcher_pause"]) |
| _ShowWatcherPause(result[0]) |
| |
| else: |
| raise errors.OpPrereqError("Command '%s' is not valid." % command, |
| errors.ECODE_INVAL) |
| |
| return 0 |
| |
| |
| def _OobPower(opts, node_list, power): |
| """Puts the node in the list to desired power state. |
| |
| @param opts: The command line options selected by the user |
| @param node_list: The list of nodes to operate on |
| @param power: True if they should be powered on, False otherwise |
| @return: The success of the operation (none failed) |
| |
| """ |
| if power: |
| command = constants.OOB_POWER_ON |
| else: |
| command = constants.OOB_POWER_OFF |
| |
| op = opcodes.OpOobCommand(node_names=node_list, |
| command=command, |
| ignore_status=True, |
| timeout=opts.oob_timeout, |
| power_delay=opts.power_delay) |
| result = SubmitOpCode(op, opts=opts) |
| errs = 0 |
| for node_result in result: |
| (node_tuple, data_tuple) = node_result |
| (_, node_name) = node_tuple |
| (data_status, _) = data_tuple |
| if data_status != constants.RS_NORMAL: |
| assert data_status != constants.RS_UNAVAIL |
| errs += 1 |
| ToStderr("There was a problem changing power for %s, please investigate", |
| node_name) |
| |
| if errs > 0: |
| return False |
| |
| return True |
| |
| |
| def _InstanceStart(opts, inst_list, start, no_remember=False): |
| """Puts the instances in the list to desired state. |
| |
| @param opts: The command line options selected by the user |
| @param inst_list: The list of instances to operate on |
| @param start: True if they should be started, False for shutdown |
| @param no_remember: If the instance state should be remembered |
| @return: The success of the operation (none failed) |
| |
| """ |
| if start: |
| opcls = opcodes.OpInstanceStartup |
| text_submit, text_success, text_failed = ("startup", "started", "starting") |
| else: |
| opcls = compat.partial(opcodes.OpInstanceShutdown, |
| timeout=opts.shutdown_timeout, |
| no_remember=no_remember) |
| text_submit, text_success, text_failed = ("shutdown", "stopped", "stopping") |
| |
| jex = JobExecutor(opts=opts) |
| |
| for inst in inst_list: |
| ToStdout("Submit %s of instance %s", text_submit, inst) |
| op = opcls(instance_name=inst) |
| jex.QueueJob(inst, op) |
| |
| results = jex.GetResults() |
| bad_cnt = len([1 for (success, _) in results if not success]) |
| |
| if bad_cnt == 0: |
| ToStdout("All instances have been %s successfully", text_success) |
| else: |
| ToStderr("There were errors while %s instances:\n" |
| "%d error(s) out of %d instance(s)", text_failed, bad_cnt, |
| len(results)) |
| return False |
| |
| return True |
| |
| |
| class _RunWhenNodesReachableHelper(object): |
| """Helper class to make shared internal state sharing easier. |
| |
| @ivar success: Indicates if all action_cb calls were successful |
| |
| """ |
| def __init__(self, node_list, action_cb, node2ip, port, feedback_fn, |
| _ping_fn=netutils.TcpPing, _sleep_fn=time.sleep): |
| """Init the object. |
| |
| @param node_list: The list of nodes to be reachable |
| @param action_cb: Callback called when a new host is reachable |
| @type node2ip: dict |
| @param node2ip: Node to ip mapping |
| @param port: The port to use for the TCP ping |
| @param feedback_fn: The function used for feedback |
| @param _ping_fn: Function to check reachabilty (for unittest use only) |
| @param _sleep_fn: Function to sleep (for unittest use only) |
| |
| """ |
| self.down = set(node_list) |
| self.up = set() |
| self.node2ip = node2ip |
| self.success = True |
| self.action_cb = action_cb |
| self.port = port |
| self.feedback_fn = feedback_fn |
| self._ping_fn = _ping_fn |
| self._sleep_fn = _sleep_fn |
| |
| def __call__(self): |
| """When called we run action_cb. |
| |
| @raises utils.RetryAgain: When there are still down nodes |
| |
| """ |
| if not self.action_cb(self.up): |
| self.success = False |
| |
| if self.down: |
| raise utils.RetryAgain() |
| else: |
| return self.success |
| |
| def Wait(self, secs): |
| """Checks if a host is up or waits remaining seconds. |
| |
| @param secs: The secs remaining |
| |
| """ |
| start = time.time() |
| for node in self.down: |
| if self._ping_fn(self.node2ip[node], self.port, timeout=_EPO_PING_TIMEOUT, |
| live_port_needed=True): |
| self.feedback_fn("Node %s became available" % node) |
| self.up.add(node) |
| self.down -= self.up |
| # If we have a node available there is the possibility to run the |
| # action callback successfully, therefore we don't wait and return |
| return |
| |
| self._sleep_fn(max(0.0, start + secs - time.time())) |
| |
| |
| def _RunWhenNodesReachable(node_list, action_cb, interval): |
| """Run action_cb when nodes become reachable. |
| |
| @param node_list: The list of nodes to be reachable |
| @param action_cb: Callback called when a new host is reachable |
| @param interval: The earliest time to retry |
| |
| """ |
| client = GetClient() |
| cluster_info = client.QueryClusterInfo() |
| if cluster_info["primary_ip_version"] == constants.IP4_VERSION: |
| family = netutils.IPAddress.family |
| else: |
| family = netutils.IP6Address.family |
| |
| node2ip = dict((node, netutils.GetHostname(node, family=family).ip) |
| for node in node_list) |
| |
| port = netutils.GetDaemonPort(constants.NODED) |
| helper = _RunWhenNodesReachableHelper(node_list, action_cb, node2ip, port, |
| ToStdout) |
| |
| try: |
| return utils.Retry(helper, interval, _EPO_REACHABLE_TIMEOUT, |
| wait_fn=helper.Wait) |
| except utils.RetryTimeout: |
| ToStderr("Time exceeded while waiting for nodes to become reachable" |
| " again:\n - %s", " - ".join(helper.down)) |
| return False |
| |
| |
| def _MaybeInstanceStartup(opts, inst_map, nodes_online, |
| _instance_start_fn=_InstanceStart): |
| """Start the instances conditional based on node_states. |
| |
| @param opts: The command line options selected by the user |
| @param inst_map: A dict of inst -> nodes mapping |
| @param nodes_online: A list of nodes online |
| @param _instance_start_fn: Callback to start instances (unittest use only) |
| @return: Success of the operation on all instances |
| |
| """ |
| start_inst_list = [] |
| for (inst, nodes) in inst_map.items(): |
| if not (nodes - nodes_online): |
| # All nodes the instance lives on are back online |
| start_inst_list.append(inst) |
| |
| for inst in start_inst_list: |
| del inst_map[inst] |
| |
| if start_inst_list: |
| return _instance_start_fn(opts, start_inst_list, True) |
| |
| return True |
| |
| |
| def _EpoOn(opts, full_node_list, node_list, inst_map): |
| """Does the actual power on. |
| |
| @param opts: The command line options selected by the user |
| @param full_node_list: All nodes to operate on (includes nodes not supporting |
| OOB) |
| @param node_list: The list of nodes to operate on (all need to support OOB) |
| @param inst_map: A dict of inst -> nodes mapping |
| @return: The desired exit status |
| |
| """ |
| if node_list and not _OobPower(opts, node_list, False): |
| ToStderr("Not all nodes seem to get back up, investigate and start" |
| " manually if needed") |
| |
| # Wait for the nodes to be back up |
| action_cb = compat.partial(_MaybeInstanceStartup, opts, dict(inst_map)) |
| |
| ToStdout("Waiting until all nodes are available again") |
| if not _RunWhenNodesReachable(full_node_list, action_cb, _EPO_PING_INTERVAL): |
| ToStderr("Please investigate and start stopped instances manually") |
| return constants.EXIT_FAILURE |
| |
| return constants.EXIT_SUCCESS |
| |
| |
| def _EpoOff(opts, node_list, inst_map): |
| """Does the actual power off. |
| |
| @param opts: The command line options selected by the user |
| @param node_list: The list of nodes to operate on (all need to support OOB) |
| @param inst_map: A dict of inst -> nodes mapping |
| @return: The desired exit status |
| |
| """ |
| if not _InstanceStart(opts, inst_map.keys(), False, no_remember=True): |
| ToStderr("Please investigate and stop instances manually before continuing") |
| return constants.EXIT_FAILURE |
| |
| if not node_list: |
| return constants.EXIT_SUCCESS |
| |
| if _OobPower(opts, node_list, False): |
| return constants.EXIT_SUCCESS |
| else: |
| return constants.EXIT_FAILURE |
| |
| |
| def Epo(opts, args, qcl=None, _on_fn=_EpoOn, _off_fn=_EpoOff, |
| _confirm_fn=ConfirmOperation, |
| _stdout_fn=ToStdout, _stderr_fn=ToStderr): |
| """EPO operations. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should contain only one element, the subcommand |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| if opts.groups and opts.show_all: |
| _stderr_fn("Only one of --groups or --all are allowed") |
| return constants.EXIT_FAILURE |
| elif args and opts.show_all: |
| _stderr_fn("Arguments in combination with --all are not allowed") |
| return constants.EXIT_FAILURE |
| |
| if qcl is None: |
| # Query client |
| qcl = GetClient() |
| |
| if opts.groups: |
| node_query_list = \ |
| itertools.chain(*qcl.QueryGroups(args, ["node_list"], False)) |
| else: |
| node_query_list = args |
| |
| result = qcl.QueryNodes(node_query_list, ["name", "master", "pinst_list", |
| "sinst_list", "powered", "offline"], |
| False) |
| |
| all_nodes = map(compat.fst, result) |
| node_list = [] |
| inst_map = {} |
| for (node, master, pinsts, sinsts, powered, offline) in result: |
| if not offline: |
| for inst in (pinsts + sinsts): |
| if inst in inst_map: |
| if not master: |
| inst_map[inst].add(node) |
| elif master: |
| inst_map[inst] = set() |
| else: |
| inst_map[inst] = set([node]) |
| |
| if master and opts.on: |
| # We ignore the master for turning on the machines, in fact we are |
| # already operating on the master at this point :) |
| continue |
| elif master and not opts.show_all: |
| _stderr_fn("%s is the master node, please do a master-failover to another" |
| " node not affected by the EPO or use --all if you intend to" |
| " shutdown the whole cluster", node) |
| return constants.EXIT_FAILURE |
| elif powered is None: |
| _stdout_fn("Node %s does not support out-of-band handling, it can not be" |
| " handled in a fully automated manner", node) |
| elif powered == opts.on: |
| _stdout_fn("Node %s is already in desired power state, skipping", node) |
| elif not offline or (offline and powered): |
| node_list.append(node) |
| |
| if not (opts.force or _confirm_fn(all_nodes, "nodes", "epo")): |
| return constants.EXIT_FAILURE |
| |
| if opts.on: |
| return _on_fn(opts, all_nodes, node_list, inst_map) |
| else: |
| return _off_fn(opts, node_list, inst_map) |
| |
| |
| def RemoveRepair(opts, args): |
| """Uncoditionally remove a repair event |
| |
| @param opts: the command line options selected by the user (ignored) |
| @type args: list |
| @param args: one element, the uuid of the event to remove |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| uuid = args[0] |
| wconfd.Client().RmMaintdIncident(uuid) |
| return 0 |
| |
| |
| def _GetCreateCommand(info): |
| buf = StringIO() |
| buf.write("gnt-cluster init") |
| PrintIPolicyCommand(buf, info["ipolicy"], False) |
| buf.write(" ") |
| buf.write(info["name"]) |
| return buf.getvalue() |
| |
| |
| def ShowCreateCommand(opts, args): |
| """Shows the command that can be used to re-create the cluster. |
| |
| Currently it works only for ipolicy specs. |
| |
| """ |
| cl = GetClient() |
| result = cl.QueryClusterInfo() |
| ToStdout(_GetCreateCommand(result)) |
| |
| |
| def _RunCommandAndReport(cmd): |
| """Run a command and report its output, iff it failed. |
| |
| @param cmd: the command to execute |
| @type cmd: list |
| @rtype: bool |
| @return: False, if the execution failed. |
| |
| """ |
| result = utils.RunCmd(cmd) |
| if result.failed: |
| ToStderr("Command %s failed: %s; Output %s" % |
| (cmd, result.fail_reason, result.output)) |
| return False |
| return True |
| |
| |
| def _VerifyCommand(cmd): |
| """Verify that a given command succeeds on all online nodes. |
| |
| As this function is intended to run during upgrades, it |
| is implemented in such a way that it still works, if all Ganeti |
| daemons are down. |
| @param cmd: a list of unquoted shell arguments |
| @type cmd: list |
| @rtype: list |
| @return: the list of node names that are online where |
| the command failed. |
| |
| """ |
| command = utils.text.ShellQuoteArgs([str(val) for val in cmd]) |
| return _VerifyCommandRaw(command) |
| |
| |
| def _VerifyCommandRaw(command): |
| """Verify that a given command succeeds on all online nodes. |
| |
| As this function is intended to run during upgrades, it |
| is implemented in such a way that it still works, if all Ganeti |
| daemons are down. |
| @param cmd: a bare string to pass to SSH. The caller must do their |
| own shell/ssh escaping. |
| @type cmd: string |
| @rtype: list |
| @return: the list of node names that are online where |
| the command failed. |
| |
| """ |
| |
| nodes = ssconf.SimpleStore().GetOnlineNodeList() |
| master_node = ssconf.SimpleStore().GetMasterNode() |
| cluster_name = ssconf.SimpleStore().GetClusterName() |
| |
| # If master node is in 'nodes', make sure master node is at list end |
| if master_node in nodes: |
| nodes.remove(master_node) |
| nodes.append(master_node) |
| |
| failed = [] |
| |
| srun = ssh.SshRunner(cluster_name=cluster_name) |
| for name in nodes: |
| result = srun.Run(name, constants.SSH_LOGIN_USER, command) |
| if result.exit_code != 0: |
| failed.append(name) |
| |
| return failed |
| |
| |
| def _VerifyVersionInstalled(versionstring): |
| """Verify that the given version of ganeti is installed on all online nodes. |
| |
| Do nothing, if this is the case, otherwise print an appropriate |
| message to stderr. |
| |
| @param versionstring: the version to check for |
| @type versionstring: string |
| @rtype: bool |
| @return: True, if the version is installed on all online nodes |
| |
| """ |
| badnodes = _VerifyCommand(["test", "-d", |
| os.path.join(pathutils.PKGLIBDIR, versionstring)]) |
| if badnodes: |
| ToStderr("Ganeti version %s not installed on nodes %s" |
| % (versionstring, ", ".join(badnodes))) |
| return False |
| |
| return True |
| |
| |
| def _GetRunning(): |
| """Determine the list of running jobs. |
| |
| @rtype: list |
| @return: the number of jobs still running |
| |
| """ |
| cl = GetClient() |
| qfilter = qlang.MakeSimpleFilter("status", |
| frozenset([constants.JOB_STATUS_RUNNING])) |
| return len(cl.Query(constants.QR_JOB, [], qfilter).data) |
| |
| |
| def _SetGanetiVersionAndEnsure(versionstring): |
| """Symlink the active version of ganeti to the given versionstring, |
| and run the ensure-dirs script. |
| |
| @type versionstring: string |
| @rtype: list |
| @return: the list of nodes where the version change failed |
| |
| """ |
| |
| # Update symlinks to point at the new version. |
| if constants.HAS_GNU_LN: |
| link_lib_cmd = [ |
| "ln", "-s", "-f", "-T", |
| os.path.join(pathutils.PKGLIBDIR, versionstring), |
| os.path.join(pathutils.SYSCONFDIR, "ganeti/lib")] |
| link_share_cmd = [ |
| "ln", "-s", "-f", "-T", |
| os.path.join(pathutils.SHAREDIR, versionstring), |
| os.path.join(pathutils.SYSCONFDIR, "ganeti/share")] |
| cmds = [link_lib_cmd, link_share_cmd] |
| else: |
| rm_lib_cmd = [ |
| "rm", "-f", os.path.join(pathutils.SYSCONFDIR, "ganeti/lib")] |
| link_lib_cmd = [ |
| "ln", "-s", "-f", os.path.join(pathutils.PKGLIBDIR, versionstring), |
| os.path.join(pathutils.SYSCONFDIR, "ganeti/lib")] |
| rm_share_cmd = [ |
| "rm", "-f", os.path.join(pathutils.SYSCONFDIR, "ganeti/share")] |
| ln_share_cmd = [ |
| "ln", "-s", "-f", os.path.join(pathutils.SHAREDIR, versionstring), |
| os.path.join(pathutils.SYSCONFDIR, "ganeti/share")] |
| cmds = [rm_lib_cmd, link_lib_cmd, rm_share_cmd, ln_share_cmd] |
| |
| # Run the ensure-dirs script to verify the new version is OK. |
| cmds.append([pathutils.ENSURE_DIRS]) |
| |
| # Submit all commands to ssh, exiting on the first failure. |
| # The command string is a single argument that's given to ssh to submit to |
| # the remote shell, so it only needs enough escaping to satisfy the remote |
| # shell, rather than the 2 levels of escaping usually required when using |
| # ssh from the commandline. |
| quoted_cmds = [utils.text.ShellQuoteArgs(cmd) for cmd in cmds] |
| cmd = " && ".join(quoted_cmds) |
| failed = _VerifyCommandRaw(cmd) |
| return list(set(failed)) |
| |
| |
| def _ExecuteCommands(fns): |
| """Execute a list of functions, in reverse order. |
| |
| @type fns: list of functions. |
| @param fns: the functions to be executed. |
| |
| """ |
| for fn in reversed(fns): |
| fn() |
| |
| |
| def _GetConfigVersion(): |
| """Determine the version the configuration file currently has. |
| |
| @rtype: tuple or None |
| @return: (major, minor, revision) if the version can be determined, |
| None otherwise |
| |
| """ |
| config_data = serializer.LoadJson(utils.ReadFile(pathutils.CLUSTER_CONF_FILE)) |
| try: |
| config_version = config_data["version"] |
| except KeyError: |
| return None |
| return utils.SplitVersion(config_version) |
| |
| |
| def _ReadIntentToUpgrade(): |
| """Read the file documenting the intent to upgrade the cluster. |
| |
| @rtype: (string, string) or (None, None) |
| @return: (old version, version to upgrade to), if the file exists, |
| and (None, None) otherwise. |
| |
| """ |
| if not os.path.isfile(pathutils.INTENT_TO_UPGRADE): |
| return (None, None) |
| |
| contentstring = utils.ReadFile(pathutils.INTENT_TO_UPGRADE) |
| contents = utils.UnescapeAndSplit(contentstring) |
| if len(contents) != 3: |
| # file syntactically mal-formed |
| return (None, None) |
| return (contents[0], contents[1]) |
| |
| |
| def _WriteIntentToUpgrade(version): |
| """Write file documenting the intent to upgrade the cluster. |
| |
| @type version: string |
| @param version: the version we intent to upgrade to |
| |
| """ |
| utils.WriteFile(pathutils.INTENT_TO_UPGRADE, |
| data=utils.EscapeAndJoin([constants.RELEASE_VERSION, version, |
| "%d" % os.getpid()])) |
| |
| |
| def _UpgradeBeforeConfigurationChange(versionstring): |
| """ |
| Carry out all the tasks necessary for an upgrade that happen before |
| the configuration file, or Ganeti version, changes. |
| |
| @type versionstring: string |
| @param versionstring: the version to upgrade to |
| @rtype: (bool, list) |
| @return: tuple of a bool indicating success and a list of rollback tasks |
| |
| """ |
| rollback = [] |
| |
| ToStdoutAndLoginfo("Verifying %s present on all nodes", versionstring) |
| if not _VerifyVersionInstalled(versionstring): |
| return (False, rollback) |
| |
| _WriteIntentToUpgrade(versionstring) |
| rollback.append( |
| lambda: utils.RunCmd(["rm", "-f", pathutils.INTENT_TO_UPGRADE])) |
| |
| ToStdoutAndLoginfo("Draining queue") |
| client = GetClient() |
| client.SetQueueDrainFlag(True) |
| |
| rollback.append(lambda: GetClient().SetQueueDrainFlag(False)) |
| |
| if utils.SimpleRetry(0, _GetRunning, |
| constants.UPGRADE_QUEUE_POLL_INTERVAL, |
| constants.UPGRADE_QUEUE_DRAIN_TIMEOUT): |
| ToStderr("Failed to completely empty the queue.") |
| return (False, rollback) |
| |
| ToStdoutAndLoginfo("Pausing the watcher for one hour.") |
| rollback.append(lambda: GetClient().SetWatcherPause(None)) |
| GetClient().SetWatcherPause(time.time() + 60 * 60) |
| |
| ToStdoutAndLoginfo("Stopping daemons on master node.") |
| if not _RunCommandAndReport([pathutils.DAEMON_UTIL, "stop-all"]): |
| return (False, rollback) |
| |
| ToStdoutAndLoginfo("Stopping daemons everywhere.") |
| rollback.append(lambda: _VerifyCommand([pathutils.DAEMON_UTIL, "start-all"])) |
| badnodes = _VerifyCommand([pathutils.DAEMON_UTIL, "stop-all"]) |
| if badnodes: |
| ToStderr("Failed to stop daemons on %s." % (", ".join(badnodes),)) |
| return (False, rollback) |
| |
| backuptar = os.path.join(pathutils.BACKUP_DIR, "ganeti%d.tar" % time.time()) |
| ToStdoutAndLoginfo("Backing up configuration as %s", backuptar) |
| if not _RunCommandAndReport(["mkdir", "-p", pathutils.BACKUP_DIR]): |
| return (False, rollback) |
| |
| # Create the archive in a safe manner, as it contains sensitive |
| # information. |
| (_, tmp_name) = tempfile.mkstemp(prefix=backuptar, dir=pathutils.BACKUP_DIR) |
| if not _RunCommandAndReport(["tar", "-cf", tmp_name, |
| "--exclude=queue/archive", |
| pathutils.DATA_DIR]): |
| return (False, rollback) |
| |
| os.rename(tmp_name, backuptar) |
| return (True, rollback) |
| |
| |
| def _VersionSpecificDowngrade(): |
| """ |
| Perform any additional downrade tasks that are version specific |
| and need to be done just after the configuration downgrade. This |
| function needs to be idempotent, so that it can be redone if the |
| downgrade procedure gets interrupted after changing the |
| configuration. |
| |
| Note that this function has to be reset with every version bump. |
| |
| @return: True upon success |
| """ |
| ToStdoutAndLoginfo("Performing version-specific downgrade tasks.") |
| |
| return True |
| |
| |
| def _SwitchVersionAndConfig(versionstring, downgrade): |
| """ |
| Switch to the new Ganeti version and change the configuration, |
| in correct order. |
| |
| @type versionstring: string |
| @param versionstring: the version to change to |
| @type downgrade: bool |
| @param downgrade: True, if the configuration should be downgraded |
| @rtype: (bool, list) |
| @return: tupe of a bool indicating success, and a list of |
| additional rollback tasks |
| |
| """ |
| rollback = [] |
| if downgrade: |
| ToStdoutAndLoginfo("Downgrading configuration") |
| if not _RunCommandAndReport([pathutils.CFGUPGRADE, "--downgrade", "-f"]): |
| return (False, rollback) |
| # Note: version specific downgrades need to be done before switching |
| # binaries, so that we still have the knowledgeable binary if the downgrade |
| # process gets interrupted at this point. |
| if not _VersionSpecificDowngrade(): |
| return (False, rollback) |
| |
| # Configuration change is the point of no return. From then onwards, it is |
| # safer to push through the up/dowgrade than to try to roll it back. |
| |
| ToStdoutAndLoginfo("Switching to version %s on all nodes", versionstring) |
| rollback.append(lambda: _SetGanetiVersionAndEnsure(constants.DIR_VERSION)) |
| badnodes = _SetGanetiVersionAndEnsure(versionstring) |
| if badnodes: |
| ToStderr("Failed to switch to Ganeti version %s on nodes %s" |
| % (versionstring, ", ".join(badnodes))) |
| if not downgrade: |
| return (False, rollback) |
| |
| # Now that we have changed to the new version of Ganeti we should |
| # not communicate over luxi any more, as luxi might have changed in |
| # incompatible ways. Therefore, manually call the corresponding ganeti |
| # commands using their canonical (version independent) path. |
| |
| if not downgrade: |
| ToStdoutAndLoginfo("Upgrading configuration") |
| if not _RunCommandAndReport([pathutils.CFGUPGRADE, "-f"]): |
| return (False, rollback) |
| |
| return (True, rollback) |
| |
| |
| def _UpgradeAfterConfigurationChange(oldversion): |
| """ |
| Carry out the upgrade actions necessary after switching to the new |
| Ganeti version and updating the configuration. |
| |
| As this part is run at a time where the new version of Ganeti is already |
| running, no communication should happen via luxi, as this is not a stable |
| interface. Also, as the configuration change is the point of no return, |
| all actions are pushed through, even if some of them fail. |
| |
| @param oldversion: the version the upgrade started from |
| @type oldversion: string |
| @rtype: int |
| @return: the intended return value |
| |
| """ |
| returnvalue = 0 |
| |
| ToStdoutAndLoginfo("Starting daemons everywhere.") |
| badnodes = _VerifyCommand([pathutils.DAEMON_UTIL, "start-all"]) |
| if badnodes: |
| ToStderr("Warning: failed to start daemons on %s." % (", ".join(badnodes),)) |
| returnvalue = 1 |
| |
| ToStdoutAndLoginfo("Redistributing the configuration.") |
| if not _RunCommandAndReport(["gnt-cluster", "redist-conf", "--yes-do-it"]): |
| returnvalue = 1 |
| |
| ToStdoutAndLoginfo("Restarting daemons everywhere.") |
| badnodes = _VerifyCommand([pathutils.DAEMON_UTIL, "stop-all"]) |
| badnodes.extend(_VerifyCommand([pathutils.DAEMON_UTIL, "start-all"])) |
| if badnodes: |
| ToStderr("Warning: failed to start daemons on %s." % |
| (", ".join(list(set(badnodes))),)) |
| returnvalue = 1 |
| |
| ToStdoutAndLoginfo("Undraining the queue.") |
| if not _RunCommandAndReport(["gnt-cluster", "queue", "undrain"]): |
| returnvalue = 1 |
| |
| _RunCommandAndReport(["rm", "-f", pathutils.INTENT_TO_UPGRADE]) |
| |
| ToStdoutAndLoginfo("Running post-upgrade hooks") |
| if not _RunCommandAndReport([pathutils.POST_UPGRADE, oldversion]): |
| returnvalue = 1 |
| |
| ToStdoutAndLoginfo("Unpausing the watcher.") |
| if not _RunCommandAndReport(["gnt-cluster", "watcher", "continue"]): |
| returnvalue = 1 |
| |
| ToStdoutAndLoginfo("Verifying cluster.") |
| if not _RunCommandAndReport(["gnt-cluster", "verify"]): |
| returnvalue = 1 |
| |
| return returnvalue |
| |
| |
| def UpgradeGanetiCommand(opts, args): |
| """Upgrade a cluster to a new ganeti version. |
| |
| @param opts: the command line options selected by the user |
| @type args: list |
| @param args: should be an empty list |
| @rtype: int |
| @return: the desired exit code |
| |
| """ |
| if ((not opts.resume and opts.to is None) |
| or (opts.resume and opts.to is not None)): |
| ToStderr("Precisely one of the options --to and --resume" |
| " has to be given") |
| return 1 |
| |
| # If we're not told to resume, verify there is no upgrade |
| # in progress. |
| if not opts.resume: |
| oldversion, versionstring = _ReadIntentToUpgrade() |
| if versionstring is not None: |
| # An upgrade is going on; verify whether the target matches |
| if versionstring == opts.to: |
| ToStderr("An upgrade is already in progress. Target version matches," |
| " resuming.") |
| opts.resume = True |
| opts.to = None |
| else: |
| ToStderr("An upgrade from %s to %s is in progress; use --resume to" |
| " finish it first" % (oldversion, versionstring)) |
| return 1 |
| |
| utils.SetupLogging(pathutils.LOG_COMMANDS, 'gnt-cluster upgrade', debug=1) |
| |
| oldversion = constants.RELEASE_VERSION |
| |
| if opts.resume: |
| ssconf.CheckMaster(False) |
| oldversion, versionstring = _ReadIntentToUpgrade() |
| if versionstring is None: |
| return 0 |
| version = utils.version.ParseVersion(versionstring) |
| if version is None: |
| return 1 |
| configversion = _GetConfigVersion() |
| if configversion is None: |
| return 1 |
| # If the upgrade we resume was an upgrade between compatible |
| # versions (like 2.10.0 to 2.10.1), the correct configversion |
| # does not guarantee that the config has been updated. |
| # However, in the case of a compatible update with the configuration |
| # not touched, we are running a different dirversion with the same |
| # config version. |
| config_already_modified = \ |
| (utils.IsCorrectConfigVersion(version, configversion) and |
| not (versionstring != constants.DIR_VERSION and |
| configversion == (constants.CONFIG_MAJOR, constants.CONFIG_MINOR, |
| constants.CONFIG_REVISION))) |
| if not config_already_modified: |
| # We have to start from the beginning; however, some daemons might have |
| # already been stopped, so the only way to get into a well-defined state |
| # is by starting all daemons again. |
| _VerifyCommand([pathutils.DAEMON_UTIL, "start-all"]) |
| else: |
| versionstring = opts.to |
| config_already_modified = False |
| version = utils.version.ParseVersion(versionstring) |
| if version is None: |
| ToStderr("Could not parse version string %s" % versionstring) |
| return 1 |
| |
| msg = utils.version.UpgradeRange(version) |
| if msg is not None: |
| ToStderr("Cannot upgrade to %s: %s" % (versionstring, msg)) |
| return 1 |
| |
| if not config_already_modified: |
| success, rollback = _UpgradeBeforeConfigurationChange(versionstring) |
| if not success: |
| _ExecuteCommands(rollback) |
| return 1 |
| else: |
| rollback = [] |
| |
| downgrade = utils.version.ShouldCfgdowngrade(version) |
| |
| success, additionalrollback = \ |
| _SwitchVersionAndConfig(versionstring, downgrade) |
| if not success: |
| rollback.extend(additionalrollback) |
| _ExecuteCommands(rollback) |
| return 1 |
| |
| return _UpgradeAfterConfigurationChange(oldversion) |
| |
| |
| commands = { |
| "init": ( |
| InitCluster, [ArgHost(min=1, max=1)], |
| [BACKEND_OPT, CP_SIZE_OPT, ENABLED_HV_OPT, GLOBAL_FILEDIR_OPT, |
| HVLIST_OPT, MAC_PREFIX_OPT, MASTER_NETDEV_OPT, MASTER_NETMASK_OPT, |
| NIC_PARAMS_OPT, NOMODIFY_ETCHOSTS_OPT, NOMODIFY_SSH_SETUP_OPT, |
| SECONDARY_IP_OPT, VG_NAME_OPT, MAINTAIN_NODE_HEALTH_OPT, UIDPOOL_OPT, |
| DRBD_HELPER_OPT, DEFAULT_IALLOCATOR_OPT, DEFAULT_IALLOCATOR_PARAMS_OPT, |
| PRIMARY_IP_VERSION_OPT, PREALLOC_WIPE_DISKS_OPT, NODE_PARAMS_OPT, |
| GLOBAL_SHARED_FILEDIR_OPT, USE_EXTERNAL_MIP_SCRIPT, DISK_PARAMS_OPT, |
| HV_STATE_OPT, DISK_STATE_OPT, ENABLED_DISK_TEMPLATES_OPT, |
| IPOLICY_STD_SPECS_OPT, GLOBAL_GLUSTER_FILEDIR_OPT, INSTALL_IMAGE_OPT, |
| ZEROING_IMAGE_OPT, COMPRESSION_TOOLS_OPT, |
| ENABLED_USER_SHUTDOWN_OPT, SSH_KEY_BITS_OPT, SSH_KEY_TYPE_OPT, |
| ] |
| + INSTANCE_POLICY_OPTS + SPLIT_ISPECS_OPTS, |
| "[opts...] <cluster_name>", "Initialises a new cluster configuration"), |
| "destroy": ( |
| DestroyCluster, ARGS_NONE, [YES_DOIT_OPT], |
| "", "Destroy cluster"), |
| "rename": ( |
| RenameCluster, [ArgHost(min=1, max=1)], |
| [FORCE_OPT, DRY_RUN_OPT], |
| "<new_name>", |
| "Renames the cluster"), |
| "redist-conf": ( |
| RedistributeConfig, ARGS_NONE, SUBMIT_OPTS + |
| [DRY_RUN_OPT, PRIORITY_OPT, FORCE_DISTRIBUTION], |
| "", "Forces a push of the configuration file and ssconf files" |
| " to the nodes in the cluster"), |
| "verify": ( |
| VerifyCluster, ARGS_NONE, |
| [VERBOSE_OPT, DEBUG_SIMERR_OPT, ERROR_CODES_OPT, NONPLUS1_OPT, |
| DRY_RUN_OPT, PRIORITY_OPT, NODEGROUP_OPT, IGNORE_ERRORS_OPT, |
| VERIFY_CLUTTER_OPT], |
| "", "Does a check on the cluster configuration"), |
| "verify-disks": ( |
| VerifyDisks, ARGS_NONE, [PRIORITY_OPT, NODEGROUP_OPT, STRICT_OPT], |
| "", "Does a check on the cluster disk status"), |
| "repair-disk-sizes": ( |
| RepairDiskSizes, ARGS_MANY_INSTANCES, [DRY_RUN_OPT, PRIORITY_OPT], |
| "[instance...]", "Updates mismatches in recorded disk sizes"), |
| "master-failover": ( |
| MasterFailover, ARGS_NONE, |
| [NOVOTING_OPT, FORCE_FAILOVER, IGNORE_OFFLINE_NODES_FAILOVER], |
| "", "Makes the current node the master"), |
| "master-ping": ( |
| MasterPing, ARGS_NONE, [], |
| "", "Checks if the master is alive"), |
| "version": ( |
| ShowClusterVersion, ARGS_NONE, [], |
| "", "Shows the cluster version"), |
| "getmaster": ( |
| ShowClusterMaster, ARGS_NONE, [], |
| "", "Shows the cluster master"), |
| "copyfile": ( |
| ClusterCopyFile, [ArgFile(min=1, max=1)], |
| [NODE_LIST_OPT, USE_REPL_NET_OPT, NODEGROUP_OPT], |
| "[-n node...] <filename>", "Copies a file to all (or only some) nodes"), |
| "command": ( |
| RunClusterCommand, [ArgCommand(min=1)], |
| [NODE_LIST_OPT, NODEGROUP_OPT, SHOW_MACHINE_OPT, FAILURE_ONLY_OPT], |
| "[-n node...] <command>", "Runs a command on all (or only some) nodes"), |
| "info": ( |
| ShowClusterConfig, ARGS_NONE, [ROMAN_OPT], |
| "[--roman]", "Show cluster configuration"), |
| "list-tags": ( |
| ListTags, ARGS_NONE, [], "", "List the tags of the cluster"), |
| "add-tags": ( |
| AddTags, [ArgUnknown()], [TAG_SRC_OPT, PRIORITY_OPT] + SUBMIT_OPTS, |
| "tag...", "Add tags to the cluster"), |
| "remove-tags": ( |
| RemoveTags, [ArgUnknown()], [TAG_SRC_OPT, PRIORITY_OPT] + SUBMIT_OPTS, |
| "tag...", "Remove tags from the cluster"), |
| "search-tags": ( |
| SearchTags, [ArgUnknown(min=1, max=1)], [PRIORITY_OPT], "", |
| "Searches the tags on all objects on" |
| " the cluster for a given pattern (regex)"), |
| "queue": ( |
| QueueOps, |
| [ArgChoice(min=1, max=1, choices=["drain", "undrain", "info"])], |
| [], "drain|undrain|info", "Change queue properties"), |
| "watcher": ( |
| WatcherOps, |
| [ArgChoice(min=1, max=1, choices=["pause", "continue", "info"]), |
| ArgSuggest(min=0, max=1, choices=["30m", "1h", "4h"])], |
| [], |
| "{pause <timespec>|continue|info}", "Change watcher properties"), |
| "modify": ( |
| SetClusterParams, ARGS_NONE, |
| [FORCE_OPT, |
| BACKEND_OPT, CP_SIZE_OPT, RQL_OPT, MAX_TRACK_OPT, INSTALL_IMAGE_OPT, |
| INSTANCE_COMMUNICATION_NETWORK_OPT, ENABLED_HV_OPT, HVLIST_OPT, |
| MAC_PREFIX_OPT, MASTER_NETDEV_OPT, MASTER_NETMASK_OPT, NIC_PARAMS_OPT, |
| VG_NAME_OPT, MAINTAIN_NODE_HEALTH_OPT, UIDPOOL_OPT, ADD_UIDS_OPT, |
| REMOVE_UIDS_OPT, DRBD_HELPER_OPT, DEFAULT_IALLOCATOR_OPT, |
| DEFAULT_IALLOCATOR_PARAMS_OPT, RESERVED_LVS_OPT, DRY_RUN_OPT, PRIORITY_OPT, |
| PREALLOC_WIPE_DISKS_OPT, NODE_PARAMS_OPT, USE_EXTERNAL_MIP_SCRIPT, |
| DISK_PARAMS_OPT, HV_STATE_OPT, DISK_STATE_OPT] + SUBMIT_OPTS + |
| [ENABLED_DISK_TEMPLATES_OPT, IPOLICY_STD_SPECS_OPT, MODIFY_ETCHOSTS_OPT, |
| ENABLED_USER_SHUTDOWN_OPT] + |
| INSTANCE_POLICY_OPTS + |
| [GLOBAL_FILEDIR_OPT, GLOBAL_SHARED_FILEDIR_OPT, ZEROING_IMAGE_OPT, |
| COMPRESSION_TOOLS_OPT] + |
| [ENABLED_DATA_COLLECTORS_OPT, DATA_COLLECTOR_INTERVAL_OPT, |
| DIAGNOSE_DATA_COLLECTOR_FILENAME_OPT, |
| MAINT_INTERVAL_OPT, MAINT_BALANCE_OPT, MAINT_BALANCE_THRESHOLD_OPT], |
| "[opts...]", |
| "Alters the parameters of the cluster"), |
| "renew-crypto": ( |
| RenewCrypto, ARGS_NONE, |
| [NEW_CLUSTER_CERT_OPT, NEW_RAPI_CERT_OPT, RAPI_CERT_OPT, |
| NEW_CONFD_HMAC_KEY_OPT, FORCE_OPT, |
| NEW_CLUSTER_DOMAIN_SECRET_OPT, CLUSTER_DOMAIN_SECRET_OPT, |
| NEW_SPICE_CERT_OPT, SPICE_CERT_OPT, SPICE_CACERT_OPT, |
| NEW_NODE_CERT_OPT, NEW_SSH_KEY_OPT, NOSSH_KEYCHECK_OPT, |
| VERBOSE_OPT, SSH_KEY_BITS_OPT, SSH_KEY_TYPE_OPT], |
| "[opts...]", |
| "Renews cluster certificates, keys and secrets"), |
| "epo": ( |
| Epo, [ArgUnknown()], |
| [FORCE_OPT, ON_OPT, GROUPS_OPT, ALL_OPT, OOB_TIMEOUT_OPT, |
| SHUTDOWN_TIMEOUT_OPT, POWER_DELAY_OPT], |
| "[opts...] [args]", |
| "Performs an emergency power-off on given args"), |
| "activate-master-ip": ( |
| ActivateMasterIp, ARGS_NONE, [], "", "Activates the master IP"), |
| "deactivate-master-ip": ( |
| DeactivateMasterIp, ARGS_NONE, [CONFIRM_OPT], "", |
| "Deactivates the master IP"), |
| "show-ispecs-cmd": ( |
| ShowCreateCommand, ARGS_NONE, [], "", |
| "Show the command line to re-create the cluster"), |
| "upgrade": ( |
| UpgradeGanetiCommand, ARGS_NONE, [TO_OPT, RESUME_OPT], "", |
| "Upgrade (or downgrade) to a new Ganeti version"), |
| "remove-repair": ( |
| RemoveRepair, [ArgUnknown()], [], "<uuid>", |
| "Remove a repair event from the list of pending events"), |
| } |
| |
| |
| #: dictionary with aliases for commands |
| aliases = { |
| "masterfailover": "master-failover", |
| "show": "info", |
| } |
| |
| |
| def Main(): |
| return GenericMain(commands, override={"tag_type": constants.TAG_CLUSTER}, |
| aliases=aliases) |