2020-03-09 10:11:07 +01:00
|
|
|
#!/usr/bin/python
|
|
|
|
# -*- coding: utf-8 -*-
|
|
|
|
|
|
|
|
# Copyright: (c) 2014, Taneli Leppä <taneli@crasman.fi>
|
|
|
|
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
|
|
|
|
|
|
|
from __future__ import absolute_import, division, print_function
|
|
|
|
__metaclass__ = type
|
|
|
|
|
|
|
|
DOCUMENTATION = '''
|
|
|
|
module: gluster_volume
|
|
|
|
short_description: Manage GlusterFS volumes
|
|
|
|
description:
|
|
|
|
- Create, remove, start, stop and tune GlusterFS volumes
|
|
|
|
options:
|
|
|
|
name:
|
|
|
|
description:
|
|
|
|
- The volume name.
|
|
|
|
required: true
|
|
|
|
aliases: ['volume']
|
|
|
|
state:
|
|
|
|
description:
|
|
|
|
- Use present/absent ensure if a volume exists or not.
|
|
|
|
Use started/stopped to control its availability.
|
|
|
|
required: true
|
|
|
|
choices: ['absent', 'present', 'started', 'stopped']
|
|
|
|
cluster:
|
|
|
|
description:
|
|
|
|
- List of hosts to use for probing and brick setup.
|
|
|
|
host:
|
|
|
|
description:
|
|
|
|
- Override local hostname (for peer probing purposes).
|
|
|
|
replicas:
|
|
|
|
description:
|
|
|
|
- Replica count for volume.
|
|
|
|
arbiters:
|
|
|
|
description:
|
|
|
|
- Arbiter count for volume.
|
|
|
|
stripes:
|
|
|
|
description:
|
|
|
|
- Stripe count for volume.
|
|
|
|
disperses:
|
|
|
|
description:
|
|
|
|
- Disperse count for volume.
|
|
|
|
redundancies:
|
|
|
|
description:
|
|
|
|
- Redundancy count for volume.
|
|
|
|
transport:
|
|
|
|
description:
|
|
|
|
- Transport type for volume.
|
|
|
|
default: tcp
|
|
|
|
choices: [ tcp, rdma, 'tcp,rdma' ]
|
|
|
|
bricks:
|
|
|
|
description:
|
|
|
|
- Brick paths on servers. Multiple brick paths can be separated by commas.
|
|
|
|
aliases: [ brick ]
|
|
|
|
start_on_create:
|
|
|
|
description:
|
|
|
|
- Controls whether the volume is started after creation or not.
|
|
|
|
type: bool
|
|
|
|
default: 'yes'
|
|
|
|
rebalance:
|
|
|
|
description:
|
|
|
|
- Controls whether the cluster is rebalanced after changes.
|
|
|
|
type: bool
|
|
|
|
default: 'no'
|
|
|
|
directory:
|
|
|
|
description:
|
|
|
|
- Directory for limit-usage.
|
|
|
|
options:
|
|
|
|
description:
|
|
|
|
- A dictionary/hash with options/settings for the volume.
|
|
|
|
quota:
|
|
|
|
description:
|
|
|
|
- Quota value for limit-usage (be sure to use 10.0MB instead of 10MB, see quota list).
|
|
|
|
force:
|
|
|
|
description:
|
|
|
|
- If brick is being created in the root partition, module will fail.
|
|
|
|
Set force to true to override this behaviour.
|
|
|
|
type: bool
|
2020-11-23 12:29:14 +01:00
|
|
|
default: false
|
2020-03-09 10:11:07 +01:00
|
|
|
notes:
|
|
|
|
- Requires cli tools for GlusterFS on servers.
|
|
|
|
- Will add new bricks, but not remove them.
|
|
|
|
author:
|
|
|
|
- Taneli Leppä (@rosmo)
|
|
|
|
'''
|
|
|
|
|
|
|
|
EXAMPLES = """
|
2020-05-16 15:07:51 +02:00
|
|
|
- name: Create gluster volume
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: present
|
|
|
|
name: test1
|
|
|
|
bricks: /bricks/brick1/g1
|
|
|
|
rebalance: yes
|
|
|
|
cluster:
|
|
|
|
- 192.0.2.10
|
|
|
|
- 192.0.2.11
|
|
|
|
run_once: true
|
|
|
|
|
2020-05-16 15:07:51 +02:00
|
|
|
- name: Tune
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: present
|
|
|
|
name: test1
|
|
|
|
options:
|
|
|
|
performance.cache-size: 256MB
|
|
|
|
|
|
|
|
- name: Set multiple options on GlusterFS volume
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: present
|
|
|
|
name: test1
|
|
|
|
options:
|
|
|
|
{ performance.cache-size: 128MB,
|
|
|
|
write-behind: 'off',
|
|
|
|
quick-read: 'on'
|
|
|
|
}
|
|
|
|
|
2020-05-16 15:07:51 +02:00
|
|
|
- name: Start gluster volume
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: started
|
|
|
|
name: test1
|
|
|
|
|
2020-05-16 15:07:51 +02:00
|
|
|
- name: Limit usage
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: present
|
|
|
|
name: test1
|
|
|
|
directory: /foo
|
|
|
|
quota: 20.0MB
|
|
|
|
|
2020-05-16 15:07:51 +02:00
|
|
|
- name: Stop gluster volume
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: stopped
|
|
|
|
name: test1
|
|
|
|
|
2020-05-16 15:07:51 +02:00
|
|
|
- name: Remove gluster volume
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: absent
|
|
|
|
name: test1
|
|
|
|
|
2020-05-16 15:07:51 +02:00
|
|
|
- name: Create gluster volume with multiple bricks
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: present
|
|
|
|
name: test2
|
|
|
|
bricks: /bricks/brick1/g2,/bricks/brick2/g2
|
|
|
|
cluster:
|
|
|
|
- 192.0.2.10
|
|
|
|
- 192.0.2.11
|
|
|
|
run_once: true
|
|
|
|
|
|
|
|
- name: Remove the bricks from gluster volume
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: present
|
|
|
|
name: testvol
|
|
|
|
bricks: /bricks/brick1/b1,/bricks/brick2/b2
|
|
|
|
cluster:
|
|
|
|
- 10.70.42.85
|
|
|
|
force: true
|
|
|
|
run_once: true
|
|
|
|
|
|
|
|
- name: Reduce cluster configuration
|
2020-07-13 21:50:31 +02:00
|
|
|
community.general.gluster_volume:
|
2020-03-09 10:11:07 +01:00
|
|
|
state: present
|
|
|
|
name: testvol
|
|
|
|
bricks: /bricks/brick3/b1,/bricks/brick4/b2
|
|
|
|
replicas: 2
|
|
|
|
cluster:
|
|
|
|
- 10.70.42.85
|
|
|
|
force: true
|
|
|
|
run_once: true
|
|
|
|
"""
|
|
|
|
|
|
|
|
import re
|
|
|
|
import socket
|
|
|
|
import time
|
|
|
|
import traceback
|
|
|
|
|
|
|
|
from ansible.module_utils.basic import AnsibleModule
|
|
|
|
from ansible.module_utils._text import to_native
|
|
|
|
|
|
|
|
glusterbin = ''
|
|
|
|
|
|
|
|
|
|
|
|
def run_gluster(gargs, **kwargs):
|
|
|
|
global glusterbin
|
|
|
|
global module
|
|
|
|
args = [glusterbin, '--mode=script']
|
|
|
|
args.extend(gargs)
|
|
|
|
try:
|
|
|
|
rc, out, err = module.run_command(args, **kwargs)
|
|
|
|
if rc != 0:
|
|
|
|
module.fail_json(msg='error running gluster (%s) command (rc=%d): %s' %
|
|
|
|
(' '.join(args), rc, out or err), exception=traceback.format_exc())
|
|
|
|
except Exception as e:
|
|
|
|
module.fail_json(msg='error running gluster (%s) command: %s' % (' '.join(args),
|
|
|
|
to_native(e)), exception=traceback.format_exc())
|
|
|
|
return out
|
|
|
|
|
|
|
|
|
|
|
|
def run_gluster_nofail(gargs, **kwargs):
|
|
|
|
global glusterbin
|
|
|
|
global module
|
|
|
|
args = [glusterbin]
|
|
|
|
args.extend(gargs)
|
|
|
|
rc, out, err = module.run_command(args, **kwargs)
|
|
|
|
if rc != 0:
|
|
|
|
return None
|
|
|
|
return out
|
|
|
|
|
|
|
|
|
|
|
|
def get_peers():
|
|
|
|
out = run_gluster(['peer', 'status'])
|
|
|
|
peers = {}
|
|
|
|
hostname = None
|
|
|
|
uuid = None
|
|
|
|
state = None
|
|
|
|
shortNames = False
|
|
|
|
for row in out.split('\n'):
|
|
|
|
if ': ' in row:
|
|
|
|
key, value = row.split(': ')
|
|
|
|
if key.lower() == 'hostname':
|
|
|
|
hostname = value
|
|
|
|
shortNames = False
|
|
|
|
if key.lower() == 'uuid':
|
|
|
|
uuid = value
|
|
|
|
if key.lower() == 'state':
|
|
|
|
state = value
|
|
|
|
peers[hostname] = [uuid, state]
|
|
|
|
elif row.lower() == 'other names:':
|
|
|
|
shortNames = True
|
|
|
|
elif row != '' and shortNames is True:
|
|
|
|
peers[row] = [uuid, state]
|
|
|
|
elif row == '':
|
|
|
|
shortNames = False
|
|
|
|
return peers
|
|
|
|
|
|
|
|
|
|
|
|
def get_volumes():
|
|
|
|
out = run_gluster(['volume', 'info'])
|
|
|
|
|
|
|
|
volumes = {}
|
|
|
|
volume = {}
|
|
|
|
for row in out.split('\n'):
|
|
|
|
if ': ' in row:
|
|
|
|
key, value = row.split(': ')
|
|
|
|
if key.lower() == 'volume name':
|
|
|
|
volume['name'] = value
|
|
|
|
volume['options'] = {}
|
|
|
|
volume['quota'] = False
|
|
|
|
if key.lower() == 'volume id':
|
|
|
|
volume['id'] = value
|
|
|
|
if key.lower() == 'status':
|
|
|
|
volume['status'] = value
|
|
|
|
if key.lower() == 'transport-type':
|
|
|
|
volume['transport'] = value
|
|
|
|
if value.lower().endswith(' (arbiter)'):
|
|
|
|
if 'arbiters' not in volume:
|
|
|
|
volume['arbiters'] = []
|
|
|
|
value = value[:-10]
|
|
|
|
volume['arbiters'].append(value)
|
|
|
|
elif key.lower() == 'number of bricks':
|
|
|
|
volume['replicas'] = value[-1:]
|
|
|
|
if key.lower() != 'bricks' and key.lower()[:5] == 'brick':
|
|
|
|
if 'bricks' not in volume:
|
|
|
|
volume['bricks'] = []
|
|
|
|
volume['bricks'].append(value)
|
|
|
|
# Volume options
|
|
|
|
if '.' in key:
|
|
|
|
if 'options' not in volume:
|
|
|
|
volume['options'] = {}
|
|
|
|
volume['options'][key] = value
|
|
|
|
if key == 'features.quota' and value == 'on':
|
|
|
|
volume['quota'] = True
|
|
|
|
else:
|
|
|
|
if row.lower() != 'bricks:' and row.lower() != 'options reconfigured:':
|
|
|
|
if len(volume) > 0:
|
|
|
|
volumes[volume['name']] = volume
|
|
|
|
volume = {}
|
|
|
|
return volumes
|
|
|
|
|
|
|
|
|
|
|
|
def get_quotas(name, nofail):
|
|
|
|
quotas = {}
|
|
|
|
if nofail:
|
|
|
|
out = run_gluster_nofail(['volume', 'quota', name, 'list'])
|
|
|
|
if not out:
|
|
|
|
return quotas
|
|
|
|
else:
|
|
|
|
out = run_gluster(['volume', 'quota', name, 'list'])
|
|
|
|
for row in out.split('\n'):
|
|
|
|
if row[:1] == '/':
|
|
|
|
q = re.split(r'\s+', row)
|
|
|
|
quotas[q[0]] = q[1]
|
|
|
|
return quotas
|
|
|
|
|
|
|
|
|
|
|
|
def wait_for_peer(host):
|
|
|
|
for x in range(0, 4):
|
|
|
|
peers = get_peers()
|
|
|
|
if host in peers and peers[host][1].lower().find('peer in cluster') != -1:
|
|
|
|
return True
|
|
|
|
time.sleep(1)
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
|
|
|
def probe(host, myhostname):
|
|
|
|
global module
|
|
|
|
out = run_gluster(['peer', 'probe', host])
|
|
|
|
if out.find('localhost') == -1 and not wait_for_peer(host):
|
|
|
|
module.fail_json(msg='failed to probe peer %s on %s' % (host, myhostname))
|
|
|
|
|
|
|
|
|
|
|
|
def probe_all_peers(hosts, peers, myhostname):
|
|
|
|
for host in hosts:
|
|
|
|
host = host.strip() # Clean up any extra space for exact comparison
|
|
|
|
if host not in peers:
|
|
|
|
probe(host, myhostname)
|
|
|
|
|
|
|
|
|
|
|
|
def create_volume(name, stripe, replica, arbiter, disperse, redundancy, transport, hosts, bricks, force):
|
|
|
|
args = ['volume', 'create']
|
|
|
|
args.append(name)
|
|
|
|
if stripe:
|
|
|
|
args.append('stripe')
|
|
|
|
args.append(str(stripe))
|
|
|
|
if replica:
|
|
|
|
args.append('replica')
|
|
|
|
args.append(str(replica))
|
|
|
|
if arbiter:
|
|
|
|
args.append('arbiter')
|
|
|
|
args.append(str(arbiter))
|
|
|
|
if disperse:
|
|
|
|
args.append('disperse')
|
|
|
|
args.append(str(disperse))
|
|
|
|
if redundancy:
|
|
|
|
args.append('redundancy')
|
|
|
|
args.append(str(redundancy))
|
|
|
|
args.append('transport')
|
|
|
|
args.append(transport)
|
|
|
|
for brick in bricks:
|
|
|
|
for host in hosts:
|
|
|
|
args.append(('%s:%s' % (host, brick)))
|
|
|
|
if force:
|
|
|
|
args.append('force')
|
|
|
|
run_gluster(args)
|
|
|
|
|
|
|
|
|
|
|
|
def start_volume(name):
|
|
|
|
run_gluster(['volume', 'start', name])
|
|
|
|
|
|
|
|
|
|
|
|
def stop_volume(name):
|
|
|
|
run_gluster(['volume', 'stop', name])
|
|
|
|
|
|
|
|
|
|
|
|
def set_volume_option(name, option, parameter):
|
|
|
|
run_gluster(['volume', 'set', name, option, parameter])
|
|
|
|
|
|
|
|
|
|
|
|
def add_bricks(name, new_bricks, stripe, replica, force):
|
|
|
|
args = ['volume', 'add-brick', name]
|
|
|
|
if stripe:
|
|
|
|
args.append('stripe')
|
|
|
|
args.append(str(stripe))
|
|
|
|
if replica:
|
|
|
|
args.append('replica')
|
|
|
|
args.append(str(replica))
|
|
|
|
args.extend(new_bricks)
|
|
|
|
if force:
|
|
|
|
args.append('force')
|
|
|
|
run_gluster(args)
|
|
|
|
|
|
|
|
|
|
|
|
def remove_bricks(name, removed_bricks, force):
|
|
|
|
# max-tries=12 with default_interval=10 secs
|
|
|
|
max_tries = 12
|
|
|
|
retries = 0
|
|
|
|
success = False
|
|
|
|
args = ['volume', 'remove-brick', name]
|
|
|
|
args.extend(removed_bricks)
|
|
|
|
# create a copy of args to use for commit operation
|
|
|
|
args_c = args[:]
|
|
|
|
args.append('start')
|
|
|
|
run_gluster(args)
|
|
|
|
# remove-brick operation needs to be followed by commit operation.
|
|
|
|
if not force:
|
|
|
|
module.fail_json(msg="Force option is mandatory.")
|
|
|
|
else:
|
|
|
|
while retries < max_tries:
|
|
|
|
last_brick = removed_bricks[-1]
|
|
|
|
out = run_gluster(['volume', 'remove-brick', name, last_brick, 'status'])
|
|
|
|
for row in out.split('\n')[1:]:
|
|
|
|
if 'completed' in row:
|
|
|
|
# remove-brick successful, call commit operation.
|
|
|
|
args_c.append('commit')
|
|
|
|
out = run_gluster(args_c)
|
|
|
|
success = True
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
time.sleep(10)
|
|
|
|
if success:
|
|
|
|
break
|
|
|
|
retries += 1
|
|
|
|
if not success:
|
|
|
|
# remove-brick still in process, needs to be committed after completion.
|
|
|
|
module.fail_json(msg="Exceeded number of tries, check remove-brick status.\n"
|
|
|
|
"Commit operation needs to be followed.")
|
|
|
|
|
|
|
|
|
|
|
|
def reduce_config(name, removed_bricks, replicas, force):
|
|
|
|
out = run_gluster(['volume', 'heal', name, 'info'])
|
|
|
|
summary = out.split("\n")
|
|
|
|
for line in summary:
|
|
|
|
if 'Number' in line and int(line.split(":")[1].strip()) != 0:
|
|
|
|
module.fail_json(msg="Operation aborted, self-heal in progress.")
|
|
|
|
args = ['volume', 'remove-brick', name, 'replica', replicas]
|
|
|
|
args.extend(removed_bricks)
|
|
|
|
if force:
|
|
|
|
args.append('force')
|
|
|
|
else:
|
|
|
|
module.fail_json(msg="Force option is mandatory")
|
|
|
|
run_gluster(args)
|
|
|
|
|
|
|
|
|
|
|
|
def do_rebalance(name):
|
|
|
|
run_gluster(['volume', 'rebalance', name, 'start'])
|
|
|
|
|
|
|
|
|
|
|
|
def enable_quota(name):
|
|
|
|
run_gluster(['volume', 'quota', name, 'enable'])
|
|
|
|
|
|
|
|
|
|
|
|
def set_quota(name, directory, value):
|
|
|
|
run_gluster(['volume', 'quota', name, 'limit-usage', directory, value])
|
|
|
|
|
|
|
|
|
|
|
|
def main():
|
|
|
|
# MAIN
|
|
|
|
|
|
|
|
global module
|
|
|
|
module = AnsibleModule(
|
|
|
|
argument_spec=dict(
|
|
|
|
name=dict(type='str', required=True, aliases=['volume']),
|
|
|
|
state=dict(type='str', required=True, choices=['absent', 'started', 'stopped', 'present']),
|
|
|
|
cluster=dict(type='list'),
|
|
|
|
host=dict(type='str'),
|
|
|
|
stripes=dict(type='int'),
|
|
|
|
replicas=dict(type='int'),
|
|
|
|
arbiters=dict(type='int'),
|
|
|
|
disperses=dict(type='int'),
|
|
|
|
redundancies=dict(type='int'),
|
|
|
|
transport=dict(type='str', default='tcp', choices=['tcp', 'rdma', 'tcp,rdma']),
|
|
|
|
bricks=dict(type='str', aliases=['brick']),
|
|
|
|
start_on_create=dict(type='bool', default=True),
|
|
|
|
rebalance=dict(type='bool', default=False),
|
|
|
|
options=dict(type='dict', default={}),
|
|
|
|
quota=dict(type='str'),
|
|
|
|
directory=dict(type='str'),
|
|
|
|
force=dict(type='bool', default=False),
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
|
|
|
global glusterbin
|
|
|
|
glusterbin = module.get_bin_path('gluster', True)
|
|
|
|
|
|
|
|
changed = False
|
|
|
|
|
|
|
|
action = module.params['state']
|
|
|
|
volume_name = module.params['name']
|
|
|
|
cluster = module.params['cluster']
|
|
|
|
brick_paths = module.params['bricks']
|
|
|
|
stripes = module.params['stripes']
|
|
|
|
replicas = module.params['replicas']
|
|
|
|
arbiters = module.params['arbiters']
|
|
|
|
disperses = module.params['disperses']
|
|
|
|
redundancies = module.params['redundancies']
|
|
|
|
transport = module.params['transport']
|
|
|
|
myhostname = module.params['host']
|
|
|
|
start_on_create = module.boolean(module.params['start_on_create'])
|
|
|
|
rebalance = module.boolean(module.params['rebalance'])
|
|
|
|
force = module.boolean(module.params['force'])
|
|
|
|
|
|
|
|
if not myhostname:
|
|
|
|
myhostname = socket.gethostname()
|
|
|
|
|
|
|
|
# Clean up if last element is empty. Consider that yml can look like this:
|
|
|
|
# cluster="{% for host in groups['glusterfs'] %}{{ hostvars[host]['private_ip'] }},{% endfor %}"
|
|
|
|
if cluster is not None and len(cluster) > 1 and cluster[-1] == '':
|
|
|
|
cluster = cluster[0:-1]
|
|
|
|
|
|
|
|
if cluster is None:
|
|
|
|
cluster = []
|
|
|
|
|
|
|
|
if brick_paths is not None and "," in brick_paths:
|
|
|
|
brick_paths = brick_paths.split(",")
|
|
|
|
else:
|
|
|
|
brick_paths = [brick_paths]
|
|
|
|
|
|
|
|
options = module.params['options']
|
|
|
|
quota = module.params['quota']
|
|
|
|
directory = module.params['directory']
|
|
|
|
|
|
|
|
# get current state info
|
|
|
|
peers = get_peers()
|
|
|
|
volumes = get_volumes()
|
|
|
|
quotas = {}
|
|
|
|
if volume_name in volumes and volumes[volume_name]['quota'] and volumes[volume_name]['status'].lower() == 'started':
|
|
|
|
quotas = get_quotas(volume_name, True)
|
|
|
|
|
|
|
|
# do the work!
|
|
|
|
if action == 'absent':
|
|
|
|
if volume_name in volumes:
|
|
|
|
if volumes[volume_name]['status'].lower() != 'stopped':
|
|
|
|
stop_volume(volume_name)
|
|
|
|
run_gluster(['volume', 'delete', volume_name])
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
if action == 'present':
|
|
|
|
probe_all_peers(cluster, peers, myhostname)
|
|
|
|
|
|
|
|
# create if it doesn't exist
|
|
|
|
if volume_name not in volumes:
|
|
|
|
create_volume(volume_name, stripes, replicas, arbiters, disperses, redundancies, transport, cluster, brick_paths, force)
|
|
|
|
volumes = get_volumes()
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
if volume_name in volumes:
|
|
|
|
if volumes[volume_name]['status'].lower() != 'started' and start_on_create:
|
|
|
|
start_volume(volume_name)
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
# switch bricks
|
|
|
|
new_bricks = []
|
|
|
|
removed_bricks = []
|
|
|
|
all_bricks = []
|
|
|
|
bricks_in_volume = volumes[volume_name]['bricks']
|
|
|
|
|
|
|
|
for node in cluster:
|
|
|
|
for brick_path in brick_paths:
|
|
|
|
brick = '%s:%s' % (node, brick_path)
|
|
|
|
all_bricks.append(brick)
|
|
|
|
if brick not in bricks_in_volume:
|
|
|
|
new_bricks.append(brick)
|
|
|
|
|
|
|
|
if not new_bricks and len(all_bricks) > 0 and \
|
|
|
|
len(all_bricks) < len(bricks_in_volume):
|
|
|
|
for brick in bricks_in_volume:
|
|
|
|
if brick not in all_bricks:
|
|
|
|
removed_bricks.append(brick)
|
|
|
|
|
|
|
|
if new_bricks:
|
|
|
|
add_bricks(volume_name, new_bricks, stripes, replicas, force)
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
if removed_bricks:
|
|
|
|
if replicas and int(replicas) < int(volumes[volume_name]['replicas']):
|
|
|
|
reduce_config(volume_name, removed_bricks, str(replicas), force)
|
|
|
|
else:
|
|
|
|
remove_bricks(volume_name, removed_bricks, force)
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
# handle quotas
|
|
|
|
if quota:
|
|
|
|
if not volumes[volume_name]['quota']:
|
|
|
|
enable_quota(volume_name)
|
|
|
|
quotas = get_quotas(volume_name, False)
|
|
|
|
if directory not in quotas or quotas[directory] != quota:
|
|
|
|
set_quota(volume_name, directory, quota)
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
# set options
|
|
|
|
for option in options.keys():
|
|
|
|
if option not in volumes[volume_name]['options'] or volumes[volume_name]['options'][option] != options[option]:
|
|
|
|
set_volume_option(volume_name, option, options[option])
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
else:
|
|
|
|
module.fail_json(msg='failed to create volume %s' % volume_name)
|
|
|
|
|
|
|
|
if action != 'absent' and volume_name not in volumes:
|
|
|
|
module.fail_json(msg='volume not found %s' % volume_name)
|
|
|
|
|
|
|
|
if action == 'started':
|
|
|
|
if volumes[volume_name]['status'].lower() != 'started':
|
|
|
|
start_volume(volume_name)
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
if action == 'stopped':
|
|
|
|
if volumes[volume_name]['status'].lower() != 'stopped':
|
|
|
|
stop_volume(volume_name)
|
|
|
|
changed = True
|
|
|
|
|
|
|
|
if changed:
|
|
|
|
volumes = get_volumes()
|
|
|
|
if rebalance:
|
|
|
|
do_rebalance(volume_name)
|
|
|
|
|
|
|
|
facts = {}
|
|
|
|
facts['glusterfs'] = {'peers': peers, 'volumes': volumes, 'quotas': quotas}
|
|
|
|
|
|
|
|
module.exit_json(changed=changed, ansible_facts=facts)
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == '__main__':
|
|
|
|
main()
|