Compare commits
77 commits
stable-202
...
main
Author | SHA1 | Date | |
---|---|---|---|
1a41a66355 | |||
da7be10d58 | |||
44e398a418 | |||
d903f4f90d | |||
ad13e65250 | |||
b63229092c | |||
4b3c189481 | |||
8d57b773a0 | |||
3c815b9a38 | |||
55cc71871f | |||
46f1330b5c | |||
d774349950 | |||
249fe90c54 | |||
87b6ea8ddb | |||
dc39cb5d85 | |||
beff4242ac | |||
cc31a2f038 | |||
6a63031f84 | |||
1c3eecdd85 | |||
16c4f3650c | |||
57b90cced4 | |||
4e67d6f1b8 | |||
9bb2afa68b | |||
26ea8e8e8d | |||
5c42f9d7d2 | |||
f87fac3c3b | |||
582d88d2f1 | |||
4f84f71070 | |||
4caed83e80 | |||
38174166a0 | |||
3b42817c57 | |||
5950751c73 | |||
d010d27f20 | |||
d5c3486425 | |||
1aea2f5674 | |||
0b3a2d43cb | |||
4b952bac40 | |||
e150fa3625 | |||
3579682de8 | |||
8f090bfcac | |||
0e189ad0ef | |||
e27cf34987 | |||
17d8d8b2de | |||
3fbfe26c66 | |||
4a6d35b984 | |||
ef1eb5b4ad | |||
dc5b61c9e6 | |||
33389e842a | |||
bb7ab09db8 | |||
bc481d44d5 | |||
5619cb8270 | |||
fbe409efff | |||
9096bf7c33 | |||
1c38cfbf3c | |||
d332a810b9 | |||
74da008b5e | |||
a6eaddbd8f | |||
0da62f097d | |||
ed3d4b1b9c | |||
bb44e6da8f | |||
de3d18b0a1 | |||
aa6be0ce34 | |||
83fcc67c91 | |||
ec4e2bdb38 | |||
a68fea86b8 | |||
a1b7a03908 | |||
7363284353 | |||
9f0eccdb3d | |||
be9f1a7152 | |||
f6cd170e6f | |||
dcb0ca8e48 | |||
e0f427fcc8 | |||
7090f7ca5e | |||
705479f29c | |||
3007fc6de4 | |||
4a270b9f65 | |||
b2942339d1 |
18 changed files with 185 additions and 56 deletions
6
files/scriptreciver/sysctl-d-gofasta.conf
Normal file
6
files/scriptreciver/sysctl-d-gofasta.conf
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
net.core.rmem_max=67108864
|
||||||
|
net.core.wmem_max=67108864
|
||||||
|
net.ipv4.tcp_rmem=4096 87380 33554432
|
||||||
|
net.ipv4.tcp_wmem=4096 87380 33554432
|
||||||
|
net.core.default_qdisc=fq
|
||||||
|
net.ipv4.tcp_congestion_control=bbr
|
|
@ -101,7 +101,24 @@ define sunetdrive::app_type (
|
||||||
content => template('sunetdrive/application/get_containers'),
|
content => template('sunetdrive/application/get_containers'),
|
||||||
mode => '0744',
|
mode => '0744',
|
||||||
}
|
}
|
||||||
if ($environment == 'test' and ($nodenumber == 3)) {
|
if ($nodenumber == 3) {
|
||||||
|
file { '/usr/lib/nagios/plugins/check_nextcloud_mounts.py':
|
||||||
|
ensure => present,
|
||||||
|
owner => 'root',
|
||||||
|
group => 'root',
|
||||||
|
content => template('sunetdrive/application/check_nextcloud_mounts.py'),
|
||||||
|
mode => '0744',
|
||||||
|
}
|
||||||
|
sunet::sudoer {'nagios_run_nextcloud_mounts_command':
|
||||||
|
user_name => 'nagios',
|
||||||
|
collection => 'nrpe_nextcloud_mounts_check',
|
||||||
|
command_line => '/usr/lib/nagios/plugins/check_nextcloud_mounts.py'
|
||||||
|
}
|
||||||
|
sunet::nagios::nrpe_command {'check_nextcloud_mounts':
|
||||||
|
command_line => '/usr/bin/sudo /usr/lib/nagios/plugins/check_nextcloud_mounts.py'
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if ($nodenumber == 3) {
|
||||||
file { '/usr/local/bin/scan_external_mounts':
|
file { '/usr/local/bin/scan_external_mounts':
|
||||||
ensure => present,
|
ensure => present,
|
||||||
force => true,
|
force => true,
|
||||||
|
@ -117,10 +134,6 @@ define sunetdrive::app_type (
|
||||||
ok_criteria => ['exit_status=0','max_age=2d'],
|
ok_criteria => ['exit_status=0','max_age=2d'],
|
||||||
warn_criteria => ['exit_status=1','max_age=3d'],
|
warn_criteria => ['exit_status=1','max_age=3d'],
|
||||||
}
|
}
|
||||||
cron { 'scan_external_mounts':
|
|
||||||
ensure => absent,
|
|
||||||
command => 'true',
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
file { '/opt/nextcloud/cron.sh':
|
file { '/opt/nextcloud/cron.sh':
|
||||||
ensure => file,
|
ensure => file,
|
||||||
|
@ -392,5 +405,4 @@ define sunetdrive::app_type (
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -15,6 +15,7 @@ define sunetdrive::db_type(
|
||||||
$backup_password = safe_hiera('backup_password')
|
$backup_password = safe_hiera('backup_password')
|
||||||
$proxysql_password = safe_hiera('proxysql_password')
|
$proxysql_password = safe_hiera('proxysql_password')
|
||||||
$mysql_user_password = safe_hiera('mysql_user_password')
|
$mysql_user_password = safe_hiera('mysql_user_password')
|
||||||
|
$roundcube_password = safe_hiera('roundcube_password')
|
||||||
$mariadb_dir = '/etc/mariadb'
|
$mariadb_dir = '/etc/mariadb'
|
||||||
$mycnf_path = 'sunetdrive/mariadb/my.cnf.erb'
|
$mycnf_path = 'sunetdrive/mariadb/my.cnf.erb'
|
||||||
$server_id = 1000 + Integer($facts['networking']['hostname'][-1])
|
$server_id = 1000 + Integer($facts['networking']['hostname'][-1])
|
||||||
|
@ -33,8 +34,10 @@ define sunetdrive::db_type(
|
||||||
$ports = [3306, 4444, 4567, 4568]
|
$ports = [3306, 4444, 4567, 4568]
|
||||||
if $location =~ /^multinode/ {
|
if $location =~ /^multinode/ {
|
||||||
$from = $db_ip + $nextcloud_ip + $backup_ip + $backup_ipv6 + $db_ipv6 + $config['kube'] + $config['kube_v6']
|
$from = $db_ip + $nextcloud_ip + $backup_ip + $backup_ipv6 + $db_ipv6 + $config['kube'] + $config['kube_v6']
|
||||||
} elsif $location == 'sunet-test' or $location == 'sunet-prod' {
|
} elsif $location == 'sunet-prod' {
|
||||||
$from = $db_ip + $nextcloud_ip + $backup_ip + $backup_ipv6 + $db_ipv6 + $config['imap'] + $config['imap_v6'] + $config['smtp'] + $config['smtp_v6']
|
$from = $db_ip + $nextcloud_ip + $backup_ip + $backup_ipv6 + $db_ipv6 + $config['imap'] + $config['imap_v6'] + $config['smtp'] + $config['smtp_v6'] + $config['webmail'] + $config['webmail_v6']
|
||||||
|
} elsif $location == 'sunet-test' {
|
||||||
|
$from = $db_ip + $nextcloud_ip + $backup_ip + $backup_ipv6 + $db_ipv6 + $config['imap'] + $config['imap_v6'] + $config['smtp'] + $config['smtp_v6'] + $config['webmail'] + $config['webmail_v6'] + $config['calendar'] + $config['calendar_v6']
|
||||||
} else {
|
} else {
|
||||||
$from = $db_ip + $nextcloud_ip + $backup_ip + $backup_ipv6 + $db_ipv6
|
$from = $db_ip + $nextcloud_ip + $backup_ip + $backup_ipv6 + $db_ipv6
|
||||||
}
|
}
|
||||||
|
@ -46,7 +49,7 @@ define sunetdrive::db_type(
|
||||||
sunet::system_user {'mysql': username => 'mysql', group => 'mysql' }
|
sunet::system_user {'mysql': username => 'mysql', group => 'mysql' }
|
||||||
|
|
||||||
|
|
||||||
$sql_files = ['02-backup_user.sql', '03-proxysql.sql', '04-nextcloud.sql']
|
$sql_files = ['02-backup_user.sql', '03-proxysql.sql', '04-nextcloud.sql', '05-roundcube.sql']
|
||||||
$sql_files.each |$sql_file|{
|
$sql_files.each |$sql_file|{
|
||||||
file { "${mariadb_dir}/init/${sql_file}":
|
file { "${mariadb_dir}/init/${sql_file}":
|
||||||
ensure => present,
|
ensure => present,
|
||||||
|
|
|
@ -46,7 +46,21 @@ class sunetdrive::multinode (
|
||||||
content => template('sunetdrive/application/get_containers'),
|
content => template('sunetdrive/application/get_containers'),
|
||||||
mode => '0744',
|
mode => '0744',
|
||||||
}
|
}
|
||||||
if ($environment == 'test') {
|
file { '/usr/lib/nagios/plugins/check_nextcloud_mounts.py':
|
||||||
|
ensure => present,
|
||||||
|
owner => 'root',
|
||||||
|
group => 'root',
|
||||||
|
content => template('sunetdrive/application/check_nextcloud_mounts.py'),
|
||||||
|
mode => '0744',
|
||||||
|
}
|
||||||
|
sunet::sudoer {'nagios_run_nextcloud_mounts_command':
|
||||||
|
user_name => 'nagios',
|
||||||
|
collection => 'nrpe_nextcloud_mounts_check',
|
||||||
|
command_line => '/usr/lib/nagios/plugins/check_nextcloud_mounts.py'
|
||||||
|
}
|
||||||
|
sunet::nagios::nrpe_command {'check_nextcloud_mounts':
|
||||||
|
command_line => '/usr/bin/sudo /usr/lib/nagios/plugins/check_nextcloud_mounts.py'
|
||||||
|
}
|
||||||
file { '/usr/local/bin/scan_external_mounts':
|
file { '/usr/local/bin/scan_external_mounts':
|
||||||
ensure => present,
|
ensure => present,
|
||||||
force => true,
|
force => true,
|
||||||
|
@ -62,11 +76,6 @@ class sunetdrive::multinode (
|
||||||
ok_criteria => ['exit_status=0','max_age=2d'],
|
ok_criteria => ['exit_status=0','max_age=2d'],
|
||||||
warn_criteria => ['exit_status=1','max_age=3d'],
|
warn_criteria => ['exit_status=1','max_age=3d'],
|
||||||
}
|
}
|
||||||
cron { 'scan_external_mounts':
|
|
||||||
ensure => absent,
|
|
||||||
command => 'true',
|
|
||||||
}
|
|
||||||
}
|
|
||||||
file { '/usr/local/bin/nocc':
|
file { '/usr/local/bin/nocc':
|
||||||
ensure => present,
|
ensure => present,
|
||||||
force => true,
|
force => true,
|
||||||
|
|
|
@ -41,7 +41,10 @@ class sunetdrive::multinode_db(){
|
||||||
group => 'root',
|
group => 'root',
|
||||||
mode => '0600',
|
mode => '0600',
|
||||||
}
|
}
|
||||||
file { '/root/tasks/listusersbydep.sh':
|
file { '/root/tasks/':
|
||||||
|
ensure => directory,
|
||||||
|
}
|
||||||
|
-> file { '/root/tasks/listusersbydep.sh':
|
||||||
ensure => file,
|
ensure => file,
|
||||||
content => template('sunetdrive/mariadb/listusersdep.sh.erb'),
|
content => template('sunetdrive/mariadb/listusersdep.sh.erb'),
|
||||||
owner => 'root',
|
owner => 'root',
|
||||||
|
@ -55,7 +58,10 @@ class sunetdrive::multinode_db(){
|
||||||
group => 'root',
|
group => 'root',
|
||||||
mode => '0700',
|
mode => '0700',
|
||||||
}
|
}
|
||||||
file {'/opt/mariadb/statistics/custdata.json':
|
file {'/opt/mariadb/statistics/':
|
||||||
|
ensure => directory,
|
||||||
|
}
|
||||||
|
-> file {'/opt/mariadb/statistics/custdata.json':
|
||||||
ensure => file,
|
ensure => file,
|
||||||
content => template('sunetdrive/mariadb/custconfig.json.erb'),
|
content => template('sunetdrive/mariadb/custconfig.json.erb'),
|
||||||
owner => 'root',
|
owner => 'root',
|
||||||
|
|
|
@ -5,7 +5,19 @@ class sunetdrive::scriptreceiver()
|
||||||
sunet::system_user {'script': username => 'script', group => 'script', managehome => true, shell => '/bin/bash' }
|
sunet::system_user {'script': username => 'script', group => 'script', managehome => true, shell => '/bin/bash' }
|
||||||
|
|
||||||
# These tasks correspond to a ${task}.erb.sh template
|
# These tasks correspond to a ${task}.erb.sh template
|
||||||
$tasks = ['list_users', 'list_files_for_user', 'create_bucket', 'backup_db', 'purge_backups', 'maintenancemode', 'restart_sunet_service', 'start_sentinel', 'stop_sentinel', 'removeswap', 'backup_multinode_db']
|
$tasks = [
|
||||||
|
'list_users',
|
||||||
|
'list_files_for_user',
|
||||||
|
'create_bucket',
|
||||||
|
'backup_db',
|
||||||
|
'purge_backups',
|
||||||
|
'maintenancemode',
|
||||||
|
'restart_sunet_service',
|
||||||
|
'start_sentinel',
|
||||||
|
'stop_sentinel',
|
||||||
|
'removeswap',
|
||||||
|
'backup_multinode_db'
|
||||||
|
]
|
||||||
|
|
||||||
$environment = sunetdrive::get_environment()
|
$environment = sunetdrive::get_environment()
|
||||||
$config = hiera_hash($environment)
|
$config = hiera_hash($environment)
|
||||||
|
@ -35,7 +47,9 @@ class sunetdrive::scriptreceiver()
|
||||||
type => 'ssh-ed25519',
|
type => 'ssh-ed25519',
|
||||||
key => $script_pub_key,
|
key => $script_pub_key,
|
||||||
}
|
}
|
||||||
|
file { '/etc/sysctl.d/gofasta.conf':
|
||||||
|
ensure => 'absent',
|
||||||
|
}
|
||||||
file { '/opt/rotate':
|
file { '/opt/rotate':
|
||||||
ensure => directory,
|
ensure => directory,
|
||||||
mode => '0750',
|
mode => '0750',
|
||||||
|
|
|
@ -11,6 +11,9 @@ class sunetdrive::sitemonitornaemon() {
|
||||||
$environment = sunetdrive::get_environment()
|
$environment = sunetdrive::get_environment()
|
||||||
$influx_passwd = safe_hiera('influx_passwd')
|
$influx_passwd = safe_hiera('influx_passwd')
|
||||||
$slack_url = safe_hiera('slack_url')
|
$slack_url = safe_hiera('slack_url')
|
||||||
|
$extra_host_groups = {
|
||||||
|
node3_hosts => join($facts['configured_hosts_in_cosmos']['all'].filter |$host| { $host =~ /^node3\./ }, ',')
|
||||||
|
}
|
||||||
|
|
||||||
file { '/usr/local/bin/slack_nagios.sh':
|
file { '/usr/local/bin/slack_nagios.sh':
|
||||||
ensure => present,
|
ensure => present,
|
||||||
|
@ -45,6 +48,11 @@ class sunetdrive::sitemonitornaemon() {
|
||||||
content => template('sunetdrive/monitor/sunetdrive_thruk_templates.conf.erb'),
|
content => template('sunetdrive/monitor/sunetdrive_thruk_templates.conf.erb'),
|
||||||
mode => '0644',
|
mode => '0644',
|
||||||
}
|
}
|
||||||
|
file { '/etc/naemon/conf.d/sunetdrive_extra_hostgroups.cfg':
|
||||||
|
ensure => present,
|
||||||
|
content => template('sunetdrive/monitor/sunetdrive_extra_hostgroups.cfg.erb'),
|
||||||
|
mode => '0644',
|
||||||
|
}
|
||||||
nagioscfg::service {'check_scriptherder':
|
nagioscfg::service {'check_scriptherder':
|
||||||
hostgroup_name => ['sunetdrive::nrpe'],
|
hostgroup_name => ['sunetdrive::nrpe'],
|
||||||
check_command => 'check_nrpe_1arg_to300!check_scriptherder',
|
check_command => 'check_nrpe_1arg_to300!check_scriptherder',
|
||||||
|
@ -99,6 +107,12 @@ class sunetdrive::sitemonitornaemon() {
|
||||||
description => 'Status of sarimner interface',
|
description => 'Status of sarimner interface',
|
||||||
contact_groups => ['alerts']
|
contact_groups => ['alerts']
|
||||||
}
|
}
|
||||||
|
nagioscfg::service {'check_nextcloud_mounts':
|
||||||
|
hostgroup_name => ['node3_hosts','sunetdrive::multinode'],
|
||||||
|
check_command => 'check_nrpe_1arg!check_nextcloud_mounts',
|
||||||
|
description => 'S3 buckets with multiple Nextcloud mounts',
|
||||||
|
contact_groups => ['alerts']
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
42
templates/application/check_nextcloud_mounts.py
Normal file
42
templates/application/check_nextcloud_mounts.py
Normal file
|
@ -0,0 +1,42 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
from collections import Counter
|
||||||
|
import json
|
||||||
|
import shlex
|
||||||
|
import subprocess
|
||||||
|
import sys
|
||||||
|
|
||||||
|
exit = 0
|
||||||
|
base_message = "OK: no duplicate mounts"
|
||||||
|
long_message = ""
|
||||||
|
|
||||||
|
get_containers = subprocess.Popen('/usr/local/bin/get_containers', stdout=subprocess.PIPE).stdout.read()
|
||||||
|
containers = get_containers.decode().splitlines()
|
||||||
|
|
||||||
|
for i, container in enumerate(containers, start=1):
|
||||||
|
buckets = []
|
||||||
|
list_command = f"/usr/local/bin/nocc {container} files_external:list --all --show-password --output json"
|
||||||
|
command = shlex.split(list_command)
|
||||||
|
mount_data_byte = subprocess.Popen(command, stdout=subprocess.PIPE).stdout.read()
|
||||||
|
try:
|
||||||
|
mount_data = json.loads(mount_data_byte.decode())
|
||||||
|
except json.decoder.JSONDecodeError as err:
|
||||||
|
if i == 1 or i != len(containers):
|
||||||
|
base_message = "WARNING: invalid json"
|
||||||
|
long_message += f"\ncontainer: {container} - json decode error: {err}"
|
||||||
|
# lets do exit 0 for now
|
||||||
|
# exit = 1
|
||||||
|
continue
|
||||||
|
for items in mount_data:
|
||||||
|
buckets.append(items["configuration"]["bucket"])
|
||||||
|
bucket_count = dict(Counter(buckets))
|
||||||
|
for k, v in bucket_count.items():
|
||||||
|
if v > 1:
|
||||||
|
base_message = "WARNING: buckets with multiple mounts"
|
||||||
|
long_message += f"\ncontainer: {container} - bucket: {k} - {v}"
|
||||||
|
# lets do exit 0 for now
|
||||||
|
# exit = 1
|
||||||
|
print(base_message)
|
||||||
|
if long_message != "":
|
||||||
|
print(long_message.lstrip())
|
||||||
|
sys.exit(exit)
|
|
@ -16,6 +16,7 @@ $CONFIG = array (
|
||||||
),
|
),
|
||||||
),
|
),
|
||||||
'appstoreenabled' => false,
|
'appstoreenabled' => false,
|
||||||
|
'auth.bruteforce.protection.enabled' => false,
|
||||||
'config_is_read_only' => true,
|
'config_is_read_only' => true,
|
||||||
'csrf.disabled' => true,
|
'csrf.disabled' => true,
|
||||||
'datadirectory' => '/var/www/html/data',
|
'datadirectory' => '/var/www/html/data',
|
||||||
|
@ -76,7 +77,7 @@ $CONFIG = array (
|
||||||
'region' => 'us-east-1',
|
'region' => 'us-east-1',
|
||||||
'hostname' => '<%= @s3_host %>',
|
'hostname' => '<%= @s3_host %>',
|
||||||
'port' => '',
|
'port' => '',
|
||||||
'useMultipartCopy' => false,
|
'useMultipartCopy' => true,
|
||||||
'objectPrefix' => 'urn:oid:',
|
'objectPrefix' => 'urn:oid:',
|
||||||
'autocreate' => false,
|
'autocreate' => false,
|
||||||
'use_ssl' => true,
|
'use_ssl' => true,
|
||||||
|
|
|
@ -15,7 +15,7 @@ if [[ "x${oc_list}" != "x" ]]; then
|
||||||
done
|
done
|
||||||
fi
|
fi
|
||||||
|
|
||||||
docker exec -ti ${MY_VARS} -u www-data ${container} php --define apc.enable_cli=1 /var/www/html/occ "$@"
|
docker exec -i ${MY_VARS} -u www-data ${container} php --define apc.enable_cli=1 /var/www/html/occ "$@"
|
||||||
exit 0
|
exit 0
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -52,7 +52,7 @@ echo '
|
||||||
"region": "'${region}'",
|
"region": "'${region}'",
|
||||||
"secret": "'${secret}'",
|
"secret": "'${secret}'",
|
||||||
"storageClass": "",
|
"storageClass": "",
|
||||||
"useMultipartCopy": false,
|
"useMultipartCopy": true,
|
||||||
"use_path_style": true,
|
"use_path_style": true,
|
||||||
"use_ssl": true
|
"use_ssl": true
|
||||||
},
|
},
|
||||||
|
|
|
@ -3,23 +3,32 @@
|
||||||
error_ids=""
|
error_ids=""
|
||||||
# Only run if this is the only instance of this script running
|
# Only run if this is the only instance of this script running
|
||||||
# note: since this script forks to run pgrep, we need -eq 2 here
|
# note: since this script forks to run pgrep, we need -eq 2 here
|
||||||
if [[ $(pgrep -f "${0}" | wc -l) -eq 2 ]]; then
|
# shellcheck disable=SC2126
|
||||||
|
if [[ $(pgrep -a -f "${0}" | grep -v scriptherder | wc -l) -eq 2 ]]; then
|
||||||
# We sleep a deterministic amount of time, which will be between 0 an 128 m and allways the same within
|
# We sleep a deterministic amount of time, which will be between 0 an 128 m and allways the same within
|
||||||
# a specific host, but will differ between hosts
|
# a specific host, but will differ between hosts
|
||||||
sleep $((16#$(ip a | grep "link/ether" | head -1 | awk -F ':' '{print $6}' | awk '{print $1}') / 2))m
|
sleep $((16#$(ip a | grep "link/ether" | head -1 | awk -F ':' '{print $6}' | awk '{print $1}') / 2))m
|
||||||
for container in $(get_containers); do
|
errors=''
|
||||||
for id in $(nocc "${container}" files_external:list --all --output json | jq '.[].mount_id' | jq .); do
|
for container in $(/usr/local/bin/get_containers); do
|
||||||
nocc "${container}" files_external:scan "${id}" | grep Error
|
error_ids="${error_ids} ${container}: "
|
||||||
|
for id in $(/usr/local/bin/nocc "${container}" files_external:list --all --output json | jq '.[].mount_id' | jq .); do
|
||||||
|
/usr/local/bin/nocc "${container}" files_external:scan "${id}" | grep Error
|
||||||
# shellcheck disable=SC2181
|
# shellcheck disable=SC2181
|
||||||
if [[ ${?} -eq 0 ]]; then
|
if [[ ${?} -eq 0 ]]; then
|
||||||
error_ids="${error_ids} ${container}: ${id}"
|
errors="${errors} ${id}"
|
||||||
|
error_ids="${error_ids} ${id}"
|
||||||
fi
|
fi
|
||||||
done
|
done
|
||||||
done
|
done
|
||||||
|
else
|
||||||
|
echo "Another instance of this script is already running, exiting"
|
||||||
|
pgrep -a -f "${0}" | grep -v scriptherder
|
||||||
|
exit 0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ -n "${error_ids}" ]]; then
|
if [[ -n "${errors}" ]]; then
|
||||||
echo "Errors found in the following mounts: ${error_ids}"
|
echo "Errors found in the following mounts: ${error_ids}"
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
echo "No errors found"
|
||||||
exit 0
|
exit 0
|
||||||
|
|
3
templates/mariadb/05-roundcube.sql.erb
Normal file
3
templates/mariadb/05-roundcube.sql.erb
Normal file
|
@ -0,0 +1,3 @@
|
||||||
|
CREATE SCHEMA roundcubemail;
|
||||||
|
CREATE USER 'roundcube'@'%' IDENTIFIED BY '<%= @roundcube_password %>';
|
||||||
|
GRANT ALL PRIVILEGES ON roundcubemail.* TO 'roundcube'@'%' IDENTIFIED BY '<%= @roundcube_password %>';
|
8
templates/monitor/sunetdrive_extra_hostgroups.cfg.erb
Normal file
8
templates/monitor/sunetdrive_extra_hostgroups.cfg.erb
Normal file
|
@ -0,0 +1,8 @@
|
||||||
|
<% @extra_host_groups.each do |group, members| -%>
|
||||||
|
# <%= group %>
|
||||||
|
define hostgroup {
|
||||||
|
hostgroup_name <%= group %>
|
||||||
|
alias <%= group %>
|
||||||
|
members <%= members %>
|
||||||
|
}
|
||||||
|
<% end -%>
|
|
@ -4,7 +4,7 @@ customer="${1}"
|
||||||
multinode="${2}"
|
multinode="${2}"
|
||||||
environment="<%= @environment %>"
|
environment="<%= @environment %>"
|
||||||
location="${customer}-${environment}"
|
location="${customer}-${environment}"
|
||||||
userjson=$(ssh "script@${multinode}" "sudo /home/script/bin/list_users.sh nextcloud-${customer}-app-1")
|
userjson=$(ssh -o StrictHostKeyChecking=no "script@${multinode}" "sudo /home/script/bin/list_users.sh nextcloud-${customer}-app-1")
|
||||||
project="statistics"
|
project="statistics"
|
||||||
bucket="drive-server-coms"
|
bucket="drive-server-coms"
|
||||||
base_dir="${project}:${bucket}"
|
base_dir="${project}:${bucket}"
|
||||||
|
|
|
@ -43,9 +43,11 @@ for eppn in $(echo "${users}" | jq -r keys[]); do
|
||||||
username=${eppn%@*}
|
username=${eppn%@*}
|
||||||
# Remove underscore from username
|
# Remove underscore from username
|
||||||
user=${username//_/-}
|
user=${username//_/-}
|
||||||
|
# convert user to lower case for bucket naming rules
|
||||||
|
user_lower=${user,,}
|
||||||
|
|
||||||
echo "$(date) - Check bucket status for ${eppn}"
|
echo "$(date) - Check bucket status for ${eppn}"
|
||||||
bucketname="${user}-${site_name//./-}"
|
bucketname="${user_lower}-${site_name//./-}"
|
||||||
if ! echo "${buckets}" | grep "${bucketname}" &> /dev/null; then
|
if ! echo "${buckets}" | grep "${bucketname}" &> /dev/null; then
|
||||||
echo "$(date) - ${eppn} has no mounts configured, adding bucket and mounts..."
|
echo "$(date) - ${eppn} has no mounts configured, adding bucket and mounts..."
|
||||||
${rclone} mkdir "${rcp}:${bucketname}"
|
${rclone} mkdir "${rcp}:${bucketname}"
|
||||||
|
|
|
@ -33,12 +33,12 @@ fi
|
||||||
key=$(grep access_key_id "${rclone_config}" | awk '{print $3}')
|
key=$(grep access_key_id "${rclone_config}" | awk '{print $3}')
|
||||||
secret=$(grep secret_access_key "${rclone_config}"| awk '{print $3}')
|
secret=$(grep secret_access_key "${rclone_config}"| awk '{print $3}')
|
||||||
endpoint=$(grep endpoint "${rclone_config}" | awk '{print $3}')
|
endpoint=$(grep endpoint "${rclone_config}" | awk '{print $3}')
|
||||||
preexisting="$(docker exec -u www-data -i "${container}" php --define apc.enable_cli=1 /var/www/html/occ files_external:list --output json "${user}" | jq -r '.[] | .configuration.bucket' | grep "${bucket}")"
|
preexisting="$(docker exec -u www-data -i "${container}" php --define apc.enable_cli=1 /var/www/html/occ files_external:list --output json --show-password "${user}" | jq -r '.[] | .configuration.bucket' | grep "${bucket}")"
|
||||||
|
|
||||||
if [[ -z ${preexisting} ]]; then
|
if [[ -z ${preexisting} ]]; then
|
||||||
docker exec -u www-data -i "${container}" php --define apc.enable_cli=1 /var/www/html/occ files_external:create "${user_bucket_name}" \
|
docker exec -u www-data -i "${container}" php --define apc.enable_cli=1 /var/www/html/occ files_external:create "${user_bucket_name}" \
|
||||||
amazons3 -c bucket="${bucket}" -c key="${key}" -c secret="${secret}" -c hostname="${endpoint}" -c use_ssl=true -c use_path_style=true -c region=us-east-1 \
|
amazons3 -c bucket="${bucket}" -c key="${key}" -c secret="${secret}" -c hostname="${endpoint}" -c use_ssl=true -c use_path_style=true -c region=us-east-1 \
|
||||||
-c useMultipartCopy=false amazons3::accesskey --user ${user}
|
-c useMultipartCopy=true amazons3::accesskey --user ${user}
|
||||||
for shareid in $(docker exec -u www-data -i ${container} php --define apc.enable_cli=1 /var/www/html/occ files_external:export ${user} | jq -r '.[].mount_id'); do
|
for shareid in $(docker exec -u www-data -i ${container} php --define apc.enable_cli=1 /var/www/html/occ files_external:export ${user} | jq -r '.[].mount_id'); do
|
||||||
docker exec -u www-data -i ${container} php --define apc.enable_cli=1 /var/www/html/occ files_external:option ${shareid} enable_sharing true
|
docker exec -u www-data -i ${container} php --define apc.enable_cli=1 /var/www/html/occ files_external:option ${shareid} enable_sharing true
|
||||||
done
|
done
|
||||||
|
|
|
@ -7,7 +7,7 @@ bucket=${4}
|
||||||
user=${5}
|
user=${5}
|
||||||
/usr/local/bin/occ files_external:create "${bucket}" \
|
/usr/local/bin/occ files_external:create "${bucket}" \
|
||||||
amazons3 -c bucket="${bucket}" -c key="${key}" -c secret="${secret}" -c hostname="${endpoint}" -c use_ssl=true -c use_path_style=true -c region=us-east-1 \
|
amazons3 -c bucket="${bucket}" -c key="${key}" -c secret="${secret}" -c hostname="${endpoint}" -c use_ssl=true -c use_path_style=true -c region=us-east-1 \
|
||||||
-c useMultipartCopy=false amazons3::accesskey --user "${user}"
|
-c useMultipartCopy=true amazons3::accesskey --user "${user}"
|
||||||
for shareid in $(/usr/local/bin/occ files_external:export "${user}" | jq -r '.[].mount_id'); do
|
for shareid in $(/usr/local/bin/nocc files_external:export "${user}" | jq -r '.[].mount_id'); do
|
||||||
/usr/local/bin/occ files_external:option "${shareid}" enable_sharing true
|
/usr/local/bin/nocc files_external:option "${shareid}" enable_sharing true
|
||||||
done
|
done
|
||||||
|
|
Loading…
Add table
Reference in a new issue