Compare commits
64 commits
Author | SHA1 | Date | |
---|---|---|---|
8e6dd476bf | |||
55b0140edd | |||
2edf4d2e86 | |||
1ce00d81cb | |||
6d8219b22c | |||
96132ae60a | |||
db86a3d02c | |||
daf4833691 | |||
cebb8f970b | |||
d0aaf99890 | |||
833f6e8d4d | |||
75bb6b0765 | |||
54ff8f23ed | |||
63931b8880 | |||
186c886e84 | |||
5e735cf453 | |||
7ebcff86cb | |||
0705409393 | |||
84315c7bae | |||
f0d3ca8f43 | |||
88fc83014b | |||
120af2ff01 | |||
44eb0031de | |||
bfa697f457 | |||
0b0503e109 | |||
7f745b3609 | |||
4bacf38188 | |||
79a549cd4c | |||
7b74984fb7 | |||
b1098a966b | |||
5d969ada38 | |||
bb35d6cf3e | |||
afd28627f3 | |||
124c6b56d2 | |||
23bacdb9c6 | |||
11029ea231 | |||
212699f571 | |||
6e38f1c4d1 | |||
f79e486371 | |||
5bec5dffed | |||
e1a447da7d | |||
|
14a2cc65f9 | ||
|
8895c3ee9e | ||
|
de48994d76 | ||
|
34d277d80f | ||
|
74878cae0f | ||
|
b9be6491cc | ||
|
30a605a81c | ||
|
e3bca44f3a | ||
|
231125be0c | ||
|
6b65f80919 | ||
|
cb1ab19099 | ||
|
a400e81fbe | ||
|
382a5322a6 | ||
|
23adccacb3 | ||
|
76a99dc532 | ||
|
127dba8c52 | ||
c8430f440e | |||
|
02e96812ae | ||
9d12af51c0 | |||
c37b096838 | |||
b2fcdac493 | |||
|
9025fff7e1 | ||
|
7c26744cbd |
61
README.md
|
@ -1,12 +1,14 @@
|
||||||
|
![Logo Risotto](logo.png "logo risotto")
|
||||||
|
|
||||||
# Risotto
|
# Risotto
|
||||||
|
|
||||||
## Installation
|
## Install dependencies
|
||||||
|
|
||||||
Install dependencies:
|
|
||||||
|
|
||||||
- ldns-utils
|
|
||||||
- python3-cheetah
|
|
||||||
- python3-dkimpy
|
- python3-dkimpy
|
||||||
|
- python3-cheetah
|
||||||
|
- ldns-utils
|
||||||
|
|
||||||
|
## Installation
|
||||||
|
|
||||||
Clone projects:
|
Clone projects:
|
||||||
|
|
||||||
|
@ -15,6 +17,10 @@ Clone projects:
|
||||||
- https://cloud.silique.fr/gitea/risotto/rougail
|
- https://cloud.silique.fr/gitea/risotto/rougail
|
||||||
- https://cloud.silique.fr/gitea/risotto/risotto
|
- https://cloud.silique.fr/gitea/risotto/risotto
|
||||||
|
|
||||||
|
## Documentation
|
||||||
|
|
||||||
|
[Documentation](doc/README.md)
|
||||||
|
|
||||||
## Set up
|
## Set up
|
||||||
|
|
||||||
Set up Risotto:
|
Set up Risotto:
|
||||||
|
@ -29,46 +35,19 @@ In risotto.conf change the dataset directory.
|
||||||
Set up infrasctructure:
|
Set up infrasctructure:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
cp server.json.example server.json
|
cp server.yml.example server.yml
|
||||||
```
|
```
|
||||||
|
|
||||||
Change the configuration.
|
Modify infrastructure description as required.
|
||||||
|
|
||||||
|
Generate the configuration:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
./bootstrap.py
|
||||||
|
```
|
||||||
|
|
||||||
Send configuration to remote server:
|
Send configuration to remote server:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
HOST=cloud.silique.fr
|
ansible-playbook -i ansible/inventory.py ansible/playbook.yml
|
||||||
./test.py
|
|
||||||
rm -f installations.tar
|
|
||||||
tar -cf installations.tar installations
|
|
||||||
scp installations.tar root@$HOST:
|
|
||||||
```
|
|
||||||
|
|
||||||
## Deploy
|
|
||||||
|
|
||||||
In host:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
cd
|
|
||||||
rm -rf installations
|
|
||||||
tar xf installations.tar
|
|
||||||
cd installations
|
|
||||||
```
|
|
||||||
|
|
||||||
Set up host:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
./install_host cloud.silique.fr
|
|
||||||
```
|
|
||||||
|
|
||||||
Build container image:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
./install_images cloud.silique.fr
|
|
||||||
```
|
|
||||||
|
|
||||||
Set up the containers and start them up:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
./install_machines cloud.silique.fr
|
|
||||||
```
|
```
|
||||||
|
|
0
ansible/__init__.py
Normal file
0
ansible/action_plugins/__init__.py
Normal file
44
ansible/action_plugins/build_images.py
Normal file
|
@ -0,0 +1,44 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
from os import listdir, makedirs
|
||||||
|
from os.path import isfile, isdir, join, dirname
|
||||||
|
from shutil import copy2, copytree, rmtree
|
||||||
|
|
||||||
|
from ansible.plugins.action import ActionBase
|
||||||
|
from risotto.utils import RISOTTO_CONFIG
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
class ActionModule(ActionBase):
|
||||||
|
def run(self, tmp=None, task_vars=None):
|
||||||
|
super(ActionModule, self).run(tmp, task_vars)
|
||||||
|
module_args = self._task.args.copy()
|
||||||
|
modules = module_args['modules']
|
||||||
|
copy_tests = module_args.get('copy_tests', False)
|
||||||
|
dataset_directories = RISOTTO_CONFIG['directories']['datasets']
|
||||||
|
install_dir = join('/tmp/risotto/images')
|
||||||
|
if isdir(install_dir):
|
||||||
|
rmtree(install_dir)
|
||||||
|
if copy_tests:
|
||||||
|
install_tests_dir = join('/tmp/risotto/tests')
|
||||||
|
if isdir(install_tests_dir):
|
||||||
|
rmtree(install_tests_dir)
|
||||||
|
for module_name, depends in modules.items():
|
||||||
|
for dataset_directory in dataset_directories:
|
||||||
|
for depend in depends:
|
||||||
|
if copy_tests:
|
||||||
|
tests_dir = join(dataset_directory, depend, 'tests')
|
||||||
|
if isdir(tests_dir):
|
||||||
|
for filename in listdir(tests_dir):
|
||||||
|
src_file = join(tests_dir, filename)
|
||||||
|
dst_file = join(install_tests_dir, module_name, filename)
|
||||||
|
copy(src_file, dst_file)
|
||||||
|
# manual = join(dataset_directory, depend, 'manual', 'image')
|
||||||
|
# if not isdir(manual):
|
||||||
|
# continue
|
||||||
|
# for filename in listdir(manual):
|
||||||
|
# src_file = join(manual, filename)
|
||||||
|
# dst_file = join(install_dir, module_name, filename)
|
||||||
|
# copy(src_file, dst_file)
|
||||||
|
return dict(ansible_facts=dict({}))
|
14
ansible/action_plugins/machinectl.py
Normal file
|
@ -0,0 +1,14 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
from ansible.plugins.action import ActionBase
|
||||||
|
|
||||||
|
|
||||||
|
class ActionModule(ActionBase):
|
||||||
|
def run(self, tmp=None, task_vars=None):
|
||||||
|
super(ActionModule, self).run(tmp, task_vars)
|
||||||
|
module_args = self._task.args.copy()
|
||||||
|
module_return = self._execute_module(module_name='machinectl',
|
||||||
|
module_args=module_args,
|
||||||
|
task_vars=task_vars, tmp=tmp)
|
||||||
|
if module_return.get('failed'):
|
||||||
|
return module_return
|
||||||
|
return {'ansible_facts': {}, 'changed': module_return['changed']}
|
280
ansible/action_plugins/rougail.py
Normal file
|
@ -0,0 +1,280 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
from os import readlink, walk, chdir, getcwd, makedirs
|
||||||
|
from os.path import join, islink, isdir
|
||||||
|
from typing import Dict, Any
|
||||||
|
from shutil import rmtree, copy2
|
||||||
|
import tarfile
|
||||||
|
from ansible.module_utils._text import to_text
|
||||||
|
from ansible import constants
|
||||||
|
|
||||||
|
from rougail.template import base
|
||||||
|
from rougail.error import TemplateError
|
||||||
|
from risotto.machine import build_files, INSTALL_DIR, INSTALL_CONFIG_DIR, INSTALL_TMPL_DIR, INSTALL_IMAGES_DIR, INSTALL_TESTS_DIR
|
||||||
|
from risotto.utils import custom_filters
|
||||||
|
|
||||||
|
|
||||||
|
try:
|
||||||
|
from ansible.plugins.action import ActionBase
|
||||||
|
from ansible.module_utils.basic import AnsibleModule
|
||||||
|
class FakeModule(AnsibleModule):
|
||||||
|
def __init__(self):
|
||||||
|
pass
|
||||||
|
from ansible.plugins.action.template import ActionModule as TmplActionModule
|
||||||
|
except:
|
||||||
|
class ActionBase():
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
raise Exception('works only with ansible')
|
||||||
|
|
||||||
|
|
||||||
|
ARCHIVES_DIR = '/tmp/new_configurations'
|
||||||
|
|
||||||
|
|
||||||
|
def is_diff(server_name,
|
||||||
|
remote_directories,
|
||||||
|
certificates,
|
||||||
|
):
|
||||||
|
ret = {}
|
||||||
|
module = FakeModule()
|
||||||
|
current_path = getcwd()
|
||||||
|
root = join(INSTALL_DIR, INSTALL_CONFIG_DIR, server_name)
|
||||||
|
chdir(root)
|
||||||
|
search_paths = [join(directory[2:], f) for directory, subdirectories, files in walk('.') for f in files]
|
||||||
|
chdir(current_path)
|
||||||
|
for path in search_paths:
|
||||||
|
if path not in remote_directories:
|
||||||
|
return True
|
||||||
|
full_path = join(root, path)
|
||||||
|
if not islink(full_path):
|
||||||
|
if remote_directories[path] != module.digest_from_file(full_path, 'sha256'):
|
||||||
|
return True
|
||||||
|
elif remote_directories[path] != readlink(full_path):
|
||||||
|
return True
|
||||||
|
remote_directories.pop(path)
|
||||||
|
if remote_directories:
|
||||||
|
for certificate in certificates:
|
||||||
|
for typ in ['name', 'private', 'authority']:
|
||||||
|
if not typ in certificate:
|
||||||
|
continue
|
||||||
|
name = certificate[typ][1:]
|
||||||
|
if name in remote_directories:
|
||||||
|
remote_directories.pop(name)
|
||||||
|
if remote_directories:
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
class ActionModule(ActionBase):
|
||||||
|
def run(self, tmp=None, task_vars=None):
|
||||||
|
super(ActionModule, self).run(tmp, task_vars)
|
||||||
|
module_args = self._task.args.copy()
|
||||||
|
hostname = module_args.pop('hostname')
|
||||||
|
only_machine = module_args.pop('only_machine')
|
||||||
|
configure_host = module_args.pop('configure_host')
|
||||||
|
copy_tests = module_args.pop('copy_tests')
|
||||||
|
# define ansible engine
|
||||||
|
base.ENGINES['ansible'] = Tmpl(task_vars,
|
||||||
|
self._task,
|
||||||
|
self._connection,
|
||||||
|
self._play_context,
|
||||||
|
self._loader,
|
||||||
|
self._templar,
|
||||||
|
self._shared_loader_obj,
|
||||||
|
)
|
||||||
|
if 'copy_templates' in module_args:
|
||||||
|
copy_templates = module_args.pop('copy_templates')
|
||||||
|
else:
|
||||||
|
copy_templates = False
|
||||||
|
directories, certificates = build_files(hostname,
|
||||||
|
only_machine,
|
||||||
|
False,
|
||||||
|
copy_tests,
|
||||||
|
)
|
||||||
|
module_args['directories'] = list(directories.values())
|
||||||
|
module_args['directories'].append('/var/lib/risotto/images_files')
|
||||||
|
remote = self._execute_module(module_name='compare',
|
||||||
|
module_args=module_args,
|
||||||
|
task_vars=task_vars,
|
||||||
|
)
|
||||||
|
if remote.get('failed'):
|
||||||
|
if 'module_stdout' in remote:
|
||||||
|
msg = remote['module_stdout']
|
||||||
|
else:
|
||||||
|
msg = remote['msg']
|
||||||
|
raise Exception(f'error in remote action: {msg}')
|
||||||
|
if copy_templates:
|
||||||
|
build_files(hostname,
|
||||||
|
only_machine,
|
||||||
|
True,
|
||||||
|
copy_tests,
|
||||||
|
)
|
||||||
|
|
||||||
|
machines_changed = []
|
||||||
|
for machine, directory in directories.items():
|
||||||
|
if directory not in remote['directories']:
|
||||||
|
machines_changed.append(machine)
|
||||||
|
continue
|
||||||
|
if is_diff(machine,
|
||||||
|
remote['directories'][directory],
|
||||||
|
certificates['certificates'].get(machine, []),
|
||||||
|
):
|
||||||
|
machines_changed.append(machine)
|
||||||
|
current_path = getcwd()
|
||||||
|
if isdir(ARCHIVES_DIR):
|
||||||
|
rmtree(ARCHIVES_DIR)
|
||||||
|
makedirs(ARCHIVES_DIR)
|
||||||
|
if machines_changed:
|
||||||
|
self._execute_module(module_name='file',
|
||||||
|
module_args={'path': ARCHIVES_DIR,
|
||||||
|
'state': 'absent',
|
||||||
|
},
|
||||||
|
task_vars=task_vars,
|
||||||
|
)
|
||||||
|
self._execute_module(module_name='file',
|
||||||
|
module_args={'path': ARCHIVES_DIR,
|
||||||
|
'state': 'directory',
|
||||||
|
},
|
||||||
|
task_vars=task_vars,
|
||||||
|
)
|
||||||
|
machines = machines_changed.copy()
|
||||||
|
if self._task.args['hostname'] in machines_changed:
|
||||||
|
machine = self._task.args['hostname']
|
||||||
|
machines.remove(machine)
|
||||||
|
chdir(f'{task_vars["host_install_dir"]}/{INSTALL_CONFIG_DIR}/{machine}')
|
||||||
|
tar_filename = f'{ARCHIVES_DIR}/host.tar'
|
||||||
|
with tarfile.open(tar_filename, 'w') as archive:
|
||||||
|
archive.add('.')
|
||||||
|
chdir(current_path)
|
||||||
|
self._transfer_file(tar_filename, tar_filename)
|
||||||
|
|
||||||
|
# archive and send
|
||||||
|
if machines:
|
||||||
|
chdir(f'{task_vars["host_install_dir"]}/{INSTALL_CONFIG_DIR}')
|
||||||
|
tar_filename = f'{ARCHIVES_DIR}/machines.tar'
|
||||||
|
with tarfile.open(tar_filename, 'w') as archive:
|
||||||
|
for machine in machines:
|
||||||
|
if machine == self._task.args['hostname']:
|
||||||
|
continue
|
||||||
|
archive.add(f'{machine}')
|
||||||
|
self._transfer_file(tar_filename, tar_filename)
|
||||||
|
else:
|
||||||
|
machines = []
|
||||||
|
# archive and send
|
||||||
|
chdir(f'{task_vars["host_install_dir"]}/{INSTALL_IMAGES_DIR}/')
|
||||||
|
tar_filename = f'{ARCHIVES_DIR}/{INSTALL_IMAGES_DIR}.tar'
|
||||||
|
with tarfile.open(tar_filename, 'w') as archive:
|
||||||
|
archive.add('.')
|
||||||
|
self._execute_module(module_name='file',
|
||||||
|
module_args={'path': '/tmp/new_configurations',
|
||||||
|
'state': 'directory',
|
||||||
|
},
|
||||||
|
task_vars=task_vars,
|
||||||
|
)
|
||||||
|
self._transfer_file(tar_filename, tar_filename)
|
||||||
|
# tests
|
||||||
|
self._execute_module(module_name='file',
|
||||||
|
module_args={'path': '/var/lib/risotto/tests',
|
||||||
|
'state': 'absent',
|
||||||
|
},
|
||||||
|
task_vars=task_vars,
|
||||||
|
)
|
||||||
|
if copy_tests:
|
||||||
|
chdir(f'{task_vars["host_install_dir"]}/{INSTALL_TESTS_DIR}/')
|
||||||
|
tar_filename = f'{ARCHIVES_DIR}/{INSTALL_TESTS_DIR}.tar'
|
||||||
|
with tarfile.open(tar_filename, 'w') as archive:
|
||||||
|
archive.add('.')
|
||||||
|
self._transfer_file(tar_filename, tar_filename)
|
||||||
|
# templates
|
||||||
|
self._execute_module(module_name='file',
|
||||||
|
module_args={'path': '/var/lib/risotto/templates',
|
||||||
|
'state': 'absent',
|
||||||
|
},
|
||||||
|
task_vars=task_vars,
|
||||||
|
)
|
||||||
|
if copy_templates:
|
||||||
|
chdir(f'{task_vars["host_install_dir"]}/')
|
||||||
|
tar_filename = f'{ARCHIVES_DIR}/{INSTALL_TMPL_DIR}.tar'
|
||||||
|
with tarfile.open(tar_filename, 'w') as archive:
|
||||||
|
archive.add(INSTALL_TMPL_DIR)
|
||||||
|
self._transfer_file(tar_filename, tar_filename)
|
||||||
|
remote = self._execute_module(module_name='unarchive',
|
||||||
|
module_args={'remote_src': True,
|
||||||
|
'src': '/tmp/new_configurations/templates.tar',
|
||||||
|
'dest': '/var/lib/risotto',
|
||||||
|
},
|
||||||
|
task_vars=task_vars,
|
||||||
|
)
|
||||||
|
chdir(current_path)
|
||||||
|
changed = machines_changed != []
|
||||||
|
return dict(ansible_facts=dict({}),
|
||||||
|
changed=changed,
|
||||||
|
machines_changed=machines,
|
||||||
|
host_changed=self._task.args['hostname'] in machines_changed,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class FakeCopy:
|
||||||
|
def __init__(self, task):
|
||||||
|
self.task = task
|
||||||
|
|
||||||
|
def run(self, *args, **kwargs):
|
||||||
|
copy2(self.task.args['src'], self.task.args['dest'])
|
||||||
|
return {}
|
||||||
|
|
||||||
|
|
||||||
|
class FakeGet:
|
||||||
|
def __init__(self, klass):
|
||||||
|
self.klass = klass
|
||||||
|
|
||||||
|
def fake_get(self, action, *args, task, **kwargs):
|
||||||
|
if action == 'ansible.legacy.copy':
|
||||||
|
return FakeCopy(task)
|
||||||
|
return self.klass.ori_get(action, *args, task=task, **kwargs)
|
||||||
|
|
||||||
|
|
||||||
|
class Tmpl(TmplActionModule):
|
||||||
|
def __init__(self, task_vars, *args):
|
||||||
|
super().__init__(*args)
|
||||||
|
self.task_vars = task_vars
|
||||||
|
|
||||||
|
def _early_needs_tmp_path(self):
|
||||||
|
# do not create tmp remotely
|
||||||
|
return False
|
||||||
|
|
||||||
|
def process(self,
|
||||||
|
filename: str,
|
||||||
|
source: str,
|
||||||
|
true_destfilename: str,
|
||||||
|
destfilename: str,
|
||||||
|
destdir: str,
|
||||||
|
variable: Any,
|
||||||
|
index: int,
|
||||||
|
rougail_variables_dict: Dict,
|
||||||
|
eosfunc: Dict,
|
||||||
|
extra_variables: Any=None,
|
||||||
|
):
|
||||||
|
if source is not None: # pragma: no cover
|
||||||
|
raise TemplateError(_('source is not supported for ansible'))
|
||||||
|
task_vars = rougail_variables_dict | self.task_vars
|
||||||
|
if variable is not None:
|
||||||
|
task_vars['rougail_variable'] = variable
|
||||||
|
if index is not None:
|
||||||
|
task_vars['rougail_index'] = index
|
||||||
|
if extra_variables:
|
||||||
|
task_vars['extra_variables'] = extra_variables
|
||||||
|
task_vars['rougail_filename'] = true_destfilename
|
||||||
|
task_vars['rougail_destination_dir'] = destdir
|
||||||
|
self._task.args['src'] = filename
|
||||||
|
self._task.args['dest'] = destfilename
|
||||||
|
# add custom filter
|
||||||
|
custom_filters.update(eosfunc)
|
||||||
|
# do not copy file in host but stay it locally
|
||||||
|
self._shared_loader_obj.action_loader.ori_get = self._shared_loader_obj.action_loader.get
|
||||||
|
self._shared_loader_obj.action_loader.get = FakeGet(self._shared_loader_obj.action_loader).fake_get
|
||||||
|
# template
|
||||||
|
ret = self.run(task_vars=task_vars)
|
||||||
|
# restore get function
|
||||||
|
self._shared_loader_obj.action_loader.get = self._shared_loader_obj.action_loader.ori_get
|
||||||
|
# remove custom filter
|
||||||
|
custom_filters.clear()
|
||||||
|
if ret.get('failed'):
|
||||||
|
raise TemplateError(f'error while templating "{filename}": {ret["msg"]}')
|
1
ansible/file.txt
Normal file
|
@ -0,0 +1 @@
|
||||||
|
{'pouet': 'a'}
|
8
ansible/filter_plugins/custom.py
Normal file
|
@ -0,0 +1,8 @@
|
||||||
|
from risotto.utils import custom_filters
|
||||||
|
|
||||||
|
|
||||||
|
class FilterModule:
|
||||||
|
"""This filter is used to load custom filter from dataset
|
||||||
|
"""
|
||||||
|
def filters(self):
|
||||||
|
return custom_filters
|
140
ansible/filter_plugins/fileslist.py
Normal file
|
@ -0,0 +1,140 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
|
||||||
|
from os.path import dirname
|
||||||
|
|
||||||
|
|
||||||
|
def _add(files, file_data, name, name_only, prefix):
|
||||||
|
if prefix is not None:
|
||||||
|
name = prefix + name
|
||||||
|
if name_only:
|
||||||
|
files.append(name)
|
||||||
|
else:
|
||||||
|
files.append({'name': name,
|
||||||
|
'owner': file_data['owner'],
|
||||||
|
'group': file_data['group'],
|
||||||
|
'mode': file_data['mode'],
|
||||||
|
})
|
||||||
|
|
||||||
|
|
||||||
|
def fileslist(data, is_host=False, name_only=False, prefix=None):
|
||||||
|
files = []
|
||||||
|
if is_host:
|
||||||
|
base_systemd = '/usr/local/lib'
|
||||||
|
else:
|
||||||
|
base_systemd = ''
|
||||||
|
_add(files,
|
||||||
|
{'owner': 'root', 'group': 'root', 'mode': '0755'},
|
||||||
|
f'/tmpfiles.d/0rougail.conf',
|
||||||
|
name_only,
|
||||||
|
prefix,
|
||||||
|
)
|
||||||
|
for service, service_data in data.items():
|
||||||
|
if not service_data['activate']:
|
||||||
|
if service_data['manage']:
|
||||||
|
if not service_data.get('undisable', False) and not service_data['engine'] and not service_data.get('target'):
|
||||||
|
_add(files,
|
||||||
|
{'owner': 'root', 'group': 'root', 'mode': '0755'},
|
||||||
|
base_systemd + '/systemd/system/' + service_data['doc'],
|
||||||
|
name_only,
|
||||||
|
prefix,
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
if service_data['manage'] and service_data['engine']:
|
||||||
|
_add(files,
|
||||||
|
{'owner': 'root', 'group': 'root', 'mode': '0755'},
|
||||||
|
base_systemd + '/systemd/system/' + service_data['doc'],
|
||||||
|
name_only,
|
||||||
|
prefix,
|
||||||
|
)
|
||||||
|
if service_data.get('target'):
|
||||||
|
_add(files,
|
||||||
|
{'owner': 'root', 'group': 'root', 'mode': '0755'},
|
||||||
|
f'/systemd/system/{service_data["target"]}.target.wants/{service_data["doc"]}',
|
||||||
|
name_only,
|
||||||
|
prefix,
|
||||||
|
)
|
||||||
|
if 'overrides' in service_data:
|
||||||
|
for override_data in service_data['overrides'].values():
|
||||||
|
_add(files,
|
||||||
|
{'owner': 'root', 'group': 'root', 'mode': '0755'},
|
||||||
|
base_systemd + '/systemd/system/' + override_data['name'] + '.d/rougail.conf',
|
||||||
|
name_only,
|
||||||
|
prefix,
|
||||||
|
)
|
||||||
|
if 'ip' in service_data:
|
||||||
|
_add(files,
|
||||||
|
{'owner': 'root', 'group': 'root', 'mode': '0755'},
|
||||||
|
base_systemd + '/systemd/system/' + service_data['doc'] + '.d/rougail_ip.conf',
|
||||||
|
name_only,
|
||||||
|
prefix,
|
||||||
|
)
|
||||||
|
if 'files' not in service_data:
|
||||||
|
continue
|
||||||
|
for file_data in service_data['files'].values():
|
||||||
|
if not file_data['activate'] or file_data['included'] == 'content':
|
||||||
|
continue
|
||||||
|
if isinstance(file_data['name'], list):
|
||||||
|
for name in file_data['name']:
|
||||||
|
_add(files, file_data, name, name_only, prefix)
|
||||||
|
else:
|
||||||
|
_add(files, file_data, file_data['name'], name_only, prefix)
|
||||||
|
|
||||||
|
return files
|
||||||
|
|
||||||
|
|
||||||
|
def directorieslist(data):
|
||||||
|
directories = {'/usr/local/lib/systemd/system/'}
|
||||||
|
for service, service_data in data.items():
|
||||||
|
if 'files' not in service_data:
|
||||||
|
continue
|
||||||
|
for file_data in service_data['files'].values():
|
||||||
|
if not file_data['activate']:
|
||||||
|
continue
|
||||||
|
if isinstance(file_data['name'], list):
|
||||||
|
for name in file_data['name']:
|
||||||
|
directories.add(dirname(name))
|
||||||
|
else:
|
||||||
|
directories.add(dirname(file_data['name']))
|
||||||
|
|
||||||
|
return list(directories)
|
||||||
|
|
||||||
|
|
||||||
|
def machineslist(data, only=None, only_name=False):
|
||||||
|
srv = []
|
||||||
|
if only is not None:
|
||||||
|
if only not in data:
|
||||||
|
raise Exception(f"cannot find {only} but only {data.keys()}")
|
||||||
|
if only_name:
|
||||||
|
srv.append(only)
|
||||||
|
else:
|
||||||
|
srv.append({'name': only,
|
||||||
|
'srv': data[only]['machine']['add_srv'],
|
||||||
|
}
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
for host, host_data in data.items():
|
||||||
|
if '.' not in host or not isinstance(host_data, dict) or 'general' not in host_data or host_data['general']['module_name'] == 'host':
|
||||||
|
continue
|
||||||
|
if only_name:
|
||||||
|
srv.append(host)
|
||||||
|
else:
|
||||||
|
srv.append({'name': host,
|
||||||
|
'srv': host_data['machine']['add_srv'],
|
||||||
|
}
|
||||||
|
)
|
||||||
|
return srv
|
||||||
|
|
||||||
|
|
||||||
|
def modulename(data, servername):
|
||||||
|
return data[servername]['general']['module_name']
|
||||||
|
|
||||||
|
|
||||||
|
class FilterModule:
|
||||||
|
def filters(self):
|
||||||
|
return {
|
||||||
|
'fileslist': fileslist,
|
||||||
|
'directorieslist': directorieslist,
|
||||||
|
'machineslist': machineslist,
|
||||||
|
'modulename': modulename,
|
||||||
|
}
|
32
ansible/filter_plugins/normalize_family.py
Normal file
|
@ -0,0 +1,32 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
"""
|
||||||
|
Silique (https://www.silique.fr)
|
||||||
|
Copyright (C) 2023
|
||||||
|
|
||||||
|
distribued with GPL-2 or later license
|
||||||
|
|
||||||
|
This program is free software; you can redistribute it and/or modify
|
||||||
|
it under the terms of the GNU General Public License as published by
|
||||||
|
the Free Software Foundation; either version 2 of the License, or
|
||||||
|
(at your option) any later version.
|
||||||
|
|
||||||
|
This program is distributed in the hope that it will be useful,
|
||||||
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
GNU General Public License for more details.
|
||||||
|
|
||||||
|
You should have received a copy of the GNU General Public License
|
||||||
|
along with this program; if not, write to the Free Software
|
||||||
|
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
||||||
|
"""
|
||||||
|
|
||||||
|
|
||||||
|
from rougail.utils import normalize_family
|
||||||
|
|
||||||
|
|
||||||
|
class FilterModule:
|
||||||
|
def filters(self):
|
||||||
|
return {
|
||||||
|
'normalize_family': normalize_family,
|
||||||
|
}
|
11
ansible/filter_plugins/raise.py
Normal file
|
@ -0,0 +1,11 @@
|
||||||
|
from jinja2.exceptions import TemplateRuntimeError
|
||||||
|
|
||||||
|
def fraise(msg):
|
||||||
|
raise TemplateRuntimeError(msg)
|
||||||
|
|
||||||
|
|
||||||
|
class FilterModule:
|
||||||
|
def filters(self):
|
||||||
|
return {
|
||||||
|
'raise': fraise,
|
||||||
|
}
|
101
ansible/host.yml
Normal file
|
@ -0,0 +1,101 @@
|
||||||
|
---
|
||||||
|
- name: "Populate service facts"
|
||||||
|
service_facts:
|
||||||
|
|
||||||
|
- name: "Set timezone"
|
||||||
|
timezone:
|
||||||
|
name: Europe/Paris
|
||||||
|
|
||||||
|
- name: Set a hostname
|
||||||
|
ansible.builtin.hostname:
|
||||||
|
name: "{{ inventory_hostname }}"
|
||||||
|
|
||||||
|
- name: "Install packages"
|
||||||
|
apt:
|
||||||
|
pkg: "{{ vars[inventory_hostname]['general']['host_packages'] }}"
|
||||||
|
update_cache: yes
|
||||||
|
state: latest
|
||||||
|
|
||||||
|
- name: "Remove packages"
|
||||||
|
apt:
|
||||||
|
pkg: "{{ vars[inventory_hostname]['general']['host_removed_packages'] }}"
|
||||||
|
update_cache: yes
|
||||||
|
state: absent
|
||||||
|
|
||||||
|
- name: "Add keyrings directory"
|
||||||
|
file:
|
||||||
|
path: /etc/apt/keyrings
|
||||||
|
state: directory
|
||||||
|
mode: "755"
|
||||||
|
|
||||||
|
- name: "Add vector signed repositories"
|
||||||
|
ansible.builtin.get_url:
|
||||||
|
url: https://repositories.timber.io/public/vector/gpg.3543DB2D0A2BC4B8.key
|
||||||
|
dest: /etc/apt/keyrings/vector.asc
|
||||||
|
|
||||||
|
- name: "Add vector repository"
|
||||||
|
ansible.builtin.apt_repository:
|
||||||
|
repo: "deb [signed-by=/etc/apt/keyrings/vector.asc] https://repositories.timber.io/public/vector/deb/debian {{ ansible_distribution_release }} main"
|
||||||
|
state: present
|
||||||
|
|
||||||
|
- name: "Install vector"
|
||||||
|
ansible.builtin.apt:
|
||||||
|
name: vector
|
||||||
|
update_cache: yes
|
||||||
|
state: present
|
||||||
|
|
||||||
|
- name: "Host is modified"
|
||||||
|
include_tasks: host_modified.yml
|
||||||
|
when: build_host.host_changed
|
||||||
|
|
||||||
|
- name: "Copy machines scripts"
|
||||||
|
ansible.builtin.copy:
|
||||||
|
src: "{{ item }}"
|
||||||
|
dest: "/usr/local/sbin"
|
||||||
|
owner: "root"
|
||||||
|
group: "root"
|
||||||
|
mode: "0755"
|
||||||
|
loop: "{{ lookup('fileglob', 'sbin/*', wantlist=True) | list }}"
|
||||||
|
|
||||||
|
- name: "Remove dest images files"
|
||||||
|
file:
|
||||||
|
path: /var/lib/risotto/images_files
|
||||||
|
state: "{{ item }}"
|
||||||
|
mode: "0700"
|
||||||
|
with_items:
|
||||||
|
- absent
|
||||||
|
- directory
|
||||||
|
|
||||||
|
- name: "Copy images files"
|
||||||
|
unarchive:
|
||||||
|
remote_src: true
|
||||||
|
src: "/tmp/new_configurations/images_files.tar"
|
||||||
|
dest: "/var/lib/risotto/images_files"
|
||||||
|
|
||||||
|
- name: "Create versions directory"
|
||||||
|
file:
|
||||||
|
path: /var/lib/risotto/machines_informations
|
||||||
|
state: directory
|
||||||
|
mode: "0700"
|
||||||
|
|
||||||
|
- name: "Empty tests files"
|
||||||
|
file:
|
||||||
|
path: /var/lib/risotto/tests
|
||||||
|
state: "{{ item }}"
|
||||||
|
mode: "0700"
|
||||||
|
with_items:
|
||||||
|
- absent
|
||||||
|
- directory
|
||||||
|
|
||||||
|
- name: "Copy tests files"
|
||||||
|
unarchive:
|
||||||
|
remote_src: true
|
||||||
|
src: "/tmp/new_configurations/tests.tar"
|
||||||
|
dest: "/var/lib/risotto/tests"
|
||||||
|
when: copy_tests
|
||||||
|
|
||||||
|
- name: "Create TLS directory"
|
||||||
|
file:
|
||||||
|
path: /var/lib/risotto/tls
|
||||||
|
state: directory
|
||||||
|
mode: "755"
|
75
ansible/host_modified.yml
Normal file
|
@ -0,0 +1,75 @@
|
||||||
|
- name: "Stop services"
|
||||||
|
ansible.builtin.service:
|
||||||
|
name: "{{ item.value['doc'] }}"
|
||||||
|
state: stopped
|
||||||
|
when: item.value['manage'] and item.value['activate'] and item.value['doc'].endswith('.service') and not item.value['doc'].endswith('@.service') and item.value['engine'] and item.value['doc'] in services
|
||||||
|
loop: "{{ vars[inventory_hostname]['services'] | dict2items }}"
|
||||||
|
loop_control:
|
||||||
|
label: "{{ item.value['doc'] }}"
|
||||||
|
|
||||||
|
- name: "Remove old config files"
|
||||||
|
file:
|
||||||
|
path: /usr/local/lib/
|
||||||
|
state: "{{ item }}"
|
||||||
|
mode: "0700"
|
||||||
|
with_items:
|
||||||
|
- absent
|
||||||
|
- directory
|
||||||
|
|
||||||
|
- name: "Copy config files"
|
||||||
|
unarchive:
|
||||||
|
remote_src: true
|
||||||
|
src: "/tmp/new_configurations/host.tar"
|
||||||
|
dest: /usr/local/lib/
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
|
||||||
|
- name: "Execute systemd-tmpfiles"
|
||||||
|
command: /usr/bin/systemd-tmpfiles --create --clean --remove -E --exclude-prefix=/tmp
|
||||||
|
|
||||||
|
- name: "Remove tmpfiles files directory"
|
||||||
|
local_action:
|
||||||
|
module: file
|
||||||
|
path: /usr/local/lib/tmpfiles.d/
|
||||||
|
state: absent
|
||||||
|
|
||||||
|
- name: "Reload systemd services configuration"
|
||||||
|
ansible.builtin.systemd:
|
||||||
|
daemon_reload: yes
|
||||||
|
|
||||||
|
- name: "Enable services"
|
||||||
|
when: item.value['manage'] and item.value['activate'] and '@.service' not in item.value['doc']
|
||||||
|
ansible.builtin.service:
|
||||||
|
name: "{{ item.value['doc'] }}"
|
||||||
|
enabled: yes
|
||||||
|
loop: "{{ vars[inventory_hostname]['services'] | dict2items }}"
|
||||||
|
loop_control:
|
||||||
|
label: "{{ item.value['doc'] }}"
|
||||||
|
|
||||||
|
- name: "Disable services"
|
||||||
|
when: item.value['manage'] and not item.value['activate'] and not item.value['undisable'] and '@.service' not in item.value['doc']
|
||||||
|
ansible.builtin.service:
|
||||||
|
name: "{{ item.value['doc'] }}"
|
||||||
|
enabled: no
|
||||||
|
loop: "{{ vars[inventory_hostname]['services'] | dict2items }}"
|
||||||
|
loop_control:
|
||||||
|
label: "{{ item.value['doc'] }}"
|
||||||
|
|
||||||
|
- name: "Start services"
|
||||||
|
when: item.value['manage'] and item.value['activate'] and item.value['doc'].endswith('.service') and not item.value['doc'].endswith('@.service') and item.value['engine']
|
||||||
|
ignore_errors: true
|
||||||
|
ansible.builtin.service:
|
||||||
|
name: "{{ item.value['doc'] }}"
|
||||||
|
state: started
|
||||||
|
loop: "{{ vars[inventory_hostname]['services'] | dict2items }}"
|
||||||
|
loop_control:
|
||||||
|
label: "{{ item.value['doc'] }}"
|
||||||
|
|
||||||
|
- name: "Restart services"
|
||||||
|
when: item.value['manage'] and item.value['activate'] and item.value['doc'].endswith('.service') and not item.value['doc'].endswith('@.service') and not item.value['engine']
|
||||||
|
ansible.builtin.service:
|
||||||
|
name: "{{ item.value['doc'] }}"
|
||||||
|
state: restarted
|
||||||
|
loop: "{{ vars[inventory_hostname]['services'] | dict2items }}"
|
||||||
|
loop_control:
|
||||||
|
label: "{{ item.value['doc'] }}"
|
136
ansible/inventory.py
Executable file
|
@ -0,0 +1,136 @@
|
||||||
|
#!/usr/bin/env python
|
||||||
|
|
||||||
|
'''
|
||||||
|
Example custom dynamic inventory script for Ansible, in Python.
|
||||||
|
'''
|
||||||
|
|
||||||
|
from argparse import ArgumentParser
|
||||||
|
from json import load as json_load, dumps, JSONEncoder
|
||||||
|
from os import remove
|
||||||
|
from os.path import isfile
|
||||||
|
from traceback import print_exc
|
||||||
|
from sys import stderr, argv
|
||||||
|
|
||||||
|
from risotto.machine import load, TIRAMISU_CACHE, VALUES_CACHE, INFORMATIONS_CACHE, ROUGAIL_NAMESPACE, ROUGAIL_NAMESPACE_DESCRIPTION
|
||||||
|
from tiramisu import Config
|
||||||
|
from tiramisu.error import PropertiesOptionError
|
||||||
|
from rougail.utils import normalize_family
|
||||||
|
from rougail import RougailSystemdTemplate, RougailConfig
|
||||||
|
from rougail.template.base import RougailLeader, RougailExtra
|
||||||
|
|
||||||
|
|
||||||
|
DEBUG = False
|
||||||
|
|
||||||
|
|
||||||
|
class RougailEncoder(JSONEncoder):
|
||||||
|
def default(self, obj):
|
||||||
|
if isinstance(obj, RougailLeader):
|
||||||
|
return obj._follower
|
||||||
|
if isinstance(obj, RougailExtra):
|
||||||
|
return obj._suboption
|
||||||
|
if isinstance(obj, PropertiesOptionError):
|
||||||
|
return 'PropertiesOptionError'
|
||||||
|
return JSONEncoder.default(self, obj)
|
||||||
|
|
||||||
|
|
||||||
|
class RisottoInventory(object):
|
||||||
|
def __init__(self):
|
||||||
|
parser = ArgumentParser()
|
||||||
|
parser.add_argument('--list', action='store_true')
|
||||||
|
parser.add_argument('--host', action='store')
|
||||||
|
parser.add_argument('--nocache', action='store_true')
|
||||||
|
parser.add_argument('--debug', action='store_true')
|
||||||
|
parser.add_argument('--pretty_print', action='store_true')
|
||||||
|
parser.add_argument('--quite', action='store_true')
|
||||||
|
self.args = parser.parse_args()
|
||||||
|
if self.args.debug:
|
||||||
|
global DEBUG
|
||||||
|
DEBUG = True
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
if self.args.list and self.args.host:
|
||||||
|
raise Exception('cannot have --list and --host together')
|
||||||
|
if self.args.list or self.args.nocache:
|
||||||
|
if isfile(TIRAMISU_CACHE):
|
||||||
|
remove(TIRAMISU_CACHE)
|
||||||
|
if isfile(VALUES_CACHE):
|
||||||
|
remove(VALUES_CACHE)
|
||||||
|
if isfile(INFORMATIONS_CACHE):
|
||||||
|
remove(INFORMATIONS_CACHE)
|
||||||
|
config = load(TIRAMISU_CACHE,
|
||||||
|
VALUES_CACHE,
|
||||||
|
INFORMATIONS_CACHE,
|
||||||
|
)
|
||||||
|
if self.args.list:
|
||||||
|
return self.do_inventory(config)
|
||||||
|
elif self.args.host:
|
||||||
|
return self.get_vars(config, self.args.host)
|
||||||
|
raise Exception('pfff')
|
||||||
|
|
||||||
|
def do_inventory(self,
|
||||||
|
config: Config,
|
||||||
|
) -> dict:
|
||||||
|
servers = [subconfig.doc() for subconfig in config.option.list('optiondescription') if subconfig.information.get('module') == 'host']
|
||||||
|
return dumps({
|
||||||
|
'group': {
|
||||||
|
'hosts': servers,
|
||||||
|
'vars': {
|
||||||
|
# FIXME
|
||||||
|
# 'ansible_ssh_host': '192.168.0.28',
|
||||||
|
'ansible_ssh_user': 'root',
|
||||||
|
'ansible_python_interpreter': '/usr/bin/python3'
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
def get_vars(self,
|
||||||
|
config: Config,
|
||||||
|
host_name: str,
|
||||||
|
) -> dict:
|
||||||
|
ret = {}
|
||||||
|
rougailconfig = RougailConfig.copy()
|
||||||
|
rougailconfig['variable_namespace'] = ROUGAIL_NAMESPACE
|
||||||
|
rougailconfig['variable_namespace_description'] = ROUGAIL_NAMESPACE_DESCRIPTION
|
||||||
|
for subconfig in config.option.list('optiondescription'):
|
||||||
|
server_name = subconfig.description()
|
||||||
|
module_name = subconfig.option(subconfig.information.get('provider:global:module_name')).value.get()
|
||||||
|
if module_name == 'host' and server_name != host_name:
|
||||||
|
continue
|
||||||
|
engine = RougailSystemdTemplate(subconfig, rougailconfig)
|
||||||
|
engine.load_variables(with_flatten=False)
|
||||||
|
if module_name != 'host' and engine.rougail_variables_dict['general']['host'] != host_name:
|
||||||
|
continue
|
||||||
|
ret[server_name] = engine.rougail_variables_dict
|
||||||
|
ret['modules'] = config.information.get('modules')
|
||||||
|
ret['delete_old_image'] = False
|
||||||
|
ret['configure_host'] = True
|
||||||
|
ret['only_machine'] = None
|
||||||
|
ret['copy_templates'] = False
|
||||||
|
ret['copy_tests'] = False
|
||||||
|
ret['host_install_dir'] = ret[host_name]['general']['host_install_dir']
|
||||||
|
return dumps(ret, cls=RougailEncoder)
|
||||||
|
|
||||||
|
|
||||||
|
# Get the inventory.
|
||||||
|
def main():
|
||||||
|
try:
|
||||||
|
inv = RisottoInventory()
|
||||||
|
values = inv.run()
|
||||||
|
if inv.args.pretty_print:
|
||||||
|
from pprint import pprint
|
||||||
|
from json import loads
|
||||||
|
pprint(loads(values))
|
||||||
|
elif not inv.args.quite:
|
||||||
|
print(values)
|
||||||
|
except Exception as err:
|
||||||
|
if DEBUG:
|
||||||
|
print_exc()
|
||||||
|
print('---', file=stderr)
|
||||||
|
extra=''
|
||||||
|
else:
|
||||||
|
extra=f'\nmore informations with commandline "{" ".join(argv)} --debug"'
|
||||||
|
print(f'{err}{extra}', file=stderr)
|
||||||
|
exit(1)
|
||||||
|
|
||||||
|
|
||||||
|
main()
|
83
ansible/library/compare.py
Normal file
|
@ -0,0 +1,83 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
from time import sleep
|
||||||
|
from os import fdopen, walk, readlink, chdir, getcwd
|
||||||
|
from os.path import join, islink, isdir
|
||||||
|
|
||||||
|
from ansible.module_utils.basic import AnsibleModule
|
||||||
|
|
||||||
|
|
||||||
|
def run_module():
|
||||||
|
# define available arguments/parameters a user can pass to the module
|
||||||
|
module_args = dict(
|
||||||
|
# shasums=dict(type='dict', required=True),
|
||||||
|
directories=dict(type='list', required=True),
|
||||||
|
)
|
||||||
|
|
||||||
|
# seed the result dict in the object
|
||||||
|
# we primarily care about changed and state
|
||||||
|
# changed is if this module effectively modified the target
|
||||||
|
# state will include any data that you want your module to pass back
|
||||||
|
# for consumption, for example, in a subsequent task
|
||||||
|
result = dict(
|
||||||
|
directories={},
|
||||||
|
)
|
||||||
|
|
||||||
|
# the AnsibleModule object will be our abstraction working with Ansible
|
||||||
|
# this includes instantiation, a couple of common attr would be the
|
||||||
|
# args/params passed to the execution, as well as if the module
|
||||||
|
# supports check mode
|
||||||
|
module = AnsibleModule(
|
||||||
|
argument_spec=module_args,
|
||||||
|
supports_check_mode=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
current_path = getcwd()
|
||||||
|
for directory in module.params['directories']:
|
||||||
|
result['directories'][directory] = {}
|
||||||
|
if not isdir(directory):
|
||||||
|
continue
|
||||||
|
chdir(directory)
|
||||||
|
search_paths = [join(directory_[2:], f) for directory_, subdirectories, files in walk('.') for f in files]
|
||||||
|
for path in search_paths:
|
||||||
|
full_path = join(directory, path)
|
||||||
|
if not islink(full_path):
|
||||||
|
result['directories'][directory][path] = module.digest_from_file(full_path, 'sha256')
|
||||||
|
else:
|
||||||
|
result['directories'][directory][path] = readlink(full_path)
|
||||||
|
chdir(current_path)
|
||||||
|
# current_path = getcwd()
|
||||||
|
# for server_name, dico in module.params['shasums'].items():
|
||||||
|
# root = dico['config_dir']
|
||||||
|
# if not isdir(root):
|
||||||
|
# result['machines_changed'].append(server_name)
|
||||||
|
# continue
|
||||||
|
# chdir(root)
|
||||||
|
# search_paths = [join(directory[2:], f) for directory, subdirectories, files in walk('.') for f in files]
|
||||||
|
# chdir(current_path)
|
||||||
|
# for path in search_paths:
|
||||||
|
# if path in dico['shasums']:
|
||||||
|
# full_path = join(root, path)
|
||||||
|
# if not islink(full_path):
|
||||||
|
# if module.digest_from_file(full_path, 'sha256') != dico['shasums'][path]:
|
||||||
|
# result['machines_changed'].append(server_name)
|
||||||
|
# break
|
||||||
|
# elif dico['shasums'][path] != readlink(full_path):
|
||||||
|
# result['machines_changed'].append(server_name)
|
||||||
|
# break
|
||||||
|
# del dico['shasums'][path]
|
||||||
|
# else:
|
||||||
|
# result['machines_changed'].append(server_name)
|
||||||
|
# break
|
||||||
|
# if server_name not in result['machines_changed'] and dico['shasums']:
|
||||||
|
# result['machines_changed'].append(server_name)
|
||||||
|
|
||||||
|
module.exit_json(**result)
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
run_module()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
219
ansible/library/machinectl.py
Normal file
|
@ -0,0 +1,219 @@
|
||||||
|
#!/usr/bin/python3
|
||||||
|
|
||||||
|
from time import sleep
|
||||||
|
from os import fdopen
|
||||||
|
from dbus import SystemBus, Array
|
||||||
|
from dbus.exceptions import DBusException
|
||||||
|
from subprocess import run
|
||||||
|
|
||||||
|
from ansible.module_utils.basic import AnsibleModule
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def stop(bus, machines):
|
||||||
|
changed = False
|
||||||
|
remote_object = bus.get_object('org.freedesktop.machine1',
|
||||||
|
'/org/freedesktop/machine1',
|
||||||
|
False,
|
||||||
|
)
|
||||||
|
res = remote_object.ListMachines(dbus_interface='org.freedesktop.machine1.Manager')
|
||||||
|
started_machines = [str(r[0]) for r in res if str(r[0]) != '.host']
|
||||||
|
for host in machines:
|
||||||
|
if host not in started_machines:
|
||||||
|
continue
|
||||||
|
changed = True
|
||||||
|
remote_object.TerminateMachine(host, dbus_interface='org.freedesktop.machine1.Manager')
|
||||||
|
idx = 0
|
||||||
|
errors = []
|
||||||
|
while True:
|
||||||
|
res = remote_object.ListMachines(dbus_interface='org.freedesktop.machine1.Manager')
|
||||||
|
started_machines = [str(r[0]) for r in res if str(r[0]) != '.host']
|
||||||
|
for host in machines:
|
||||||
|
if host in started_machines:
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
break
|
||||||
|
sleep(1)
|
||||||
|
idx += 1
|
||||||
|
if idx == 120:
|
||||||
|
errors.append('Cannot not stopped: ' + ','.join(started_machines))
|
||||||
|
break
|
||||||
|
return changed, errors
|
||||||
|
|
||||||
|
|
||||||
|
def start(bus, machines):
|
||||||
|
changed = False
|
||||||
|
remote_object = bus.get_object('org.freedesktop.machine1',
|
||||||
|
'/org/freedesktop/machine1',
|
||||||
|
False,
|
||||||
|
)
|
||||||
|
res = remote_object.ListMachines(dbus_interface='org.freedesktop.machine1.Manager')
|
||||||
|
started_machines = [str(r[0]) for r in res if str(r[0]) != '.host']
|
||||||
|
remote_object_system = bus.get_object('org.freedesktop.systemd1',
|
||||||
|
'/org/freedesktop/systemd1',
|
||||||
|
False,
|
||||||
|
)
|
||||||
|
for host in machines:
|
||||||
|
if host in started_machines:
|
||||||
|
continue
|
||||||
|
changed = True
|
||||||
|
service = f'systemd-nspawn@{host}.service'
|
||||||
|
remote_object_system.StartUnit(service, 'fail', dbus_interface='org.freedesktop.systemd1.Manager')
|
||||||
|
errors = []
|
||||||
|
idx = 0
|
||||||
|
while True:
|
||||||
|
res = remote_object.ListMachines(dbus_interface='org.freedesktop.machine1.Manager')
|
||||||
|
started_machines = [str(r[0]) for r in res if str(r[0]) != '.host']
|
||||||
|
for host in machines:
|
||||||
|
if host not in started_machines:
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
break
|
||||||
|
sleep(1)
|
||||||
|
idx += 1
|
||||||
|
if idx == 120:
|
||||||
|
hosts = set(machines) - set(started_machines)
|
||||||
|
errors.append('Cannot not start: ' + ','.join(hosts))
|
||||||
|
break
|
||||||
|
if not errors:
|
||||||
|
idx = 0
|
||||||
|
for host in machines:
|
||||||
|
cmd = ['/usr/bin/systemctl', 'is-system-running']
|
||||||
|
error = False
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
ret = []
|
||||||
|
res = remote_object.OpenMachineShell(host,
|
||||||
|
'',
|
||||||
|
cmd[0],
|
||||||
|
Array(cmd, signature='s'),
|
||||||
|
Array(['TERM=dumb'], signature='s'),
|
||||||
|
dbus_interface='org.freedesktop.machine1.Manager',
|
||||||
|
)
|
||||||
|
fd = res[0].take()
|
||||||
|
fh = fdopen(fd)
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
ret.append(fh.readline().strip())
|
||||||
|
except OSError as err:
|
||||||
|
if err.errno != 5:
|
||||||
|
raise err from err
|
||||||
|
break
|
||||||
|
if not ret:
|
||||||
|
errors.append(f'Cannot check {host} status')
|
||||||
|
error = True
|
||||||
|
break
|
||||||
|
if ret[0] in ['running', 'degraded']:
|
||||||
|
break
|
||||||
|
except DBusException:
|
||||||
|
pass
|
||||||
|
idx += 1
|
||||||
|
sleep(1)
|
||||||
|
if idx == 120:
|
||||||
|
errors.append(f'Cannot not start {host} ({ret})')
|
||||||
|
break
|
||||||
|
if error:
|
||||||
|
continue
|
||||||
|
if ret and ret[0] == 'running':
|
||||||
|
continue
|
||||||
|
cmd = ['/usr/bin/systemctl', '--state=failed', '--no-legend', '--no-page']
|
||||||
|
res = remote_object.OpenMachineShell(host,
|
||||||
|
'',
|
||||||
|
cmd[0],
|
||||||
|
Array(cmd, signature='s'),
|
||||||
|
Array(['TERM=dumb'], signature='s'),
|
||||||
|
dbus_interface='org.freedesktop.machine1.Manager',
|
||||||
|
)
|
||||||
|
fd = res[0].take()
|
||||||
|
fh = fdopen(fd)
|
||||||
|
ret = []
|
||||||
|
idx2 = 0
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
ret.append(fh.readline().strip())
|
||||||
|
except OSError as err:
|
||||||
|
if err.errno != 5:
|
||||||
|
raise err from err
|
||||||
|
break
|
||||||
|
idx2 += 1
|
||||||
|
if idx2 == 120:
|
||||||
|
errors.append(f'Cannot not get status to {host}')
|
||||||
|
break
|
||||||
|
errors.append(f'{host}: ' + '\n'.join(ret))
|
||||||
|
return changed, errors
|
||||||
|
|
||||||
|
|
||||||
|
def enable(machines):
|
||||||
|
cmd = ['/usr/bin/machinectl', 'enable'] + machines
|
||||||
|
run(cmd)
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
def run_module():
|
||||||
|
# define available arguments/parameters a user can pass to the module
|
||||||
|
module_args = dict(
|
||||||
|
state=dict(type='str', required=True),
|
||||||
|
machines=dict(type='list', required=True),
|
||||||
|
tls_machine=dict(type='str', required=False),
|
||||||
|
)
|
||||||
|
|
||||||
|
# seed the result dict in the object
|
||||||
|
# we primarily care about changed and state
|
||||||
|
# changed is if this module effectively modified the target
|
||||||
|
# state will include any data that you want your module to pass back
|
||||||
|
# for consumption, for example, in a subsequent task
|
||||||
|
result = dict(
|
||||||
|
changed=False,
|
||||||
|
message=''
|
||||||
|
)
|
||||||
|
|
||||||
|
# the AnsibleModule object will be our abstraction working with Ansible
|
||||||
|
# this includes instantiation, a couple of common attr would be the
|
||||||
|
# args/params passed to the execution, as well as if the module
|
||||||
|
# supports check mode
|
||||||
|
module = AnsibleModule(
|
||||||
|
argument_spec=module_args,
|
||||||
|
supports_check_mode=True
|
||||||
|
)
|
||||||
|
# if the user is working with this module in only check mode we do not
|
||||||
|
# want to make any changes to the environment, just return the current
|
||||||
|
# state with no modifications
|
||||||
|
if module.check_mode:
|
||||||
|
module.exit_json(**result)
|
||||||
|
|
||||||
|
# manipulate or modify the state as needed (this is going to be the
|
||||||
|
# part where your module will do what it needs to do)
|
||||||
|
machines = module.params['machines']
|
||||||
|
tls_machine = module.params.get('tls_machine')
|
||||||
|
if module.params['state'] == 'stopped':
|
||||||
|
if tls_machine and tls_machine in machines:
|
||||||
|
machines.remove(tls_machine)
|
||||||
|
bus = SystemBus()
|
||||||
|
result['changed'], errors = stop(bus, machines)
|
||||||
|
if errors:
|
||||||
|
errors = '\n\n'.join(errors)
|
||||||
|
module.fail_json(msg=f'Some machines are not stopping correctly {errors}', **result)
|
||||||
|
elif module.params['state'] == 'started':
|
||||||
|
bus = SystemBus()
|
||||||
|
result['changed'], errors = start(bus, machines)
|
||||||
|
if errors:
|
||||||
|
errors = '\n\n'.join(errors)
|
||||||
|
module.fail_json(msg=f'Some machines are not running correctly {errors}', **result)
|
||||||
|
elif module.params['state'] == 'enabled':
|
||||||
|
result['changed'] = enable(machines)
|
||||||
|
else:
|
||||||
|
module.fail_json(msg=f"Unknown state: {module.params['state']}")
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# in the event of a successful module execution, you will want to
|
||||||
|
# simple AnsibleModule.exit_json(), passing the key/value results
|
||||||
|
module.exit_json(**result)
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
run_module()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
2
ansible/machine.yml
Normal file
|
@ -0,0 +1,2 @@
|
||||||
|
- name: "Create informations for {{ item.name }}"
|
||||||
|
ansible.builtin.shell: "/usr/bin/echo {{ vars | modulename(item.name) }} > /var/lib/risotto/machines_informations/{{ item.name }}.image"
|
69
ansible/machines.yml
Normal file
|
@ -0,0 +1,69 @@
|
||||||
|
- name: "Rebuild images"
|
||||||
|
ansible.builtin.shell: "/usr/local/sbin/update_images {{ vars[vars['inventory_hostname']]['general']['tls_server'] }} do_not_start"
|
||||||
|
register: ret
|
||||||
|
failed_when: ret.rc != 0
|
||||||
|
|
||||||
|
- name: "Stop machine TLS"
|
||||||
|
machinectl:
|
||||||
|
state: stopped
|
||||||
|
machines: "{{ vars[vars['inventory_hostname']]['general']['tls_server'] }}"
|
||||||
|
when: vars[vars['inventory_hostname']]['general']['tls_server'] in machines_changed
|
||||||
|
|
||||||
|
- name: "Remove TLS files directory"
|
||||||
|
file:
|
||||||
|
path: "/var/lib/risotto/configurations/{{ vars[vars['inventory_hostname']]['general']['tls_server'] }}"
|
||||||
|
state: absent
|
||||||
|
when: vars[vars['inventory_hostname']]['general']['tls_server'] in machines_changed
|
||||||
|
|
||||||
|
- name: "Copy TLS configuration"
|
||||||
|
unarchive:
|
||||||
|
src: /tmp/new_configurations/machines.tar
|
||||||
|
dest: "/var/lib/risotto/configurations/"
|
||||||
|
include: "{{ vars[vars['inventory_hostname']]['general']['tls_server'] }}"
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
when: vars[vars['inventory_hostname']]['general']['tls_server'] in machines_changed
|
||||||
|
|
||||||
|
- name: "Start machine TLS"
|
||||||
|
machinectl:
|
||||||
|
state: started
|
||||||
|
machines: "{{ vars[vars['inventory_hostname']]['general']['tls_server'] }}"
|
||||||
|
when: vars[vars['inventory_hostname']]['general']['tls_server'] in machines_changed
|
||||||
|
|
||||||
|
- name: "Stop machines with new configuration {{ machines_changed }}"
|
||||||
|
machinectl:
|
||||||
|
state: stopped
|
||||||
|
machines: "{{ machines_changed }}"
|
||||||
|
tls_machine: "{{ vars[vars['inventory_hostname']]['general']['tls_server'] }}"
|
||||||
|
|
||||||
|
- name: "Remove files directory"
|
||||||
|
file:
|
||||||
|
path: "/var/lib/risotto/configurations/{{ item }}"
|
||||||
|
state: absent
|
||||||
|
loop: "{{ machines_changed }}"
|
||||||
|
|
||||||
|
- name: "Copy configuration"
|
||||||
|
unarchive:
|
||||||
|
src: /tmp/new_configurations/machines.tar
|
||||||
|
dest: /var/lib/risotto/configurations/
|
||||||
|
owner: root
|
||||||
|
group: root
|
||||||
|
when: machines_changed
|
||||||
|
|
||||||
|
- name: "Enable machines"
|
||||||
|
machinectl:
|
||||||
|
state: enabled
|
||||||
|
machines: "{{ vars | machineslist(only_name=True) }}"
|
||||||
|
tls_machine: "{{ vars[vars['inventory_hostname']]['general']['tls_server'] }}"
|
||||||
|
|
||||||
|
- name: "Start machines"
|
||||||
|
machinectl:
|
||||||
|
state: started
|
||||||
|
machines: "{{ vars | machineslist(only_name=True) }}"
|
||||||
|
tls_machine: "{{ vars[vars['inventory_hostname']]['general']['tls_server'] }}"
|
||||||
|
|
||||||
|
- name: "Remove compressed files directory"
|
||||||
|
local_action:
|
||||||
|
module: file
|
||||||
|
path: /tmp/new_configurations
|
||||||
|
state: absent
|
1
ansible/password
Symbolic link
|
@ -0,0 +1 @@
|
||||||
|
../password/
|
35
ansible/playbook.yml
Normal file
|
@ -0,0 +1,35 @@
|
||||||
|
---
|
||||||
|
- name: Risotto
|
||||||
|
hosts: all
|
||||||
|
tasks:
|
||||||
|
- name: "Build host files"
|
||||||
|
rougail:
|
||||||
|
hostname: "{{ vars['inventory_hostname'] }}"
|
||||||
|
only_machine: "{{ only_machine }}"
|
||||||
|
configure_host: "{{ configure_host }}"
|
||||||
|
copy_tests: "{{ copy_tests }}"
|
||||||
|
copy_templates: "{{ copy_templates }}"
|
||||||
|
register: build_host
|
||||||
|
|
||||||
|
- name: "Change"
|
||||||
|
ansible.builtin.debug:
|
||||||
|
var: build_host
|
||||||
|
|
||||||
|
- name: "Configure the host"
|
||||||
|
include_tasks: host.yml
|
||||||
|
when: configure_host == true
|
||||||
|
|
||||||
|
- name: "Prepare machine configuration"
|
||||||
|
include_tasks: machine.yml
|
||||||
|
when: item.name in build_host.machines_changed
|
||||||
|
loop: "{{ vars | machineslist(only=only_machine) }}"
|
||||||
|
|
||||||
|
# - name: "Remove images"
|
||||||
|
# include_tasks: remove_image.yml
|
||||||
|
# loop: "{{ vars | machineslist(only=only_machine) }}"
|
||||||
|
# when: delete_old_image == true
|
||||||
|
#
|
||||||
|
- name: "Install and apply configurations"
|
||||||
|
include_tasks: machines.yml
|
||||||
|
vars:
|
||||||
|
machines_changed: "{{ build_host.machines_changed }}"
|
14
ansible/remove_image.yml
Normal file
|
@ -0,0 +1,14 @@
|
||||||
|
- name: "Stop machine {{ item.name }}"
|
||||||
|
machinectl:
|
||||||
|
state: stopped
|
||||||
|
machines: "{{ item.name }}"
|
||||||
|
|
||||||
|
- name: "Remove old machine {{ item.name }}"
|
||||||
|
file:
|
||||||
|
path: /var/lib/machines/{{ item.name }}
|
||||||
|
state: absent
|
||||||
|
|
||||||
|
- name: "Remove old image {{ vars | modulename(item.name) }}"
|
||||||
|
file:
|
||||||
|
path: "/var/lib/risotto/images/{{ vars | modulename(item.name) }}"
|
||||||
|
state: absent
|
27
ansible/sbin/backup_images
Executable file
|
@ -0,0 +1,27 @@
|
||||||
|
#!/bin/bash -ex
|
||||||
|
|
||||||
|
BACKUP_DIR="/root/backup"
|
||||||
|
|
||||||
|
MACHINES=""
|
||||||
|
for nspawn in $(ls /etc/systemd/nspawn/*.nspawn); do
|
||||||
|
nspawn_file=$(basename $nspawn)
|
||||||
|
machine=${nspawn_file%.*}
|
||||||
|
if [ -d "/var/lib/risotto/srv/$machine" ]; then
|
||||||
|
MACHINES="$MACHINES$machine "
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
cd /var/lib/risotto/srv/
|
||||||
|
mkdir -p "$BACKUP_DIR"
|
||||||
|
for machine in $MACHINES; do
|
||||||
|
BACKUP_FILE="$BACKUP_DIR/backup_$machine.tar.bz2"
|
||||||
|
rm -f "$BACKUP_FILE"
|
||||||
|
if [ -f "/var/lib/risotto/configurations/$machine/sbin/risotto_backup" ]; then
|
||||||
|
machinectl -q shell $machine /usr/local/lib/sbin/risotto_backup
|
||||||
|
tar --ignore-failed-read -cJf $BACKUP_FILE $machine/backup
|
||||||
|
elif [ ! -f "/var/lib/risotto/configurations/$machine/no_risotto_backup" ]; then
|
||||||
|
tar --ignore-failed-read -cJf $BACKUP_FILE $machine
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
exit 0
|
192
ansible/sbin/build_image
Executable file
|
@ -0,0 +1,192 @@
|
||||||
|
#!/bin/bash -ex
|
||||||
|
|
||||||
|
IMAGE_NAME=$1
|
||||||
|
|
||||||
|
if [ -z "$1" ]; then
|
||||||
|
ONLY_IF_DATASET_MODIF=false
|
||||||
|
else
|
||||||
|
ONLY_IF_DATASET_MODIF=true
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -z "$IMAGE_NAME" ]; then
|
||||||
|
echo "PAS DE NOM DE MODULE"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
# root dir configuration
|
||||||
|
RISOTTO_DIR="/var/lib/risotto"
|
||||||
|
RISOTTO_IMAGE_DIR="$RISOTTO_DIR/images"
|
||||||
|
# image configuration
|
||||||
|
IMAGE_BASE_RISOTTO_BASE_DIR="$RISOTTO_IMAGE_DIR/image_bases"
|
||||||
|
IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP="$RISOTTO_IMAGE_DIR/tmp/$IMAGE_NAME"
|
||||||
|
IMAGE_NAME_RISOTTO_IMAGE_DIR="$RISOTTO_IMAGE_DIR/$IMAGE_NAME"
|
||||||
|
IMAGE_DIR_RECIPIENT_IMAGE="$RISOTTO_DIR/images_files/$IMAGE_NAME"
|
||||||
|
|
||||||
|
|
||||||
|
rm -f /var/log/risotto/build_image.log
|
||||||
|
mkdir -p "$RISOTTO_IMAGE_DIR" "$RISOTTO_IMAGE_DIR/tmp/"
|
||||||
|
PKG=""
|
||||||
|
BASE_DIR=""
|
||||||
|
for script in $(ls "$IMAGE_DIR_RECIPIENT_IMAGE"/preinstall/*.sh 2> /dev/null); do
|
||||||
|
. "$script"
|
||||||
|
done
|
||||||
|
|
||||||
|
if [ -z "$OS_NAME" ]; then
|
||||||
|
echo "NO OS NAME DEFINED"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
if [ -z "$RELEASEVER" ]; then
|
||||||
|
echo "NO RELEASEVER DEFINED"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
if [ -z "$INSTALL_TOOL" ]; then
|
||||||
|
echo "NO INSTALL TOOL DEFINED"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
BASE_NAME="$OS_NAME-$RELEASEVER"
|
||||||
|
BASE_DIR="$IMAGE_BASE_RISOTTO_BASE_DIR/$BASE_NAME"
|
||||||
|
TMP_BASE_DIR="$IMAGE_BASE_RISOTTO_BASE_DIR/tmp/$BASE_NAME"
|
||||||
|
BASE_PKGS_FILE="$IMAGE_BASE_RISOTTO_BASE_DIR-$BASE_NAME.pkgs"
|
||||||
|
BASE_LOCK="$IMAGE_BASE_RISOTTO_BASE_DIR-$BASE_NAME.build"
|
||||||
|
|
||||||
|
|
||||||
|
function dnf_opt_base() {
|
||||||
|
INSTALL_DIR=$1
|
||||||
|
echo "--setopt=install_weak_deps=False --setopt=fastestmirror=True --nodocs --noplugins --installroot=$INSTALL_DIR --releasever $RELEASEVER"
|
||||||
|
}
|
||||||
|
|
||||||
|
function dnf_opt() {
|
||||||
|
INSTALL_DIR=$1
|
||||||
|
INSTALL_PKG=$2
|
||||||
|
OPT=$(dnf_opt_base "$INSTALL_DIR")
|
||||||
|
echo "$OPT install $INSTALL_PKG"
|
||||||
|
}
|
||||||
|
function new_package_base() {
|
||||||
|
if [ "$INSTALL_TOOL" = "dnf" ]; then
|
||||||
|
OPT=$(dnf_opt "$TMP_BASE_DIR" "$BASE_PKG")
|
||||||
|
dnf --assumeno $OPT | grep ^" " > "$BASE_PKGS_FILE".new
|
||||||
|
else
|
||||||
|
debootstrap --include="$BASE_PKG" --variant=minbase "$RELEASEVER" "$TMP_BASE_DIR" >> /var/log/risotto/build_image.log
|
||||||
|
chroot "$TMP_BASE_DIR" dpkg-query -f '${binary:Package} ${source:Version}\n' -W > "$BASE_PKGS_FILE".new
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
function install_base() {
|
||||||
|
if [ "$INSTALL_TOOL" = "dnf" ]; then
|
||||||
|
OPT=$(dnf_opt "$TMP_BASE_DIR" "$BASE_PKG")
|
||||||
|
dnf --assumeyes $OPT
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
function new_package() {
|
||||||
|
if [ "$INSTALL_TOOL" = "dnf" ]; then
|
||||||
|
OPT=$(dnf_opt_base "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP")
|
||||||
|
set +e
|
||||||
|
dnf --assumeno $OPT update >> /var/log/risotto/build_image.log
|
||||||
|
set -e
|
||||||
|
OPT=$(dnf_opt "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" "$PKG")
|
||||||
|
dnf --assumeno $OPT | grep ^" " > "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs.new
|
||||||
|
else
|
||||||
|
chroot "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" apt update >> /var/log/risotto/build_image.log 2>&1
|
||||||
|
chroot "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" apt install --no-install-recommends --yes $PKG -s 2>/dev/null|grep ^"Inst " > "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs.new
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
function install_pkg() {
|
||||||
|
if [ "$INSTALL_TOOL" = "dnf" ]; then
|
||||||
|
OPT=$(dnf_opt "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" "$PKG")
|
||||||
|
dnf --assumeyes $OPT
|
||||||
|
else
|
||||||
|
if [ "$ONLY_IF_DATASET_MODIF" = true ]; then
|
||||||
|
chroot "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" apt update
|
||||||
|
fi
|
||||||
|
chroot "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" bash -c "export DEBIAN_FRONTEND=noninteractive; apt install --no-install-recommends --yes $PKG"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
if [ ! -f "$BASE_LOCK" ] || [ ! -d "$BASE_DIR" ]; then
|
||||||
|
echo " - reinstallation de l'image de base"
|
||||||
|
new_package_base
|
||||||
|
diff -u "$BASE_PKGS_FILE" "$BASE_PKGS_FILE".new &> /dev/null && NEW_BASE=false || NEW_BASE=true
|
||||||
|
if [ ! -d "$BASE_DIR" ] || [ "$NEW_BASE" = true ]; then
|
||||||
|
mkdir -p "$IMAGE_BASE_RISOTTO_BASE_DIR"
|
||||||
|
rm -rf "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP"
|
||||||
|
install_base
|
||||||
|
if [ -f "$BASE_PKGS_FILE" ]; then
|
||||||
|
mv "$BASE_PKGS_FILE" "$BASE_PKGS_FILE".old
|
||||||
|
fi
|
||||||
|
mv "$BASE_PKGS_FILE".new "$BASE_PKGS_FILE"
|
||||||
|
rm -rf "$BASE_DIR"
|
||||||
|
mv "$TMP_BASE_DIR" "$BASE_DIR"
|
||||||
|
fi
|
||||||
|
touch "$BASE_LOCK"
|
||||||
|
fi
|
||||||
|
|
||||||
|
rm -rf "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP"
|
||||||
|
cp --reflink=auto -a "$BASE_DIR/" "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP"
|
||||||
|
if [ -n "$COPR" ]; then
|
||||||
|
#FIXME signature...
|
||||||
|
mkdir -p "$REPO_DIR"
|
||||||
|
cd "$REPO_DIR"
|
||||||
|
wget -q "$COPR"
|
||||||
|
cd - > /dev/null
|
||||||
|
fi
|
||||||
|
if [ "$FUSION" = true ]; then
|
||||||
|
dnf -y install "https://download1.rpmfusion.org/free/fedora/rpmfusion-free-release-$RELEASEVER.noarch.rpm" --installroot="$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" >> /var/log/risotto/build_image.log
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -f "$IMAGE_NAME_RISOTTO_IMAGE_DIR".base.pkgs ] && [ -f "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs ]; then
|
||||||
|
echo " - différence(s) avec les paquets de base"
|
||||||
|
diff -u "$IMAGE_NAME_RISOTTO_IMAGE_DIR".base.pkgs "$BASE_PKGS_FILE" && INSTALL=false || INSTALL=true
|
||||||
|
[ ! -d "$IMAGE_NAME_RISOTTO_IMAGE_DIR" ] && INSTALL=true
|
||||||
|
else
|
||||||
|
INSTALL=true
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ "$ONLY_IF_DATASET_MODIF" = false ] || [ ! -f "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs ]; then
|
||||||
|
new_package
|
||||||
|
else
|
||||||
|
cp --reflink=auto "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs.new
|
||||||
|
fi
|
||||||
|
if [ "$INSTALL" = false ]; then
|
||||||
|
echo " - différence(s) avec les paquets de l'image"
|
||||||
|
diff -u "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs.new && INSTALL=false || INSTALL=true
|
||||||
|
fi
|
||||||
|
find "$IMAGE_DIR_RECIPIENT_IMAGE" -type f -exec md5sum '{}' \; > "$IMAGE_NAME_RISOTTO_IMAGE_DIR".md5sum.new
|
||||||
|
if [ "$INSTALL" = false ]; then
|
||||||
|
echo " - différence(s) du dataset"
|
||||||
|
diff -u "$IMAGE_NAME_RISOTTO_IMAGE_DIR".md5sum "$IMAGE_NAME_RISOTTO_IMAGE_DIR".md5sum.new && INSTALL=false || INSTALL=true
|
||||||
|
fi
|
||||||
|
if [ "$INSTALL" = true ]; then
|
||||||
|
echo " - installation"
|
||||||
|
if [ -f "$IMAGE_NAME_RISOTTO_IMAGE_DIR".version ]; then
|
||||||
|
VERSION=$(cat "$IMAGE_NAME_RISOTTO_IMAGE_DIR".version)
|
||||||
|
else
|
||||||
|
VERSION=0
|
||||||
|
fi
|
||||||
|
if [ -d "$IMAGE_NAME_RISOTTO_IMAGE_DIR" ]; then
|
||||||
|
cd "$IMAGE_NAME_RISOTTO_IMAGE_DIR"
|
||||||
|
make_changelog "$IMAGE_NAME" "$VERSION" "$OS_NAME" "$RELEASEVER" > "$IMAGE_NAME_RISOTTO_IMAGE_DIR"_"$RELEASEVER"_"$VERSION"_changelog.md
|
||||||
|
cd - > /dev/null
|
||||||
|
fi
|
||||||
|
install_pkg
|
||||||
|
sleep 2
|
||||||
|
|
||||||
|
for script in $(ls $IMAGE_DIR_RECIPIENT_IMAGE/postinstall/*.sh 2> /dev/null); do
|
||||||
|
. "$script"
|
||||||
|
done
|
||||||
|
|
||||||
|
ROOT=$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP make_volatile /etc
|
||||||
|
if [ ! "$?" = 0 ]; then
|
||||||
|
echo "make_volatile failed"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
rm -rf "$IMAGE_NAME_RISOTTO_IMAGE_DIR"
|
||||||
|
mv "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP" "$IMAGE_NAME_RISOTTO_IMAGE_DIR"
|
||||||
|
cp --reflink=auto -f "$BASE_PKGS_FILE" "$IMAGE_NAME_RISOTTO_IMAGE_DIR".base.pkgs
|
||||||
|
mv -f "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs.new "$IMAGE_NAME_RISOTTO_IMAGE_DIR".pkgs
|
||||||
|
mv -f "$IMAGE_NAME_RISOTTO_IMAGE_DIR".md5sum.new "$IMAGE_NAME_RISOTTO_IMAGE_DIR".md5sum
|
||||||
|
VERSION=$((VERSION + 1))
|
||||||
|
echo "$VERSION" > "$IMAGE_NAME_RISOTTO_IMAGE_DIR".version
|
||||||
|
fi
|
||||||
|
rm -rf "$IMAGE_NAME_RISOTTO_IMAGE_DIR_TMP"
|
||||||
|
|
||||||
|
echo " => OK"
|
||||||
|
exit 0
|
36
ansible/sbin/compare_image
Executable file
|
@ -0,0 +1,36 @@
|
||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
SRV=$1
|
||||||
|
if [ -z "$SRV" ]; then
|
||||||
|
echo "usage: $0 machine"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
dirname="/var/lib/risotto/templates/$SRV"
|
||||||
|
if [ ! -d "$dirname" ]; then
|
||||||
|
echo "cannot find $dirname"
|
||||||
|
echo "usage: $0 machine"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
cd $dirname
|
||||||
|
find -type f -not -path "./secrets/*" -not -path "./tmpfiles.d/*" -not -path "./sysusers.d/*" -not -path "./systemd/*" -not -path "./tests/*" -not -path "./etc/pki/*" | while read a; do
|
||||||
|
machine_path="/var/lib/machines/$SRV"
|
||||||
|
cfile="$machine_path/usr/share/factory/$a"
|
||||||
|
if [ -f "$cfile" ]; then
|
||||||
|
diff -u "$dirname/$a" "$cfile"
|
||||||
|
else
|
||||||
|
FIRST_LINE="$(head -n 1 $a)"
|
||||||
|
if [[ "$FIRST_LINE" == "#RISOTTO: file://"* ]]; then
|
||||||
|
other=${FIRST_LINE:16}
|
||||||
|
diff -u "$dirname/$a" "$machine_path$other"
|
||||||
|
elif [[ "$FIRST_LINE" == "#RISOTTO: https://"* ]]; then
|
||||||
|
other=${FIRST_LINE:10}
|
||||||
|
echo $other
|
||||||
|
wget -q $other -O /tmp/template.tmp
|
||||||
|
diff -u "$dirname/$a" /tmp/template.tmp
|
||||||
|
elif [ ! "$FIRST_LINE" = "#RISOTTO: do not compare" ]; then
|
||||||
|
echo "cannot find \"$cfile\" ($dirname/$a)"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
cd - > /dev/null
|
52
ansible/sbin/diagnose
Executable file
|
@ -0,0 +1,52 @@
|
||||||
|
#!/bin/bash -e
|
||||||
|
|
||||||
|
MACHINES=""
|
||||||
|
for nspawn in $(ls /etc/systemd/nspawn/*.nspawn); do
|
||||||
|
nspawn_file=$(basename $nspawn)
|
||||||
|
machine=${nspawn_file%.*}
|
||||||
|
MACHINES="$MACHINES$machine "
|
||||||
|
done
|
||||||
|
STARTED=""
|
||||||
|
DEGRADED=""
|
||||||
|
found=true
|
||||||
|
idx=0
|
||||||
|
while [ $found = true ]; do
|
||||||
|
found=false
|
||||||
|
echo "tentative $idx"
|
||||||
|
for machine in $MACHINES; do
|
||||||
|
if ! echo $STARTED | grep -q " $machine "; then
|
||||||
|
status=$(machinectl -q shell $machine /usr/bin/systemctl is-system-running 2>/dev/null || echo "not started")
|
||||||
|
if echo "$status" | grep -q degraded; then
|
||||||
|
STARTED="$STARTED $machine "
|
||||||
|
DEGRADED="$DEGRADED $machine"
|
||||||
|
elif echo "$status" | grep -q running; then
|
||||||
|
STARTED="$STARTED $machine "
|
||||||
|
else
|
||||||
|
found=true
|
||||||
|
echo "status actuel de $machine : $status"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
sleep 2
|
||||||
|
idx=$((idx+1))
|
||||||
|
if [ $idx = 60 ]; then
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
retcode=0
|
||||||
|
for machine in $MACHINES; do
|
||||||
|
if ! echo "$STARTED" | grep -q " $machine "; then
|
||||||
|
echo
|
||||||
|
echo "========= $machine"
|
||||||
|
machinectl -q shell $machine /usr/bin/systemctl is-system-running 2>/dev/null || systemctl status systemd-nspawn@$machine.service || true
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
echo $DEGRADED
|
||||||
|
for machine in $DEGRADED; do
|
||||||
|
echo
|
||||||
|
echo "========= $machine"
|
||||||
|
machinectl -q shell $machine /usr/bin/systemctl --state=failed --no-legend --no-pager
|
||||||
|
retcode=1
|
||||||
|
done
|
||||||
|
|
||||||
|
exit $retcode
|
182
ansible/sbin/make_changelog
Executable file
|
@ -0,0 +1,182 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
|
||||||
|
import logging
|
||||||
|
from dnf.conf import Conf
|
||||||
|
from dnf.cli.cli import BaseCli, Cli
|
||||||
|
from dnf.cli.output import Output
|
||||||
|
from dnf.cli.option_parser import OptionParser
|
||||||
|
from dnf.i18n import _, ucd
|
||||||
|
from datetime import datetime, timezone
|
||||||
|
from sys import argv
|
||||||
|
from os import getcwd, unlink
|
||||||
|
from os.path import isfile, join
|
||||||
|
from glob import glob
|
||||||
|
from subprocess import run
|
||||||
|
|
||||||
|
|
||||||
|
# List new or removed file
|
||||||
|
def read_dnf_pkg_file(os_name, filename1, filename2):
|
||||||
|
if os_name == 'debian':
|
||||||
|
idx_pkg = 0, 1
|
||||||
|
idx_version = 1, 2
|
||||||
|
header_idx = 0, 0
|
||||||
|
else:
|
||||||
|
idx_pkg = 0, 0
|
||||||
|
idx_version = 2, 2
|
||||||
|
header_idx = 2, 2
|
||||||
|
pass
|
||||||
|
pkgs = {}
|
||||||
|
for fidx, filename in enumerate((filename1, filename2)):
|
||||||
|
if not isfile(filename):
|
||||||
|
continue
|
||||||
|
with open(filename, 'r') as pkgs_fh:
|
||||||
|
for idx, pkg_line in enumerate(pkgs_fh.readlines()):
|
||||||
|
if idx < header_idx[fidx]:
|
||||||
|
# header
|
||||||
|
continue
|
||||||
|
sp_line = pkg_line.strip().split()
|
||||||
|
if len(sp_line) < idx_version[fidx] + 1:
|
||||||
|
continue
|
||||||
|
pkg = sp_line[idx_pkg[fidx]]
|
||||||
|
version = sp_line[idx_version[fidx]]
|
||||||
|
#if pkg in pkgs:
|
||||||
|
# raise Exception(f'package already set {pkg}?')
|
||||||
|
if os_name == 'debian' and version.startswith('('):
|
||||||
|
version = version[1:]
|
||||||
|
pkgs[pkg] = version
|
||||||
|
return pkgs
|
||||||
|
|
||||||
|
|
||||||
|
def list_packages(title, packages, packages_info):
|
||||||
|
print(f'# {title}\n')
|
||||||
|
if not packages:
|
||||||
|
print('*Aucun*')
|
||||||
|
packages = list(packages)
|
||||||
|
packages = sorted(packages)
|
||||||
|
for idx, pkg in enumerate(packages):
|
||||||
|
print(f' - {pkg} ({packages_info[pkg]})')
|
||||||
|
print()
|
||||||
|
|
||||||
|
|
||||||
|
# List updated packages
|
||||||
|
class CustomOutput(Output):
|
||||||
|
def listPkgs(self, *args, **kwargs):
|
||||||
|
# do not display list
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
def format_changelog_markdown(changelog):
|
||||||
|
"""Return changelog formatted as in spec file"""
|
||||||
|
text = '\n'.join([f' {line}' for line in changelog['text'].split('\n')])
|
||||||
|
chlog_str = ' - %s %s\n\n%s\n' % (
|
||||||
|
changelog['timestamp'].strftime("%a %b %d %X %Y"),
|
||||||
|
ucd(changelog['author']),
|
||||||
|
ucd(text))
|
||||||
|
return chlog_str
|
||||||
|
|
||||||
|
|
||||||
|
def print_changelogs_markdown(packages):
|
||||||
|
# group packages by src.rpm to avoid showing duplicate changelogs
|
||||||
|
self = BASE
|
||||||
|
bysrpm = dict()
|
||||||
|
for p in packages:
|
||||||
|
# there are packages without source_name, use name then.
|
||||||
|
bysrpm.setdefault(p.source_name or p.name, []).append(p)
|
||||||
|
for source_name in sorted(bysrpm.keys()):
|
||||||
|
bin_packages = bysrpm[source_name]
|
||||||
|
print('- ' + _("Changelogs for {}").format(', '.join([str(pkg) for pkg in bin_packages])))
|
||||||
|
print()
|
||||||
|
for chl in self.latest_changelogs(bin_packages[0]):
|
||||||
|
print(format_changelog_markdown(chl))
|
||||||
|
|
||||||
|
|
||||||
|
def dnf_update(image_name, releasever):
|
||||||
|
conf = Conf()
|
||||||
|
# obsoletes are already listed
|
||||||
|
conf.obsoletes = False
|
||||||
|
with BaseCli(conf) as base:
|
||||||
|
global BASE
|
||||||
|
BASE = base
|
||||||
|
base.print_changelogs = print_changelogs_markdown
|
||||||
|
custom_output = CustomOutput(base.output.base, base.output.conf)
|
||||||
|
base.output = custom_output
|
||||||
|
cli = Cli(base)
|
||||||
|
image_dir = join(getcwd(), image_name)
|
||||||
|
cli.configure(['--setopt=install_weak_deps=False', '--nodocs', '--noplugins', '--installroot=' + image_dir, '--releasever', releasever, 'check-update', '--changelog'], OptionParser())
|
||||||
|
logger = logging.getLogger("dnf")
|
||||||
|
for h in logger.handlers:
|
||||||
|
logger.removeHandler(h)
|
||||||
|
logger.addHandler(logging.NullHandler())
|
||||||
|
cli.run()
|
||||||
|
|
||||||
|
|
||||||
|
def main(os_name, image_name, old_version, releasever):
|
||||||
|
date = datetime.now(timezone.utc).isoformat()
|
||||||
|
if old_version == 0:
|
||||||
|
title = f"Création de l'image {image_name}"
|
||||||
|
subtitle = f"Les paquets de la première image {image_name} sur base Fedora {releasever}"
|
||||||
|
else:
|
||||||
|
title = f"Nouvelle version de l'image {image_name}"
|
||||||
|
subtitle = f"Différence des paquets de l'image {image_name} sur base Fedora {releasever} entre la version {old_version} et {old_version + 1}"
|
||||||
|
print(f"""+++
|
||||||
|
title = "{title}"
|
||||||
|
description = "{subtitle}"
|
||||||
|
date = {date}
|
||||||
|
updated = {date}
|
||||||
|
draft = false
|
||||||
|
template = "blog/page.html"
|
||||||
|
|
||||||
|
[taxonomies]
|
||||||
|
authors = ["Automate"]
|
||||||
|
|
||||||
|
[extra]
|
||||||
|
lead = "{subtitle}."
|
||||||
|
type = "installe"
|
||||||
|
+++
|
||||||
|
""")
|
||||||
|
new_dict = read_dnf_pkg_file(os_name, f'/var/lib/risotto/images/image_bases-{os_name}-{releasever}.pkgs', f'/var/lib/risotto/images/{image_name}.pkgs.new')
|
||||||
|
new_pkg = new_dict.keys()
|
||||||
|
old_file = f'/var/lib/risotto/images/{image_name}.pkgs'
|
||||||
|
if not old_version or not isfile(old_file):
|
||||||
|
list_packages('Liste des paquets', new_pkg, new_dict)
|
||||||
|
else:
|
||||||
|
ori_dict = read_dnf_pkg_file(os_name, f'/var/lib/risotto/images/{image_name}.base.pkgs', old_file)
|
||||||
|
ori_pkg = ori_dict.keys()
|
||||||
|
list_packages('Les paquets supprimés', ori_pkg - new_pkg, ori_dict)
|
||||||
|
list_packages('Les paquets ajoutés', new_pkg - ori_pkg, new_dict)
|
||||||
|
print('# Les paquets mises à jour\n')
|
||||||
|
if os_name == 'fedora':
|
||||||
|
dnf_update(image_name, releasever)
|
||||||
|
else:
|
||||||
|
for filename in glob('*.deb'):
|
||||||
|
unlink(filename)
|
||||||
|
for package in ori_pkg & new_dict:
|
||||||
|
if ori_dict[package] == new_dict[package]:
|
||||||
|
continue
|
||||||
|
info = run(['apt', 'download', package], capture_output=True)
|
||||||
|
if info.returncode:
|
||||||
|
raise Exception(f'cannot download {package}: {info}')
|
||||||
|
packages = list(glob('*.deb'))
|
||||||
|
packages.sort()
|
||||||
|
for package in packages:
|
||||||
|
info = run(['chroot', '.', 'apt-listchanges', '--which', 'both', '-f', 'text', package], capture_output=True)
|
||||||
|
if info.returncode:
|
||||||
|
raise Exception(f'cannot list changes for {package}: {info}')
|
||||||
|
header = True
|
||||||
|
for line in info.stdout.decode().split('\n'):
|
||||||
|
if not header:
|
||||||
|
print(line)
|
||||||
|
if line.startswith('-----------------------'):
|
||||||
|
header = False
|
||||||
|
print()
|
||||||
|
unlink(package)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
image_name = argv[1]
|
||||||
|
old_version = int(argv[2])
|
||||||
|
os_name = argv[3]
|
||||||
|
releasever = argv[4]
|
||||||
|
main(os_name, image_name, old_version, releasever)
|
76
ansible/sbin/make_volatile
Executable file
|
@ -0,0 +1,76 @@
|
||||||
|
#!/bin/bash -e
|
||||||
|
if [ -z $ROOT ]; then
|
||||||
|
echo "PAS DE ROOT"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
echo "$ROOT"
|
||||||
|
DESTDIR="$ROOT/usr/lib/tmpfiles.d"
|
||||||
|
CONF_DST="/usr/share/factory"
|
||||||
|
EXCLUDES="^($ROOT/etc/passwd|$ROOT/etc/group|$ROOT/etc/.updated|$ROOT/etc/.pwd.lock|$ROOT/etc/systemd/network/dhcp.network|$ROOT/etc/sudoers.d/qemubuild)$"
|
||||||
|
ONLY_COPY="^($ROOT/etc/localtime)$"
|
||||||
|
FORCE_LINKS="^($ROOT/etc/udev/hwdb.bin)$"
|
||||||
|
|
||||||
|
function execute() {
|
||||||
|
chroot $ROOT $@
|
||||||
|
}
|
||||||
|
|
||||||
|
function file_dir_in_tmpfiles() {
|
||||||
|
letter=$1
|
||||||
|
directory=$2
|
||||||
|
local_directory=$(echo $directory|sed "s@^$ROOT@@g")
|
||||||
|
mode=$(execute "/usr/bin/stat" "--format" "%a" "$local_directory" | grep -o "[0-9.]\+")
|
||||||
|
user=$(execute "/usr/bin/stat" "--format" "%U" "$local_directory" | grep -o "[0-9a-zA-Z.-]\+")
|
||||||
|
group=$(execute "/usr/bin/stat" "--format" "%G" "$local_directory" | grep -o "[0-9a-zA-Z.-]\+")
|
||||||
|
echo "$letter $local_directory $mode $user $group - -"
|
||||||
|
}
|
||||||
|
|
||||||
|
function calc_symlink_in_tmpfiles() {
|
||||||
|
dest_name=$1
|
||||||
|
local_dest_name=$2
|
||||||
|
src_file=$(readlink "$dest_name")
|
||||||
|
symlink_in_tmpfiles "$local_dest_name" "$src_file"
|
||||||
|
}
|
||||||
|
|
||||||
|
function symlink_in_tmpfiles() {
|
||||||
|
dest_name=$1
|
||||||
|
src_file=$2
|
||||||
|
echo "L+ $dest_name - - - - $src_file"
|
||||||
|
}
|
||||||
|
|
||||||
|
function main() {
|
||||||
|
dir_config_orig=$1
|
||||||
|
name="${dir_config_orig//\//-}"
|
||||||
|
dir_config_orig=$ROOT$dir_config_orig
|
||||||
|
|
||||||
|
mkdir -p "$DESTDIR"
|
||||||
|
mkdir -p "$ROOTCONF_DST$dir_config_orig"
|
||||||
|
systemd_conf="$DESTDIR/risotto$name.conf"
|
||||||
|
rm -f $systemd_conf
|
||||||
|
shopt -s globstar
|
||||||
|
for src_file in $dir_config_orig/**; do
|
||||||
|
local_src=$(echo $src_file|sed "s@$ROOT@@g")
|
||||||
|
dest_file="$ROOT$CONF_DST$local_src"
|
||||||
|
if [[ "$src_file" =~ $EXCLUDES ]]; then
|
||||||
|
echo "$src_file: exclude" >&2
|
||||||
|
elif [[ -L "$src_file" ]]; then
|
||||||
|
calc_symlink_in_tmpfiles "$src_file" "$local_src" >> $systemd_conf
|
||||||
|
elif [[ "$src_file" =~ $FORCE_LINKS ]]; then
|
||||||
|
symlink_in_tmpfiles "$src_file" "$dest_file" >> $systemd_conf
|
||||||
|
elif [[ -d "$src_file" ]]; then
|
||||||
|
file_dir_in_tmpfiles 'd' "$src_file" >> $systemd_conf
|
||||||
|
[[ ! -d "$dest_file" ]] && mkdir -p "$dest_file"
|
||||||
|
#echo "$src_file: directory ok"
|
||||||
|
else
|
||||||
|
if [[ ! "$src_file" =~ $ONLY_COPY ]]; then
|
||||||
|
file_dir_in_tmpfiles "C" "$src_file" >> $systemd_conf
|
||||||
|
fi
|
||||||
|
[[ -e "$dest_file" ]] && rm -f "$dest_file"
|
||||||
|
# not a symlink... an hardlink
|
||||||
|
ln "$src_file" "$dest_file"
|
||||||
|
#echo "$src_file: file ok"
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
}
|
||||||
|
main "$1"
|
||||||
|
echo "fin"
|
||||||
|
exit 0
|
30
ansible/sbin/test_images
Executable file
|
@ -0,0 +1,30 @@
|
||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
QUIT_ON_ERROR=true
|
||||||
|
# QUIT_ON_ERROR=false
|
||||||
|
CONFIG_DIR="/var/lib/risotto/configurations"
|
||||||
|
INFO_DIR="/var/lib/risotto/machines_informations"
|
||||||
|
TEST_DIR="/var/lib/risotto/tests"
|
||||||
|
TEST_DIR_NAME="tests"
|
||||||
|
|
||||||
|
if [ ! -d /var/lib/risotto/tests/ ]; then
|
||||||
|
echo "no tests directory"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
py_test_option="-s"
|
||||||
|
if [ "$QUIT_ON_ERROR" = true ]; then
|
||||||
|
set -e
|
||||||
|
py_test_option="$py_test_option -x"
|
||||||
|
fi
|
||||||
|
|
||||||
|
for nspawn in $(ls /etc/systemd/nspawn/*.nspawn); do
|
||||||
|
nspawn_file=$(basename $nspawn)
|
||||||
|
machine=${nspawn_file%.*}
|
||||||
|
image=$(cat $INFO_DIR/$machine.image)
|
||||||
|
imagedir=$TEST_DIR/$image
|
||||||
|
machine_test_dir=$CONFIG_DIR/$machine/$TEST_DIR_NAME
|
||||||
|
export MACHINE_TEST_DIR=$machine_test_dir
|
||||||
|
echo "- $machine"
|
||||||
|
py.test-3 $py_test_option "$imagedir"
|
||||||
|
done
|
87
ansible/sbin/update_images
Executable file
|
@ -0,0 +1,87 @@
|
||||||
|
#!/bin/bash -e
|
||||||
|
|
||||||
|
TLS_SERVER=$1
|
||||||
|
if [ -z "$TLS_SERVER" ]; then
|
||||||
|
echo "$0 nom_tls_server"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
DO_NOT_START=$2
|
||||||
|
REBOOT_EVERY_MONDAY=$3
|
||||||
|
# root dir configuration
|
||||||
|
RISOTTO_DIR="/var/lib/risotto"
|
||||||
|
RISOTTO_IMAGE_DIR="$RISOTTO_DIR/images"
|
||||||
|
# image configuration
|
||||||
|
IMAGE_BASE_RISOTTO_BASE_DIR="$RISOTTO_IMAGE_DIR/image_bases"
|
||||||
|
|
||||||
|
if [ -z "$1" ]; then
|
||||||
|
rm -f $IMAGE_BASE_RISOTTO_BASE_DIR*.build
|
||||||
|
fi
|
||||||
|
|
||||||
|
mkdir -p /var/log/risotto
|
||||||
|
|
||||||
|
ls /var/lib/risotto/images_files/ | while read image; do
|
||||||
|
if [ -d /var/lib/risotto/images_files/"$image" ]; then
|
||||||
|
echo
|
||||||
|
echo "Install image $image" | tee -a /var/log/risotto/update_images.log
|
||||||
|
/usr/local/sbin/build_image "$image" || echo "PROBLEME" | tee -a /var/log/risotto/update_images.log
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
idx=0
|
||||||
|
if [ -z "$DO_NOT_START" ]; then
|
||||||
|
machinectl reboot "$TLS_SERVER" || machinectl start "$TLS_SERVER"
|
||||||
|
while true; do
|
||||||
|
status=$(machinectl -q shell "$TLS_SERVER" /usr/bin/systemctl is-system-running 2>/dev/null || echo "not started")
|
||||||
|
if echo "$status" | grep -q degraded || echo "$status" | grep -q running; then
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
idx=$((idx+1))
|
||||||
|
if [ $idx = 60 ]; then
|
||||||
|
echo "le serveur $TLS_SERVER n'a pas encore redémarré"
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
sleep 2
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
|
||||||
|
MACHINES=""
|
||||||
|
for nspawn in $(ls /etc/systemd/nspawn/*.nspawn); do
|
||||||
|
nspawn_file=$(basename "$nspawn")
|
||||||
|
machine=${nspawn_file%.*}
|
||||||
|
MACHINES="$MACHINES$machine "
|
||||||
|
MACHINE_MACHINES_DIR="/var/lib/machines/$machine"
|
||||||
|
IMAGE_NAME_RISOTTO_IMAGE_NAME="$(cat $RISOTTO_DIR/machines_informations/$machine.image)"
|
||||||
|
MACHINE_INFO="$RISOTTO_DIR/machines_informations/"
|
||||||
|
VERSION_MACHINE="$MACHINE_INFO/$machine.version"
|
||||||
|
if [ -n "$REBOOT_EVERY_MONDAY" ] && [ "$(date +%u)" = 1 ]; then
|
||||||
|
# update TLS certificate every monday, so stop container
|
||||||
|
machinectl stop "$machine" 2> /dev/null || true
|
||||||
|
while true; do
|
||||||
|
machinectl status "$machine" > /dev/null 2>&1 || break
|
||||||
|
sleep 1
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
if [ ! -d "$MACHINE_MACHINES_DIR/etc" ]; then
|
||||||
|
rm -f "$VERSION_MACHINE"
|
||||||
|
fi
|
||||||
|
diff -q "$RISOTTO_IMAGE_DIR/$IMAGE_NAME_RISOTTO_IMAGE_NAME".version "$VERSION_MACHINE" &> /dev/null || (
|
||||||
|
echo "Reinstall machine $machine"
|
||||||
|
machinectl stop "$machine" 2> /dev/null || true
|
||||||
|
while true; do
|
||||||
|
machinectl status "$machine" > /dev/null 2>&1 || break
|
||||||
|
sleep 1
|
||||||
|
done
|
||||||
|
rm -rf "$MACHINE_MACHINES_DIR"
|
||||||
|
mkdir "$MACHINE_MACHINES_DIR"
|
||||||
|
cp -a --reflink=auto "$RISOTTO_IMAGE_DIR/$IMAGE_NAME_RISOTTO_IMAGE_NAME/"* "$MACHINE_MACHINES_DIR"
|
||||||
|
cp -a --reflink=auto "$RISOTTO_IMAGE_DIR/$IMAGE_NAME_RISOTTO_IMAGE_NAME".version "$VERSION_MACHINE"
|
||||||
|
)
|
||||||
|
done
|
||||||
|
if [ -z "$DO_NOT_START" ]; then
|
||||||
|
echo "start $MACHINES"
|
||||||
|
machinectl start $MACHINES
|
||||||
|
sleep 5
|
||||||
|
journalctl -n 100 --no-pager
|
||||||
|
diagnose
|
||||||
|
fi
|
||||||
|
exit 0
|
35
doc/README.md
Normal file
|
@ -0,0 +1,35 @@
|
||||||
|
![Logo Risotto](../logo.png "logo risotto")
|
||||||
|
|
||||||
|
# Risotto
|
||||||
|
|
||||||
|
## A dataset
|
||||||
|
|
||||||
|
- [Dataset example](dataset_example/dataset.md)
|
||||||
|
- [Official dataset](https://cloud.silique.fr/gitea/risotto/dataset/src/branch/main/seed/README.md)
|
||||||
|
|
||||||
|
## Infrastructure
|
||||||
|
|
||||||
|
- [Infrastructure](infrastructure.md)
|
||||||
|
- [Examples](dataset_example/infrastructure.md)
|
||||||
|
|
||||||
|
## risotto.conf
|
||||||
|
|
||||||
|
```toml
|
||||||
|
[directories]
|
||||||
|
datasets = ['<path_to_dataset_base>/seed']
|
||||||
|
dest = 'installations'
|
||||||
|
dest_templates = 'templates'
|
||||||
|
|
||||||
|
[cert_authority]
|
||||||
|
email = '<email>'
|
||||||
|
country = 'FR'
|
||||||
|
locality = 'Dijon'
|
||||||
|
state = 'France'
|
||||||
|
org_name = 'Silique'
|
||||||
|
org_unit_name = 'Cloud'
|
||||||
|
```
|
||||||
|
|
||||||
|
## Usage
|
||||||
|
|
||||||
|
![Schema](schema.png "Schéma")
|
||||||
|
|
405
doc/authentification.svg
Normal file
|
@ -0,0 +1,405 @@
|
||||||
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||||
|
<!-- Created with Inkscape (http://www.inkscape.org/) -->
|
||||||
|
|
||||||
|
<svg
|
||||||
|
width="210mm"
|
||||||
|
height="297mm"
|
||||||
|
viewBox="0 0 210 297"
|
||||||
|
version="1.1"
|
||||||
|
id="svg5"
|
||||||
|
inkscape:version="1.2.1 (9c6d41e410, 2022-07-14)"
|
||||||
|
sodipodi:docname="authentification.svg"
|
||||||
|
xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
|
||||||
|
xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
|
||||||
|
xmlns="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:svg="http://www.w3.org/2000/svg">
|
||||||
|
<sodipodi:namedview
|
||||||
|
id="namedview7"
|
||||||
|
pagecolor="#ffffff"
|
||||||
|
bordercolor="#666666"
|
||||||
|
borderopacity="1.0"
|
||||||
|
inkscape:pageshadow="2"
|
||||||
|
inkscape:pageopacity="0.0"
|
||||||
|
inkscape:pagecheckerboard="0"
|
||||||
|
inkscape:document-units="mm"
|
||||||
|
showgrid="false"
|
||||||
|
inkscape:zoom="1.557211"
|
||||||
|
inkscape:cx="355.12207"
|
||||||
|
inkscape:cy="181.73517"
|
||||||
|
inkscape:window-width="2048"
|
||||||
|
inkscape:window-height="1083"
|
||||||
|
inkscape:window-x="0"
|
||||||
|
inkscape:window-y="0"
|
||||||
|
inkscape:window-maximized="1"
|
||||||
|
inkscape:current-layer="layer1"
|
||||||
|
inkscape:showpageshadow="2"
|
||||||
|
inkscape:deskcolor="#d1d1d1" />
|
||||||
|
<defs
|
||||||
|
id="defs2">
|
||||||
|
<rect
|
||||||
|
x="465.27745"
|
||||||
|
y="390.53444"
|
||||||
|
width="155.19784"
|
||||||
|
height="121.34324"
|
||||||
|
id="rect9339" />
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker47493"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Sstart"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(0.2,0,0,0.2,1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path47491" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow2Send"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow2Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 3.4652294 2.5981128"
|
||||||
|
markerWidth="3.465229"
|
||||||
|
markerHeight="2.5981126"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.3,0,0,-0.3,0.69,0)"
|
||||||
|
d="M 8.7185878,4.0337352 -2.2072895,0.01601326 8.7185884,-4.0017078 c -1.7454984,2.3720609 -1.7354408,5.6174519 -6e-7,8.035443 z"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:0.625;stroke-linejoin:round"
|
||||||
|
id="path45715" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow2Sstart"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow2Sstart"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 3.4652294 2.5981128"
|
||||||
|
markerWidth="3.465229"
|
||||||
|
markerHeight="2.5981126"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(0.3,0,0,0.3,-0.69,0)"
|
||||||
|
d="M 8.7185878,4.0337352 -2.2072895,0.01601326 8.7185884,-4.0017078 c -1.7454984,2.3720609 -1.7354408,5.6174519 -6e-7,8.035443 z"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:0.625;stroke-linejoin:round"
|
||||||
|
id="path45712" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible;"
|
||||||
|
id="Arrow1Send"
|
||||||
|
refX="0.0"
|
||||||
|
refY="0.0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="scale(0.2) rotate(180) translate(6,0)"
|
||||||
|
style="fill-rule:evenodd;fill:context-stroke;stroke:context-stroke;stroke-width:1.0pt;"
|
||||||
|
d="M 0.0,0.0 L 5.0,-5.0 L -12.5,0.0 L 5.0,5.0 L 0.0,0.0 z "
|
||||||
|
id="path45697" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow1Sstart"
|
||||||
|
refX="0.0"
|
||||||
|
refY="0.0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Sstart"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="scale(0.2) translate(6,0)"
|
||||||
|
style="fill-rule:evenodd;fill:context-stroke;stroke:context-stroke;stroke-width:1.0pt"
|
||||||
|
d="M 0.0,0.0 L 5.0,-5.0 L -12.5,0.0 L 5.0,5.0 L 0.0,0.0 z "
|
||||||
|
id="path45694" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible;"
|
||||||
|
id="Arrow1Lend"
|
||||||
|
refX="0.0"
|
||||||
|
refY="0.0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Lend"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="scale(0.8) rotate(180) translate(12.5,0)"
|
||||||
|
style="fill-rule:evenodd;fill:context-stroke;stroke:context-stroke;stroke-width:1.0pt;"
|
||||||
|
d="M 0.0,0.0 L 5.0,-5.0 L -12.5,0.0 L 5.0,5.0 L 0.0,0.0 z "
|
||||||
|
id="path45685" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 4.4434635 2.539122"
|
||||||
|
markerWidth="4.4434633"
|
||||||
|
markerHeight="2.5391221"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-1"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 4.4434635 2.539122"
|
||||||
|
markerWidth="4.4434633"
|
||||||
|
markerHeight="2.5391221"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-8" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-9"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-2" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-9-3"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-2-7" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6-9"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2-2" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow2Sstart-8"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow2Sstart"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(0.3,0,0,0.3,-0.69,0)"
|
||||||
|
d="M 8.7185878,4.0337352 -2.2072895,0.01601326 8.7185884,-4.0017078 c -1.7454984,2.3720609 -1.7354408,5.6174519 -6e-7,8.035443 z"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:0.625;stroke-linejoin:round"
|
||||||
|
id="path45712-9" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6-9-7"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2-2-3" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6-4"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 4.4434635 2.539122"
|
||||||
|
markerWidth="4.4434633"
|
||||||
|
markerHeight="2.5391221"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2-7" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow2Send-3"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow2Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 3.4652294 2.5981128"
|
||||||
|
markerWidth="3.465229"
|
||||||
|
markerHeight="2.5981126"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.3,0,0,-0.3,0.69,0)"
|
||||||
|
d="M 8.7185878,4.0337352 -2.2072895,0.01601326 8.7185884,-4.0017078 c -1.7454984,2.3720609 -1.7354408,5.6174519 -6e-7,8.035443 z"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:0.625;stroke-linejoin:round"
|
||||||
|
id="path45715-6" />
|
||||||
|
</marker>
|
||||||
|
</defs>
|
||||||
|
<g
|
||||||
|
inkscape:label="Calque 1"
|
||||||
|
inkscape:groupmode="layer"
|
||||||
|
id="layer1">
|
||||||
|
<ellipse
|
||||||
|
style="fill:#000000;fill-opacity:1;fill-rule:evenodd;stroke:#000000;stroke-width:5.86795;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path986"
|
||||||
|
cx="62.406502"
|
||||||
|
cy="64.119804"
|
||||||
|
rx="4.5660253"
|
||||||
|
ry="4.5660257" />
|
||||||
|
<ellipse
|
||||||
|
style="fill:#000000;fill-opacity:1;fill-rule:evenodd;stroke:#000000;stroke-width:5.86795;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path986-6-5"
|
||||||
|
cx="142.10091"
|
||||||
|
cy="64.120003"
|
||||||
|
rx="4.5660253"
|
||||||
|
ry="4.5660257" />
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:4.93889px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;stroke-width:0.264583"
|
||||||
|
x="63.8083"
|
||||||
|
y="77.278412"
|
||||||
|
id="text3135"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="63.8083"
|
||||||
|
y="77.278412"
|
||||||
|
id="tspan10911">IMAP</tspan></text>
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:4.93889px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;stroke-width:0.264583"
|
||||||
|
x="135.16226"
|
||||||
|
y="78.779442"
|
||||||
|
id="text3135-7-6"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;stroke-width:0.264583"
|
||||||
|
x="135.16226"
|
||||||
|
y="78.779442"
|
||||||
|
id="tspan10911-3-2">LDAP</tspan></text>
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="M 75.106998,57.935664 H 122.14408"
|
||||||
|
id="path46175-02"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send-3)"
|
||||||
|
d="M 75.340285,68.861719 H 122.37734"
|
||||||
|
id="path46175-02-7"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
inkscape:connector-curvature="0"
|
||||||
|
style="opacity:0.98;fill:#000000;fill-opacity:1;stroke:none;stroke-width:0.175296;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-dashoffset:0;stroke-opacity:1"
|
||||||
|
d="m 94.078762,40.916652 c -0.0389,2.57e-4 -0.0774,5.85e-4 -0.11611,0.0022 -1.77767,0.0113 -3.37259,1.592182 -3.37712,3.374607 -0.0202,0.420172 0.007,0.840654 0,1.260955 -0.28547,-7e-6 -0.57094,-7e-6 -0.85649,0 0.023,1.846787 -0.0461,3.697844 0.036,5.54194 0.17721,1.1875 1.24351,2.26136 2.48695,2.203553 1.36149,-0.0022 2.72716,0.04211 4.086269,-0.0275 1.275754,-0.219817 2.171678,-1.529827 2.074938,-2.797815 0.0144,-1.639617 0,-3.279313 0.007,-4.918966 -0.284237,-0.0072 -0.568484,0.005 -0.852724,-0.0036 0.0216,-0.998381 0.0684,-2.089696 -0.500617,-2.955111 -0.615417,-1.026965 -1.788466,-1.688137 -2.987566,-1.680443 z m 0.0165,1.425752 c 1.01001,0.01389 2.00786,0.850284 1.97878,1.902665 0.0202,0.436339 0.0331,0.872937 0.0425,1.309642 -1.35875,-5.85e-4 -2.71751,0.0022 -4.07619,-0.0022 0.007,-0.683077 -0.17908,-1.429948 0.19471,-2.044983 0.33945,-0.651636 1.01793,-1.150287 1.76284,-1.163575 0.0324,-0.0015 0.0648,-0.0022 0.0974,-0.0015 z"
|
||||||
|
id="path3355" />
|
||||||
|
<rect
|
||||||
|
style="fill:none;fill-rule:evenodd;stroke:#040000;stroke-width:1.27229;stroke-linecap:square;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="rect8947"
|
||||||
|
width="29.594231"
|
||||||
|
height="6.274775"
|
||||||
|
x="79.703773"
|
||||||
|
y="82.478172" />
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:11.9191px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;stroke-width:0.15963"
|
||||||
|
x="80.080597"
|
||||||
|
y="91.894714"
|
||||||
|
id="text9263"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
id="tspan9261"
|
||||||
|
style="font-style:normal;font-variant:normal;font-weight:bold;font-stretch:normal;font-size:11.9191px;font-family:'Abyssinica SIL';-inkscape-font-specification:'Abyssinica SIL Bold';stroke-width:0.15963"
|
||||||
|
x="80.080597"
|
||||||
|
y="91.894714">*****</tspan></text>
|
||||||
|
<rect
|
||||||
|
style="fill:none;fill-rule:evenodd;stroke:#040000;stroke-width:1.27229;stroke-linecap:square;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="rect8947-5"
|
||||||
|
width="29.594231"
|
||||||
|
height="6.274775"
|
||||||
|
x="79.942833"
|
||||||
|
y="74.1101" />
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:7.05556px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;stroke-width:0.15963"
|
||||||
|
x="80.848824"
|
||||||
|
y="79.293304"
|
||||||
|
id="text9263-6"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
id="tspan9261-2"
|
||||||
|
style="font-style:normal;font-variant:normal;font-weight:bold;font-stretch:normal;font-size:7.05556px;font-family:'Abyssinica SIL';-inkscape-font-specification:'Abyssinica SIL Bold';stroke-width:0.15963"
|
||||||
|
x="80.848824"
|
||||||
|
y="79.293304">domaine</tspan></text>
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
transform="scale(0.26458333)"
|
||||||
|
id="text9337"
|
||||||
|
style="font-size:40px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;white-space:pre;shape-inside:url(#rect9339);display:inline" />
|
||||||
|
</g>
|
||||||
|
</svg>
|
After Width: | Height: | Size: 15 KiB |
87
doc/dataset_example/dataset.md
Normal file
|
@ -0,0 +1,87 @@
|
||||||
|
# Risotto dataset simple examples
|
||||||
|
|
||||||
|
This tutorial aims to show how create a dataset to deploy a [Caddy](https://caddyserver.com/) server via Risotto.
|
||||||
|
|
||||||
|
Attention it has no other virtues than to be educational. It is not intended for production use.
|
||||||
|
|
||||||
|
See [Rougail documentation for more details about dictionaries, templates and patches](https://cloud.silique.fr/gitea/risotto/rougail/src/branch/main/doc/README.md).
|
||||||
|
|
||||||
|
The project can be divided into three application services:
|
||||||
|
|
||||||
|
- caddy-common: an application service containing the information common to the two other application services
|
||||||
|
- caddy-https: a standalone http/https server
|
||||||
|
- caddy-https-rp: a https only server served behind a reverse proxy
|
||||||
|
|
||||||
|
## caddy-common
|
||||||
|
|
||||||
|
Start by creating the project tree:
|
||||||
|
|
||||||
|
```
|
||||||
|
seed/caddy-common/
|
||||||
|
├── dictionaries
|
||||||
|
├── templates
|
||||||
|
└── manual
|
||||||
|
└── image
|
||||||
|
└── preinstall
|
||||||
|
```
|
||||||
|
|
||||||
|
Then describe the application service in [seed/caddy-common/applicationservice.yml](seed/caddy-common/applicationservice.yml).
|
||||||
|
|
||||||
|
Also a dictionary [seed/caddy-common/dictionaries/20-caddy.yml](seed/caddy-common/dictionaries/20-caddy.yml) with
|
||||||
|
|
||||||
|
- the activation of the caddy service in the "multi-user" target. This service needs some templates:
|
||||||
|
|
||||||
|
- the main configuration's [/etc/caddy/Caddyfile](seed/caddy-common/templates/Caddyfile) to include other /etc/caddy/Caddyfile.d/\*.caddyfile
|
||||||
|
- /etc/caddy/Caddyfile.d/risotto.caddyfile with appropriate configuration (this file is not part of this application service)
|
||||||
|
- a [sysusers](https://www.freedesktop.org/software/systemd/man/sysusers.d.html) file [/sysusers.d/0caddy.conf](seed/caddy-common/templates/sysuser-caddy.conf) to create the system user "caddy"
|
||||||
|
- a [tmpfiles](https://www.freedesktop.org/software/systemd/man/tmpfiles.d.html) file [/tmpfiles.d/0caddy.conf](seed/caddy-common/templates/tmpfile-caddy.conf) to create the directory "caddy_root_directory" and volatile directory "/var/lib/caddy"
|
||||||
|
|
||||||
|
- a family "caddy" (Caddy web server) with a filename variable "caddy_root_directory" (The root path of the site) with default value "/srv/caddy".
|
||||||
|
|
||||||
|
Finally, create a script to build the image with the caddy package: [seed/caddy-common/manual/image/preinstall/caddy.sh](seed/caddy-common/manual/image/preinstall/caddy.sh).
|
||||||
|
|
||||||
|
## caddy-https
|
||||||
|
|
||||||
|
Start by creating the project tree:
|
||||||
|
|
||||||
|
```
|
||||||
|
seed/caddy-https-rp/
|
||||||
|
├── dictionaries
|
||||||
|
└── templates
|
||||||
|
```
|
||||||
|
|
||||||
|
Then describe the application service in [seed/caddy-https/applicationservice.yml](seed/caddy-https/applicationservice.yml) with OS and caddy-common dependencies.
|
||||||
|
|
||||||
|
Also create a dictionary [seed/caddy-https/dictionaries/25-caddy.yml](seed/caddy-https/dictionaries/25-caddy.yml) to define the variables:
|
||||||
|
|
||||||
|
- caddy_domain: the domain where Caddy should listen to
|
||||||
|
- caddy_ca_file, caddy_crt_file and caddy_key_file: certificat for this domain
|
||||||
|
- redefine the variable incoming_ports to open the ports 80 and 443
|
||||||
|
|
||||||
|
And new templates:
|
||||||
|
|
||||||
|
- [seed/caddy-https/templates/risotto.caddyfile](seed/caddy-https/templates/risotto.caddyfile)
|
||||||
|
- [seed/caddy-https/templates/ca_HTTP.crt](seed/caddy-https/templates/ca_HTTP.crt)
|
||||||
|
- [seed/caddy-https/templates/caddy.key](seed/caddy-https/templates/caddy.key)
|
||||||
|
- [seed/caddy-https/templates/caddy.crt](seed/caddy-https/templates/caddy.crt)
|
||||||
|
|
||||||
|
## caddy-https-rp
|
||||||
|
|
||||||
|
Start by creating the project tree:
|
||||||
|
|
||||||
|
```
|
||||||
|
seed/caddy-https-rp/
|
||||||
|
├── dictionaries
|
||||||
|
├── patches
|
||||||
|
└── templates
|
||||||
|
```
|
||||||
|
|
||||||
|
Then describe the application service in [seed/caddy-https-rp/applicationservice.yml](seed/caddy-https-rp/applicationservice.yml) with OS, caddy-common and reverse-proxy-client dependencies.
|
||||||
|
|
||||||
|
By default, reverse proxy certificate is only readable by "root" user. In the dictionary [seed/caddy-https-rp/dictionaries/25-caddy.yml](seed/caddy-https-rp/dictionaries/25-caddy.yml) we change the user to "caddy".
|
||||||
|
|
||||||
|
And add Caddy configuration's file [seed/caddy-https-rp/templates/risotto.caddyfile](seed/caddy-https-rp/templates/risotto.caddyfile).
|
||||||
|
This template use mainly variables defined in reverse-proxy application service.
|
||||||
|
|
||||||
|
Finally add a patch to modify Caddyfile to not starts Caddy in port 80: [seed/caddy-https-rp/patches/Caddyfile.patch](seed/caddy-https-rp/patches/Caddyfile.patch).
|
||||||
|
Patches should only use if a template file is define in an other dataset. You should instead add a condition in the template. But for educational reasons we made a patch in this case.
|
38
doc/dataset_example/infrastructure.md
Normal file
|
@ -0,0 +1,38 @@
|
||||||
|
# Examples
|
||||||
|
|
||||||
|
## Caddy as HTTPS server
|
||||||
|
|
||||||
|
The [servers.yml](servers.caddy-https.yml):
|
||||||
|
|
||||||
|
- we create only the zone "external"
|
||||||
|
- we create a module "caddy"
|
||||||
|
- we define an host "host.example.net":
|
||||||
|
|
||||||
|
- servers are containerized with [machined](https://freedesktop.org/wiki/Software/systemd/machined/), so service application is "host-systemd-machined"
|
||||||
|
- the provide application service is "provider-systemd-machined"
|
||||||
|
|
||||||
|
- we define a server "caddy"
|
||||||
|
|
||||||
|
## Caddy behind a Nginx reverse proxy
|
||||||
|
|
||||||
|
The [servers.yml](servers.caddy-https-rp.yml):
|
||||||
|
|
||||||
|
- we create the zone "external" and a zone "revprox" between "revprox" and "caddy" servers
|
||||||
|
- we create three module:
|
||||||
|
|
||||||
|
- "revprox": the reverse proxy (with "letsencrypt" application service if needed)
|
||||||
|
- "nsd": to manage local DNS name
|
||||||
|
- "caddy"
|
||||||
|
|
||||||
|
- we define an host "host.example.net":
|
||||||
|
|
||||||
|
- servers are containerized with [machined](https://freedesktop.org/wiki/Software/systemd/machined/), so service application is "host-systemd-machined"
|
||||||
|
- the provide application service is "provider-systemd-machined"
|
||||||
|
|
||||||
|
- we define servers:
|
||||||
|
|
||||||
|
- revprox in zones "external" and "revprox"
|
||||||
|
- nsd in zone "revprox"
|
||||||
|
- caddy in zone "revprox"
|
||||||
|
|
||||||
|
You must add a index.html file in "/var/lib/risotto/srv/caddy.in.example.net/caddy/".
|
|
@ -0,0 +1,2 @@
|
||||||
|
format: '0.1'
|
||||||
|
description: Caddy's common files
|
|
@ -0,0 +1,25 @@
|
||||||
|
services:
|
||||||
|
- service:
|
||||||
|
- name: caddy
|
||||||
|
target: multi-user
|
||||||
|
file:
|
||||||
|
- text: /etc/caddy/Caddyfile
|
||||||
|
engine: 'none'
|
||||||
|
- text: /etc/caddy/Caddyfile.d/risotto.caddyfile
|
||||||
|
- text: /sysusers.d/0caddy.conf
|
||||||
|
source: sysuser-caddy.conf
|
||||||
|
engine: 'none'
|
||||||
|
- text: /tmpfiles.d/0caddy.conf
|
||||||
|
source: tmpfile-caddy.conf
|
||||||
|
engine: 'none'
|
||||||
|
variables:
|
||||||
|
- family:
|
||||||
|
- name: caddy
|
||||||
|
description: Caddy web server
|
||||||
|
variables:
|
||||||
|
- variable:
|
||||||
|
- name: caddy_root_directory
|
||||||
|
type: filename
|
||||||
|
description: The root path of the site
|
||||||
|
value:
|
||||||
|
- text: /srv/caddy
|
|
@ -0,0 +1 @@
|
||||||
|
PKG="$PKG caddy"
|
43
doc/dataset_example/seed/caddy-common/templates/Caddyfile
Normal file
|
@ -0,0 +1,43 @@
|
||||||
|
# The Caddyfile is an easy way to configure your Caddy web server.
|
||||||
|
#
|
||||||
|
# https://caddyserver.com/docs/caddyfile
|
||||||
|
|
||||||
|
#>GNUNUX
|
||||||
|
# Global options
|
||||||
|
{
|
||||||
|
# remove administration tool
|
||||||
|
admin off
|
||||||
|
}
|
||||||
|
#<GNUNUX
|
||||||
|
|
||||||
|
# The configuration below serves a welcome page over HTTP on port 80. To use
|
||||||
|
# your own domain name with automatic HTTPS, ensure your A/AAAA DNS record is
|
||||||
|
# pointing to this machine's public IP, then replace `http://` with your domain
|
||||||
|
# name. Refer to the documentation for full instructions on the address
|
||||||
|
# specification.
|
||||||
|
#
|
||||||
|
# https://caddyserver.com/docs/caddyfile/concepts#addresses
|
||||||
|
#GNUNUX http:// {
|
||||||
|
|
||||||
|
# Set this path to your site's directory.
|
||||||
|
#GNUNUX root * /usr/share/caddy
|
||||||
|
|
||||||
|
# Enable the static file server.
|
||||||
|
#GNUNUX file_server
|
||||||
|
|
||||||
|
# Another common task is to set up a reverse proxy:
|
||||||
|
# reverse_proxy localhost:8080
|
||||||
|
|
||||||
|
# Or serve a PHP site through php-fpm:
|
||||||
|
# php_fastcgi localhost:9000
|
||||||
|
|
||||||
|
# Refer to the directive documentation for more options.
|
||||||
|
# https://caddyserver.com/docs/caddyfile/directives
|
||||||
|
|
||||||
|
#GNUNUX}
|
||||||
|
|
||||||
|
|
||||||
|
# As an alternative to editing the above site block, you can add your own site
|
||||||
|
# block files in the Caddyfile.d directory, and they will be included as long
|
||||||
|
# as they use the .caddyfile extension.
|
||||||
|
import Caddyfile.d/*.caddyfile
|
|
@ -0,0 +1,2 @@
|
||||||
|
g caddy 998 -
|
||||||
|
u caddy 998:998 "Caddy web server" /var/lib/caddy /sbin/nologin
|
|
@ -0,0 +1,2 @@
|
||||||
|
d /var/lib/caddy 750 caddy caddy - -
|
||||||
|
d %%caddy_root_directory 750 root caddy - -
|
|
@ -0,0 +1,6 @@
|
||||||
|
format: '0.1'
|
||||||
|
description: Caddy
|
||||||
|
depends:
|
||||||
|
- base-fedora-36
|
||||||
|
- reverse-proxy-client
|
||||||
|
- caddy-common
|
|
@ -0,0 +1,9 @@
|
||||||
|
variables:
|
||||||
|
- family:
|
||||||
|
- name: revprox
|
||||||
|
variables:
|
||||||
|
- variable:
|
||||||
|
- name: revprox_client_cert_owner
|
||||||
|
redefine: true
|
||||||
|
value:
|
||||||
|
- text: caddy
|
|
@ -0,0 +1 @@
|
||||||
|
PKG="$PKG caddy"
|
|
@ -0,0 +1,11 @@
|
||||||
|
--- a/Caddyfile 2022-12-21 11:51:32.834081202 +0100
|
||||||
|
+++ b/Caddyfile 2022-12-21 11:51:26.354030537 +0100
|
||||||
|
@@ -7,6 +7,8 @@
|
||||||
|
{
|
||||||
|
# remove administration tool
|
||||||
|
admin off
|
||||||
|
+ # do not start caddy on port 80
|
||||||
|
+ auto_https disable_redirects
|
||||||
|
}
|
||||||
|
#<GNUNUX
|
||||||
|
|
|
@ -0,0 +1,20 @@
|
||||||
|
# listen to all reverse proxy domains
|
||||||
|
%for %%domain in %%revprox_client_external_domainnames
|
||||||
|
https://%%domain {
|
||||||
|
# import reverse proxy certificate
|
||||||
|
# do not try to check zerossl and let's encrypt file
|
||||||
|
tls %%revprox_client_cert_file %%revprox_client_key_file {
|
||||||
|
ca_root %%revprox_client_ca_file
|
||||||
|
}
|
||||||
|
# log to the console
|
||||||
|
log {
|
||||||
|
output stdout
|
||||||
|
format console
|
||||||
|
level info
|
||||||
|
}
|
||||||
|
# root directory
|
||||||
|
root * %%caddy_root_directory
|
||||||
|
# it's a file server
|
||||||
|
file_server
|
||||||
|
}
|
||||||
|
%end for
|
|
@ -0,0 +1,2 @@
|
||||||
|
g caddy 998 -
|
||||||
|
u caddy 998:998 "Caddy web server" /var/lib/caddy /sbin/nologin
|
|
@ -0,0 +1,2 @@
|
||||||
|
d /srv/caddy 750 root caddy - -
|
||||||
|
d /var/lib/caddy 750 caddy caddy - -
|
|
@ -0,0 +1,5 @@
|
||||||
|
format: '0.1'
|
||||||
|
description: Caddy as standalone HTTPs serveur
|
||||||
|
depends:
|
||||||
|
- base-fedora-36
|
||||||
|
- caddy-common
|
|
@ -0,0 +1,72 @@
|
||||||
|
services:
|
||||||
|
- service:
|
||||||
|
- name: caddy
|
||||||
|
file:
|
||||||
|
- file_type: variable
|
||||||
|
text: caddy_ca_file
|
||||||
|
source: ca_HTTP.crt
|
||||||
|
- file_type: variable
|
||||||
|
text: caddy_crt_file
|
||||||
|
source: caddy.crt
|
||||||
|
- file_type: variable
|
||||||
|
text: caddy_key_file
|
||||||
|
source: caddy.key
|
||||||
|
variables:
|
||||||
|
- family:
|
||||||
|
- name: network
|
||||||
|
variables:
|
||||||
|
- variable:
|
||||||
|
- name: incoming_ports
|
||||||
|
redefine: true
|
||||||
|
value:
|
||||||
|
- text: 80
|
||||||
|
- text: 443
|
||||||
|
- name: caddy
|
||||||
|
variables:
|
||||||
|
- variable:
|
||||||
|
- name: caddy_domain
|
||||||
|
type: domainname
|
||||||
|
description: Domain name
|
||||||
|
- name: caddy_ca_file
|
||||||
|
type: filename
|
||||||
|
description: Caddy CA filename
|
||||||
|
hidden: true
|
||||||
|
- name: caddy_key_file
|
||||||
|
type: filename
|
||||||
|
description: Caddy private key filename
|
||||||
|
hidden: true
|
||||||
|
- name: caddy_crt_file
|
||||||
|
type: filename
|
||||||
|
description: Caddy public key filename
|
||||||
|
hidden: true
|
||||||
|
constraints:
|
||||||
|
- fill:
|
||||||
|
- name: calc_value
|
||||||
|
param:
|
||||||
|
- type: variable
|
||||||
|
text: tls_ca_directory
|
||||||
|
- text: ca_HTTP.crt
|
||||||
|
- name: join
|
||||||
|
text: /
|
||||||
|
target:
|
||||||
|
- text: caddy_ca_file
|
||||||
|
- fill:
|
||||||
|
- name: calc_value
|
||||||
|
param:
|
||||||
|
- type: variable
|
||||||
|
text: tls_cert_directory
|
||||||
|
- text: caddy.crt
|
||||||
|
- name: join
|
||||||
|
text: /
|
||||||
|
target:
|
||||||
|
- text: caddy_crt_file
|
||||||
|
- fill:
|
||||||
|
- name: calc_value
|
||||||
|
param:
|
||||||
|
- type: variable
|
||||||
|
text: tls_key_directory
|
||||||
|
- text: caddy.key
|
||||||
|
- name: join
|
||||||
|
text: /
|
||||||
|
target:
|
||||||
|
- text: caddy_key_file
|
|
@ -0,0 +1 @@
|
||||||
|
PKG="$PKG caddy"
|
57
doc/dataset_example/seed/caddy-https/templates/Caddyfile
Normal file
|
@ -0,0 +1,57 @@
|
||||||
|
# The Caddyfile is an easy way to configure your Caddy web server.
|
||||||
|
#
|
||||||
|
# https://caddyserver.com/docs/caddyfile
|
||||||
|
|
||||||
|
|
||||||
|
# The configuration below serves a welcome page over HTTP on port 80. To use
|
||||||
|
# your own domain name with automatic HTTPS, ensure your A/AAAA DNS record is
|
||||||
|
# pointing to this machine's public IP, then replace `http://` with your domain
|
||||||
|
# name. Refer to the documentation for full instructions on the address
|
||||||
|
# specification.
|
||||||
|
#
|
||||||
|
# https://caddyserver.com/docs/caddyfile/concepts#addresses
|
||||||
|
#>GNUNUX
|
||||||
|
#http:// {
|
||||||
|
#listen only in https
|
||||||
|
{
|
||||||
|
admin off
|
||||||
|
}
|
||||||
|
|
||||||
|
%for %%domain in %%revprox_client_external_domainnames
|
||||||
|
https://%%domain {
|
||||||
|
tls %%revprox_client_cert_file %%revprox_client_key_file {
|
||||||
|
ca_root %%revprox_client_ca_file
|
||||||
|
}
|
||||||
|
log {
|
||||||
|
output stdout
|
||||||
|
format console
|
||||||
|
level info
|
||||||
|
}
|
||||||
|
#<GNUNUX
|
||||||
|
|
||||||
|
# Set this path to your site's directory.
|
||||||
|
#>GNUNUX
|
||||||
|
# root * /usr/share/caddy
|
||||||
|
root * /srv/caddy
|
||||||
|
#<GNUNUX
|
||||||
|
|
||||||
|
# Enable the static file server.
|
||||||
|
file_server
|
||||||
|
|
||||||
|
# Another common task is to set up a reverse proxy:
|
||||||
|
# reverse_proxy localhost:8080
|
||||||
|
|
||||||
|
# Or serve a PHP site through php-fpm:
|
||||||
|
# php_fastcgi localhost:9000
|
||||||
|
|
||||||
|
# Refer to the directive documentation for more options.
|
||||||
|
# https://caddyserver.com/docs/caddyfile/directives
|
||||||
|
|
||||||
|
}
|
||||||
|
%end for
|
||||||
|
|
||||||
|
|
||||||
|
# As an alternative to editing the above site block, you can add your own site
|
||||||
|
# block files in the Caddyfile.d directory, and they will be included as long
|
||||||
|
# as they use the .caddyfile extension.
|
||||||
|
#GNUNUX import Caddyfile.d/*.caddyfile
|
|
@ -0,0 +1 @@
|
||||||
|
%%get_chain(cn=%%caddy_domain, authority_cn=%%caddy_domain, authority_name="HTTP", hide=%%hide_secret)
|
1
doc/dataset_example/seed/caddy-https/templates/caddy.crt
Normal file
|
@ -0,0 +1 @@
|
||||||
|
%%get_certificate(%%caddy_domain, 'HTTP', type="server", hide=%%hide_secret)
|
1
doc/dataset_example/seed/caddy-https/templates/caddy.key
Normal file
|
@ -0,0 +1 @@
|
||||||
|
%%get_private_key(cn=%%caddy_domain, authority_name='HTTP', type="server", hide=%%hide_secret)
|
|
@ -0,0 +1,18 @@
|
||||||
|
# listen to all reverse proxy domains
|
||||||
|
https://%%caddy_domain {
|
||||||
|
# use certificate
|
||||||
|
# do not try to check zerossl and let's encrypt file
|
||||||
|
tls %%caddy_crt_file %%caddy_key_file {
|
||||||
|
ca_root %%caddy_ca_file
|
||||||
|
}
|
||||||
|
# log to the console
|
||||||
|
log {
|
||||||
|
output stdout
|
||||||
|
format console
|
||||||
|
level info
|
||||||
|
}
|
||||||
|
# root directory
|
||||||
|
root * %%caddy_root_directory
|
||||||
|
# it's a file server
|
||||||
|
file_server
|
||||||
|
}
|
|
@ -0,0 +1,2 @@
|
||||||
|
g caddy 998 -
|
||||||
|
u caddy 998:998 "Caddy web server" /var/lib/caddy /sbin/nologin
|
|
@ -0,0 +1,2 @@
|
||||||
|
d /srv/caddy 750 root caddy - -
|
||||||
|
d /var/lib/caddy 750 caddy caddy - -
|
48
doc/dataset_example/servers.caddy-https-rp.yml
Normal file
|
@ -0,0 +1,48 @@
|
||||||
|
zones:
|
||||||
|
external:
|
||||||
|
network: 192.168.45.0/24
|
||||||
|
host_ip: 192.168.45.1
|
||||||
|
start_ip: 192.168.45.10
|
||||||
|
domain_name: in.example.net
|
||||||
|
revprox:
|
||||||
|
network: 192.168.46.0/24
|
||||||
|
host_ip: 192.168.46.1
|
||||||
|
start_ip: 192.168.46.10
|
||||||
|
domain_name: revprox.in.example.net
|
||||||
|
modules:
|
||||||
|
revprox:
|
||||||
|
- nginx-reverse-proxy
|
||||||
|
- letsencrypt
|
||||||
|
nsd:
|
||||||
|
- nsd
|
||||||
|
caddy:
|
||||||
|
- caddy-https-rp
|
||||||
|
hosts:
|
||||||
|
host.example.net:
|
||||||
|
applicationservices:
|
||||||
|
- host-systemd-machined
|
||||||
|
applicationservice_provider: provider-systemd-machined
|
||||||
|
values:
|
||||||
|
general.network.interfaces.interface_names:
|
||||||
|
- ens3
|
||||||
|
general.network.output_interface: ens3
|
||||||
|
servers:
|
||||||
|
nsd:
|
||||||
|
module: nsd
|
||||||
|
informations:
|
||||||
|
zones_name:
|
||||||
|
- revprox
|
||||||
|
revprox:
|
||||||
|
module: revprox
|
||||||
|
informations:
|
||||||
|
zones_name:
|
||||||
|
- external
|
||||||
|
- revprox
|
||||||
|
caddy:
|
||||||
|
module: caddy
|
||||||
|
informations:
|
||||||
|
zones_name:
|
||||||
|
- revprox
|
||||||
|
values:
|
||||||
|
general.revprox.revprox_client.revprox_client_external_domainnames:
|
||||||
|
- caddy.example.net
|
26
doc/dataset_example/servers.caddy-https.yml
Normal file
|
@ -0,0 +1,26 @@
|
||||||
|
zones:
|
||||||
|
external:
|
||||||
|
network: 192.168.45.0/24
|
||||||
|
host_ip: 192.168.45.1
|
||||||
|
start_ip: 192.168.45.10
|
||||||
|
domain_name: in.example.net
|
||||||
|
modules:
|
||||||
|
caddy:
|
||||||
|
- caddy-https
|
||||||
|
hosts:
|
||||||
|
host.example.net:
|
||||||
|
applicationservices:
|
||||||
|
- host-systemd-machined
|
||||||
|
applicationservice_provider: provider-systemd-machined
|
||||||
|
values:
|
||||||
|
general.network.interfaces.interface_names:
|
||||||
|
- ens3
|
||||||
|
general.network.output_interface: ens3
|
||||||
|
servers:
|
||||||
|
caddy:
|
||||||
|
module: caddy
|
||||||
|
informations:
|
||||||
|
zones_name:
|
||||||
|
- external
|
||||||
|
values:
|
||||||
|
general.caddy.caddy_domain: caddy.example.net
|
BIN
doc/example_smtp.png
Normal file
After Width: | Height: | Size: 100 KiB |
466
doc/example_smtp.svg
Normal file
|
@ -0,0 +1,466 @@
|
||||||
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||||
|
<!-- Created with Inkscape (http://www.inkscape.org/) -->
|
||||||
|
|
||||||
|
<svg
|
||||||
|
width="178.66008mm"
|
||||||
|
height="172.7524mm"
|
||||||
|
viewBox="0 0 178.66008 172.7524"
|
||||||
|
version="1.1"
|
||||||
|
id="svg5"
|
||||||
|
inkscape:version="1.2.1 (9c6d41e410, 2022-07-14)"
|
||||||
|
sodipodi:docname="example_smtp.svg"
|
||||||
|
inkscape:export-filename="example_smtp.png"
|
||||||
|
inkscape:export-xdpi="149.26"
|
||||||
|
inkscape:export-ydpi="149.26"
|
||||||
|
xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
|
||||||
|
xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
|
||||||
|
xmlns="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:svg="http://www.w3.org/2000/svg">
|
||||||
|
<sodipodi:namedview
|
||||||
|
id="namedview7"
|
||||||
|
pagecolor="#ffffff"
|
||||||
|
bordercolor="#666666"
|
||||||
|
borderopacity="1.0"
|
||||||
|
inkscape:pageshadow="2"
|
||||||
|
inkscape:pageopacity="0.0"
|
||||||
|
inkscape:pagecheckerboard="0"
|
||||||
|
inkscape:document-units="mm"
|
||||||
|
showgrid="false"
|
||||||
|
inkscape:zoom="0.55055723"
|
||||||
|
inkscape:cx="173.46062"
|
||||||
|
inkscape:cy="599.39273"
|
||||||
|
inkscape:window-width="1920"
|
||||||
|
inkscape:window-height="1011"
|
||||||
|
inkscape:window-x="0"
|
||||||
|
inkscape:window-y="0"
|
||||||
|
inkscape:window-maximized="1"
|
||||||
|
inkscape:current-layer="layer1"
|
||||||
|
inkscape:showpageshadow="2"
|
||||||
|
inkscape:deskcolor="#d1d1d1" />
|
||||||
|
<defs
|
||||||
|
id="defs2">
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker47493"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Sstart"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(0.2,0,0,0.2,1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path47491" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow2Send"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow2Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 3.4652294 2.5981128"
|
||||||
|
markerWidth="3.465229"
|
||||||
|
markerHeight="2.5981126"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.3,0,0,-0.3,0.69,0)"
|
||||||
|
d="M 8.7185878,4.0337352 -2.2072895,0.01601326 8.7185884,-4.0017078 c -1.7454984,2.3720609 -1.7354408,5.6174519 -6e-7,8.035443 z"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:0.625;stroke-linejoin:round"
|
||||||
|
id="path45715" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow2Sstart"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow2Sstart"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 3.4652294 2.5981128"
|
||||||
|
markerWidth="3.465229"
|
||||||
|
markerHeight="2.5981126"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(0.3,0,0,0.3,-0.69,0)"
|
||||||
|
d="M 8.7185878,4.0337352 -2.2072895,0.01601326 8.7185884,-4.0017078 c -1.7454984,2.3720609 -1.7354408,5.6174519 -6e-7,8.035443 z"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:0.625;stroke-linejoin:round"
|
||||||
|
id="path45712" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow1Send"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path45697" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow1Sstart"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Sstart"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(0.2,0,0,0.2,1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path45694" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow1Lend"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Lend"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.8,0,0,-0.8,-10,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path45685" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 4.4434635 2.539122"
|
||||||
|
markerWidth="4.4434633"
|
||||||
|
markerHeight="2.5391221"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-1"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 4.4434635 2.539122"
|
||||||
|
markerWidth="4.4434633"
|
||||||
|
markerHeight="2.5391221"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-8" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-9"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-2" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-9-3"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-2-7" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6-9"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2-2" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="Arrow2Sstart-8"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow2Sstart"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(0.3,0,0,0.3,-0.69,0)"
|
||||||
|
d="M 8.7185878,4.0337352 -2.2072895,0.01601326 8.7185884,-4.0017078 c -1.7454984,2.3720609 -1.7354408,5.6174519 -6e-7,8.035443 z"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:0.625;stroke-linejoin:round"
|
||||||
|
id="path45712-9" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6-9-7"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2-2-3" />
|
||||||
|
</marker>
|
||||||
|
<marker
|
||||||
|
style="overflow:visible"
|
||||||
|
id="marker46179-3-6-4"
|
||||||
|
refX="0"
|
||||||
|
refY="0"
|
||||||
|
orient="auto"
|
||||||
|
inkscape:stockid="Arrow1Send"
|
||||||
|
inkscape:isstock="true"
|
||||||
|
viewBox="0 0 4.4434635 2.539122"
|
||||||
|
markerWidth="4.4434633"
|
||||||
|
markerHeight="2.5391221"
|
||||||
|
preserveAspectRatio="xMidYMid">
|
||||||
|
<path
|
||||||
|
transform="matrix(-0.2,0,0,-0.2,-1.2,0)"
|
||||||
|
style="fill:context-stroke;fill-rule:evenodd;stroke:context-stroke;stroke-width:1pt"
|
||||||
|
d="M 0,0 5,-5 -12.5,0 5,5 Z"
|
||||||
|
id="path46177-6-2-7" />
|
||||||
|
</marker>
|
||||||
|
</defs>
|
||||||
|
<g
|
||||||
|
inkscape:label="Calque 1"
|
||||||
|
inkscape:groupmode="layer"
|
||||||
|
id="layer1"
|
||||||
|
transform="translate(-15.292364,-14.109702)">
|
||||||
|
<rect
|
||||||
|
style="fill:#ffffff;fill-opacity:1;fill-rule:evenodd;stroke:#f6f7d7;stroke-width:0.600001;stroke-linecap:round;stroke-linejoin:round;paint-order:fill markers stroke"
|
||||||
|
id="rect443"
|
||||||
|
width="178.06007"
|
||||||
|
height="172.15239"
|
||||||
|
x="15.592364"
|
||||||
|
y="14.409702" />
|
||||||
|
<circle
|
||||||
|
style="fill:none;fill-rule:evenodd;stroke:#000000;stroke-width:4;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path846"
|
||||||
|
cx="96.73632"
|
||||||
|
cy="103.80212"
|
||||||
|
r="52.962326" />
|
||||||
|
<ellipse
|
||||||
|
style="fill:#000000;fill-opacity:1;fill-rule:evenodd;stroke:#000000;stroke-width:5.86795;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path986"
|
||||||
|
cx="62.406502"
|
||||||
|
cy="64.119804"
|
||||||
|
rx="4.5660253"
|
||||||
|
ry="4.5660257" />
|
||||||
|
<ellipse
|
||||||
|
style="fill:#000000;fill-opacity:1;fill-rule:evenodd;stroke:#000000;stroke-width:5.86795;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path986-6"
|
||||||
|
cx="62.407001"
|
||||||
|
cy="144.45392"
|
||||||
|
rx="4.5660253"
|
||||||
|
ry="4.5660257" />
|
||||||
|
<ellipse
|
||||||
|
style="fill:#000000;fill-opacity:1;fill-rule:evenodd;stroke:#000000;stroke-width:5.86795;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path986-6-5"
|
||||||
|
cx="98.457001"
|
||||||
|
cy="79.992493"
|
||||||
|
rx="4.5660253"
|
||||||
|
ry="4.5660257" />
|
||||||
|
<ellipse
|
||||||
|
style="fill:#000000;fill-opacity:1;fill-rule:evenodd;stroke:#000000;stroke-width:5.86795;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path986-6-5-7"
|
||||||
|
cx="98.45739"
|
||||||
|
cy="122.40948"
|
||||||
|
rx="4.5660253"
|
||||||
|
ry="4.5660257" />
|
||||||
|
<ellipse
|
||||||
|
style="fill:#000000;fill-opacity:1;fill-rule:evenodd;stroke:#000000;stroke-width:5.86795;stroke-linecap:square;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path986-6-5-9"
|
||||||
|
cx="149.40425"
|
||||||
|
cy="102.7455"
|
||||||
|
rx="4.5660253"
|
||||||
|
ry="4.5660257" />
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:4.93889px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;stroke-width:0.264583"
|
||||||
|
x="37.39616"
|
||||||
|
y="61.122501"
|
||||||
|
id="text3135"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
id="tspan3133"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="37.39616"
|
||||||
|
y="61.122501">IMAP (993)</tspan><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="37.39616"
|
||||||
|
y="67.296112"
|
||||||
|
id="tspan10911">SMTP (587)</tspan></text>
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:4.93889px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;stroke-width:0.264583"
|
||||||
|
x="62.512436"
|
||||||
|
y="157.8011"
|
||||||
|
id="text3135-7"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="62.512436"
|
||||||
|
y="157.8011"
|
||||||
|
id="tspan10911-3">SMTP</tspan><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="62.512436"
|
||||||
|
y="163.97472"
|
||||||
|
id="tspan51838">relay</tspan><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="62.512436"
|
||||||
|
y="170.14833"
|
||||||
|
id="tspan51842">(25)</tspan></text>
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:4.93889px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;stroke-width:0.264583"
|
||||||
|
x="91.762589"
|
||||||
|
y="70.189774"
|
||||||
|
id="text3135-7-6"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;stroke-width:0.264583"
|
||||||
|
x="91.762589"
|
||||||
|
y="70.189774"
|
||||||
|
id="tspan10911-3-2">LDAP</tspan></text>
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:4.93889px;line-height:1.25;font-family:sans-serif;text-align:center;letter-spacing:0px;word-spacing:0px;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="162.68114"
|
||||||
|
y="114.31043"
|
||||||
|
id="text3135-7-6-1"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="162.68114"
|
||||||
|
y="114.31043"
|
||||||
|
id="tspan10911-3-2-2">DNS</tspan><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="162.68114"
|
||||||
|
y="120.48405"
|
||||||
|
id="tspan21295">Résolver</tspan></text>
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:4.93889px;line-height:1.25;font-family:sans-serif;text-align:center;letter-spacing:0px;word-spacing:0px;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="98.739502"
|
||||||
|
y="135.19682"
|
||||||
|
id="text3135-7-6-1-0"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="98.739502"
|
||||||
|
y="135.19682"
|
||||||
|
id="tspan10911-3-2-2-9">DNS</tspan><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;font-size:4.93889px;text-align:center;text-anchor:middle;stroke-width:0.264583"
|
||||||
|
x="98.739502"
|
||||||
|
y="141.37044"
|
||||||
|
id="tspan22983">autoritaire</tspan></text>
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="M 98.411,89.863152 V 107.09425"
|
||||||
|
id="path46175"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="m 69.168475,71.05256 15.523919,4.588191"
|
||||||
|
id="path46175-02"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-start:url(#Arrow2Sstart);marker-end:url(#Arrow2Send)"
|
||||||
|
d="m 58.918881,78.428313 c 0,0 -7.642846,11.083665 -8.1137,23.703427 -0.554549,14.86295 7.598141,26.95783 7.598141,26.95783"
|
||||||
|
id="path46175-02-5"
|
||||||
|
sodipodi:nodetypes="csc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="m 73.470622,64.314043 c 0,0 23.019562,-13.687982 46.104948,-0.359501 9.42693,5.44269 13.02345,12.067909 16.41683,17.107652 3.97188,5.898933 4.72416,9.274399 4.72416,9.274399"
|
||||||
|
id="path46175-7"
|
||||||
|
sodipodi:nodetypes="cssc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="m 140.00042,103.64348 -27.82831,13.38506"
|
||||||
|
id="path46175-0"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="m 158.873,103.231 19.41573,1.3e-4"
|
||||||
|
id="path46175-0-6"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="M 38.595,36.174542 51.766,51.796621"
|
||||||
|
id="path46175-0-6-8"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-start:url(#Arrow2Sstart);marker-end:url(#Arrow2Send)"
|
||||||
|
d="M 51.766,154.77542 38.595,168.15219"
|
||||||
|
id="path46175-0-6-2-6"
|
||||||
|
sodipodi:nodetypes="cc" />
|
||||||
|
<path
|
||||||
|
style="fill:none;stroke:#000000;stroke-width:4;stroke-linecap:butt;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;marker-end:url(#Arrow2Send)"
|
||||||
|
d="m 73.436962,143.41602 c 0,0 17.397816,13.36128 44.232048,1.48383 16.06906,-7.11254 23.91983,-29.57648 23.91983,-29.57648"
|
||||||
|
id="path57407"
|
||||||
|
sodipodi:nodetypes="csc" />
|
||||||
|
</g>
|
||||||
|
</svg>
|
After Width: | Height: | Size: 18 KiB |
41
doc/infrastructure.md
Normal file
|
@ -0,0 +1,41 @@
|
||||||
|
# Infrastructure
|
||||||
|
|
||||||
|
The infrastructure is define in a uniq YAML file: servers.yml:
|
||||||
|
|
||||||
|
## Zones
|
||||||
|
|
||||||
|
The idea:
|
||||||
|
|
||||||
|
- separate the networks according to the uses
|
||||||
|
- there is no route to each other
|
||||||
|
|
||||||
|
Ideally only one area has an Internet access.
|
||||||
|
Internet access is, in fact, firewall rules.
|
||||||
|
This network is usually called "external".
|
||||||
|
|
||||||
|
The other networks are only there for the communication between server and client.
|
||||||
|
|
||||||
|
The host must have an IP in this network.
|
||||||
|
IP inside this network are deliver automaticly.
|
||||||
|
|
||||||
|
A network is call a "zone".
|
||||||
|
|
||||||
|
## Modules
|
||||||
|
|
||||||
|
A module is simply a list of application services. An system image is build with informations define in application service.
|
||||||
|
|
||||||
|
## Hosts
|
||||||
|
|
||||||
|
A host is a server on which container or VM are running.
|
||||||
|
Define the host means define:
|
||||||
|
|
||||||
|
- application services to configure the host and VM
|
||||||
|
- application service provider to define the provider to apply on each VM
|
||||||
|
- values to adapt the configuration
|
||||||
|
- servers, the list of VM with :
|
||||||
|
|
||||||
|
- the corresponding module
|
||||||
|
- informations (like zone)
|
||||||
|
- values
|
||||||
|
|
||||||
|
Host must only be a Debian 11 (Bullseye) from now.
|
BIN
doc/schema.png
Normal file
After Width: | Height: | Size: 323 KiB |
3750
doc/schema.svg
Normal file
After Width: | Height: | Size: 167 KiB |
233
funcs.py
|
@ -1,233 +0,0 @@
|
||||||
from tiramisu import valid_network_netmask, valid_ip_netmask, valid_broadcast, valid_in_network, valid_not_equal as valid_differ, valid_not_equal, calc_value
|
|
||||||
from ipaddress import ip_address
|
|
||||||
from os.path import dirname, abspath, join as _join, isdir as _isdir, isfile as _isfile
|
|
||||||
from typing import List
|
|
||||||
from json import load
|
|
||||||
from secrets import token_urlsafe as _token_urlsafe
|
|
||||||
|
|
||||||
from rougail.utils import normalize_family
|
|
||||||
|
|
||||||
from risotto.utils import multi_function, CONFIGS
|
|
||||||
from risotto.x509 import gen_cert as _x509_gen_cert, gen_ca as _x509_gen_ca, gen_pub as _x509_gen_pub, has_pub as _x509_has_pub
|
|
||||||
# =============================================================
|
|
||||||
# fork of risotto-setting/src/risotto_setting/config/config.py
|
|
||||||
|
|
||||||
with open('servers.json', 'r') as server_fh:
|
|
||||||
ZONES_SERVER = load(server_fh)
|
|
||||||
|
|
||||||
|
|
||||||
ZONES = None
|
|
||||||
DOMAINS = None
|
|
||||||
HERE = dirname(abspath(__file__))
|
|
||||||
|
|
||||||
|
|
||||||
def load_zones():
|
|
||||||
global ZONES
|
|
||||||
if ZONES is not None:
|
|
||||||
return
|
|
||||||
ZONES = ZONES_SERVER['zones']
|
|
||||||
for server_name, server in ZONES_SERVER['servers'].items():
|
|
||||||
if 'informations' not in server:
|
|
||||||
continue
|
|
||||||
server_zones = server['informations']['zones_name']
|
|
||||||
server_extra_domainnames = server['informations'].get('extra_domainnames', [])
|
|
||||||
if len(server_zones) > 1 and len(server_zones) != len(server_extra_domainnames) + 1:
|
|
||||||
raise Exception(f'the server "{server_name}" has more that one zone, please set correct number of extra_domainnames ({len(server_zones) - 1} instead of {len(server_extra_domainnames)})')
|
|
||||||
|
|
||||||
for idx, zone_name in enumerate(server_zones):
|
|
||||||
zone_domain_name = ZONES[zone_name]['domain_name']
|
|
||||||
if idx == 0:
|
|
||||||
zone_server_name = server_name
|
|
||||||
else:
|
|
||||||
zone_server_name = server_extra_domainnames[idx - 1]
|
|
||||||
server_domain_name = zone_server_name.split('.', 1)[1]
|
|
||||||
if zone_domain_name and zone_domain_name != server_domain_name:
|
|
||||||
raise Exception(f'wrong server_name "{zone_server_name}" in zone "{zone_name}" should ends with "{zone_domain_name}"')
|
|
||||||
ZONES[zone_name].setdefault('hosts', []).append(server_name)
|
|
||||||
|
|
||||||
|
|
||||||
def load_domains():
|
|
||||||
load_zones()
|
|
||||||
global DOMAINS
|
|
||||||
if DOMAINS is not None:
|
|
||||||
return
|
|
||||||
DOMAINS = {}
|
|
||||||
for zone_name, zone in ZONES_SERVER['zones'].items():
|
|
||||||
if 'domain_name' in zone:
|
|
||||||
hosts = []
|
|
||||||
ips = []
|
|
||||||
for host in ZONES[zone_name].get('hosts', []):
|
|
||||||
hosts.append(host.split('.', 1)[0])
|
|
||||||
ips.append(get_ip(host, [zone_name], 0))
|
|
||||||
DOMAINS[zone['domain_name']] = (tuple(hosts), tuple(ips))
|
|
||||||
|
|
||||||
|
|
||||||
def get_ip(server_name: str,
|
|
||||||
zones_name: List[str],
|
|
||||||
index: str,
|
|
||||||
) -> str:
|
|
||||||
if server_name is None:
|
|
||||||
return
|
|
||||||
load_zones()
|
|
||||||
index = int(index)
|
|
||||||
zone_name = zones_name[index]
|
|
||||||
if zone_name not in ZONES:
|
|
||||||
raise ValueError(f"cannot set IP in unknown zone '{zone_name}'")
|
|
||||||
zone = ZONES[zone_name]
|
|
||||||
if server_name not in zone['hosts']:
|
|
||||||
raise ValueError(f"cannot set IP in unknown server '{server_name}'")
|
|
||||||
server_index = zone['hosts'].index(server_name)
|
|
||||||
# print(server_name, zones_name, index, str(ip_address(zone['start_ip']) + server_index))
|
|
||||||
return str(ip_address(zone['start_ip']) + server_index)
|
|
||||||
|
|
||||||
|
|
||||||
@multi_function
|
|
||||||
def get_chain(authority_cn,
|
|
||||||
authority_name,
|
|
||||||
):
|
|
||||||
if not authority_name or authority_name is None:
|
|
||||||
if isinstance(authority_name, list):
|
|
||||||
return []
|
|
||||||
return
|
|
||||||
if not isinstance(authority_cn, list):
|
|
||||||
is_list = False
|
|
||||||
authority_cn = [authority_cn]
|
|
||||||
else:
|
|
||||||
is_list = True
|
|
||||||
authorities = []
|
|
||||||
|
|
||||||
for auth_cn in authority_cn:
|
|
||||||
ret = _x509_gen_ca(auth_cn,
|
|
||||||
authority_name,
|
|
||||||
HERE,
|
|
||||||
)
|
|
||||||
if not is_list:
|
|
||||||
return ret
|
|
||||||
authorities.append(ret)
|
|
||||||
return authorities
|
|
||||||
|
|
||||||
|
|
||||||
@multi_function
|
|
||||||
def get_certificate(cn,
|
|
||||||
authority_name,
|
|
||||||
authority_cn=None,
|
|
||||||
extra_domainnames=[],
|
|
||||||
type='server',
|
|
||||||
):
|
|
||||||
if isinstance(cn, list) and extra_domainnames:
|
|
||||||
raise Exception('cn cannot be a list with extra_domainnames set')
|
|
||||||
if not cn or authority_name is None:
|
|
||||||
if isinstance(cn, list):
|
|
||||||
return []
|
|
||||||
return
|
|
||||||
return _x509_gen_cert(cn,
|
|
||||||
extra_domainnames,
|
|
||||||
authority_cn,
|
|
||||||
authority_name,
|
|
||||||
type,
|
|
||||||
'crt',
|
|
||||||
HERE,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@multi_function
|
|
||||||
def get_private_key(cn,
|
|
||||||
authority_name=None,
|
|
||||||
authority_cn=None,
|
|
||||||
type='server',
|
|
||||||
):
|
|
||||||
if not cn:
|
|
||||||
if isinstance(cn, list):
|
|
||||||
return []
|
|
||||||
return
|
|
||||||
if authority_name is None:
|
|
||||||
if _x509_has_pub(cn, HERE):
|
|
||||||
return _x509_gen_pub(cn,
|
|
||||||
'key',
|
|
||||||
HERE,
|
|
||||||
)
|
|
||||||
if isinstance(cn, list):
|
|
||||||
return []
|
|
||||||
return
|
|
||||||
return _x509_gen_cert(cn,
|
|
||||||
[],
|
|
||||||
authority_cn,
|
|
||||||
authority_name,
|
|
||||||
type,
|
|
||||||
'key',
|
|
||||||
HERE,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def get_public_key(cn):
|
|
||||||
if not cn:
|
|
||||||
return
|
|
||||||
return _x509_gen_pub(cn,
|
|
||||||
'pub',
|
|
||||||
HERE,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def zone_information(zone_name: str,
|
|
||||||
type: str,
|
|
||||||
multi: bool=False,
|
|
||||||
index: int=None,
|
|
||||||
) -> str:
|
|
||||||
if not zone_name:
|
|
||||||
return
|
|
||||||
if type == 'gateway' and index != 0:
|
|
||||||
return
|
|
||||||
load_zones()
|
|
||||||
if zone_name not in ZONES:
|
|
||||||
raise ValueError(f"cannot get zone informations in unknown zone '{zone_name}'")
|
|
||||||
zone = ZONES[zone_name]
|
|
||||||
if type not in zone:
|
|
||||||
raise ValueError(f"unknown type '{type}' in zone '{zone_name}'")
|
|
||||||
value = zone[type]
|
|
||||||
if multi:
|
|
||||||
value = [value]
|
|
||||||
return value
|
|
||||||
|
|
||||||
|
|
||||||
def get_internal_zones() -> List[str]:
|
|
||||||
load_domains()
|
|
||||||
return list(DOMAINS.keys())
|
|
||||||
|
|
||||||
|
|
||||||
@multi_function
|
|
||||||
def get_zones_info(type: str) -> str:
|
|
||||||
ret = []
|
|
||||||
for data in ZONES_SERVER['zones'].values():
|
|
||||||
ret.append(data[type])
|
|
||||||
return ret
|
|
||||||
|
|
||||||
@multi_function
|
|
||||||
def get_internal_zone_names() -> List[str]:
|
|
||||||
load_zones()
|
|
||||||
return list(ZONES.keys())
|
|
||||||
|
|
||||||
|
|
||||||
def get_internal_zone_information(zone: str,
|
|
||||||
info: str,
|
|
||||||
) -> str:
|
|
||||||
load_domains()
|
|
||||||
if info == 'cidr':
|
|
||||||
return ZONES[zone]['gateway'] + '/' + ZONES[zone]['network'].split('/')[-1]
|
|
||||||
return ZONES[zone][info]
|
|
||||||
|
|
||||||
|
|
||||||
def get_internal_info_in_zone(zone: str,
|
|
||||||
auto: bool,
|
|
||||||
type: str,
|
|
||||||
index: int=None,
|
|
||||||
) -> List[str]:
|
|
||||||
if not auto:
|
|
||||||
return
|
|
||||||
for domain_name, domain in DOMAINS.items():
|
|
||||||
if zone == domain_name:
|
|
||||||
if type == 'host':
|
|
||||||
return list(domain[0])
|
|
||||||
else:
|
|
||||||
return domain[1][index]
|
|
||||||
|
|
||||||
# =============================================================
|
|
BIN
logo.png
Normal file
After Width: | Height: | Size: 8.4 KiB |
115
logo.svg
Normal file
|
@ -0,0 +1,115 @@
|
||||||
|
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||||
|
<!-- Created with Inkscape (http://www.inkscape.org/) -->
|
||||||
|
|
||||||
|
<svg
|
||||||
|
width="37.297001mm"
|
||||||
|
height="38.629002mm"
|
||||||
|
viewBox="0 0 37.297 38.629002"
|
||||||
|
version="1.1"
|
||||||
|
id="svg5"
|
||||||
|
inkscape:version="1.2.1 (9c6d41e410, 2022-07-14)"
|
||||||
|
sodipodi:docname="logo.svg"
|
||||||
|
inkscape:export-filename="logo.png"
|
||||||
|
inkscape:export-xdpi="149.26"
|
||||||
|
inkscape:export-ydpi="149.26"
|
||||||
|
xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
|
||||||
|
xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
|
||||||
|
xmlns="http://www.w3.org/2000/svg"
|
||||||
|
xmlns:svg="http://www.w3.org/2000/svg">
|
||||||
|
<sodipodi:namedview
|
||||||
|
id="namedview7"
|
||||||
|
pagecolor="#ffffff"
|
||||||
|
bordercolor="#666666"
|
||||||
|
borderopacity="1.0"
|
||||||
|
inkscape:pageshadow="2"
|
||||||
|
inkscape:pageopacity="0.0"
|
||||||
|
inkscape:pagecheckerboard="0"
|
||||||
|
inkscape:document-units="mm"
|
||||||
|
showgrid="false"
|
||||||
|
inkscape:zoom="4.404458"
|
||||||
|
inkscape:cx="63.685475"
|
||||||
|
inkscape:cy="75.378174"
|
||||||
|
inkscape:window-width="1920"
|
||||||
|
inkscape:window-height="1011"
|
||||||
|
inkscape:window-x="0"
|
||||||
|
inkscape:window-y="0"
|
||||||
|
inkscape:window-maximized="1"
|
||||||
|
inkscape:current-layer="layer1"
|
||||||
|
inkscape:showpageshadow="2"
|
||||||
|
inkscape:deskcolor="#d1d1d1" />
|
||||||
|
<defs
|
||||||
|
id="defs2" />
|
||||||
|
<g
|
||||||
|
inkscape:label="Calque 1"
|
||||||
|
inkscape:groupmode="layer"
|
||||||
|
id="layer1"
|
||||||
|
transform="translate(-75.0784,-36.897831)">
|
||||||
|
<rect
|
||||||
|
style="fill:#f6f7d7;fill-opacity:1;fill-rule:evenodd;stroke:none;stroke-width:2.04884;stroke-linecap:square;paint-order:fill markers stroke"
|
||||||
|
id="rect8118"
|
||||||
|
width="37.297001"
|
||||||
|
height="38.629002"
|
||||||
|
x="75.0784"
|
||||||
|
y="36.897831" />
|
||||||
|
<rect
|
||||||
|
style="fill:none;fill-opacity:1;fill-rule:evenodd;stroke:none;stroke-width:1.5;stroke-linecap:square;stroke-linejoin:miter;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="rect19192"
|
||||||
|
width="37.29702"
|
||||||
|
height="38.628922"
|
||||||
|
x="75.0784"
|
||||||
|
y="36.897831" />
|
||||||
|
<rect
|
||||||
|
style="fill:#008700;fill-opacity:1;fill-rule:evenodd;stroke:#008700;stroke-width:1.84143;stroke-linecap:square;stroke-linejoin:round;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="rect18918"
|
||||||
|
width="29.788723"
|
||||||
|
height="6.963315"
|
||||||
|
x="78.625404"
|
||||||
|
y="40.178349" />
|
||||||
|
<rect
|
||||||
|
style="fill:#008700;fill-opacity:1;fill-rule:evenodd;stroke:#008700;stroke-width:1.84143;stroke-linecap:square;stroke-linejoin:round;stroke-miterlimit:4;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="rect18918-5"
|
||||||
|
width="29.788723"
|
||||||
|
height="6.963315"
|
||||||
|
x="78.625114"
|
||||||
|
y="65.0494" />
|
||||||
|
<text
|
||||||
|
xml:space="preserve"
|
||||||
|
style="font-size:10.5833px;line-height:1.15;font-family:sans-serif;text-align:center;letter-spacing:0px;word-spacing:0px;text-anchor:middle;fill:#4d4d4d;stroke-width:0.265;stroke-miterlimit:4;stroke-dasharray:none"
|
||||||
|
x="93.5"
|
||||||
|
y="47.586319"
|
||||||
|
id="text2080"><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
id="tspan2078"
|
||||||
|
style="font-weight:bold;text-align:center;text-anchor:middle;fill:#f6f7d7;fill-opacity:1;stroke-width:0.265;stroke-miterlimit:4;stroke-dasharray:none"
|
||||||
|
x="93.5"
|
||||||
|
y="47.586319">RIS</tspan><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;text-align:center;text-anchor:middle;fill:#4d4d4d;stroke-width:0.265;stroke-miterlimit:4;stroke-dasharray:none"
|
||||||
|
x="93.5"
|
||||||
|
y="59.757114"
|
||||||
|
id="tspan2082">OTTO</tspan><tspan
|
||||||
|
sodipodi:role="line"
|
||||||
|
style="font-weight:bold;text-align:center;text-anchor:middle;fill:#f6f7d7;fill-opacity:1;stroke-width:0.265;stroke-miterlimit:4;stroke-dasharray:none"
|
||||||
|
x="93.5"
|
||||||
|
y="71.92791"
|
||||||
|
id="tspan7995" /></text>
|
||||||
|
<circle
|
||||||
|
style="fill:#008700;fill-opacity:1;fill-rule:evenodd;stroke:#f6f7d7;stroke-width:0.56;stroke-linecap:square;stroke-dasharray:none;stroke-opacity:1;paint-order:fill markers stroke"
|
||||||
|
id="path19218"
|
||||||
|
cx="103.00674"
|
||||||
|
cy="68.43734"
|
||||||
|
r="1.7277808" />
|
||||||
|
<path
|
||||||
|
style="fill:#f6f7d7;fill-opacity:1;stroke:#f6f7d7;stroke-width:0.56;stroke-linecap:round;stroke-linejoin:round;stroke-dasharray:none;stroke-opacity:1"
|
||||||
|
d="M 82.1984,66.707831 H 95.287674"
|
||||||
|
id="path19357" />
|
||||||
|
<path
|
||||||
|
style="fill:#f6f7d7;fill-opacity:1;stroke:#f6f7d7;stroke-width:0.6;stroke-linecap:round;stroke-linejoin:round;stroke-dasharray:none;stroke-opacity:1"
|
||||||
|
d="M 82.1984,70.167831 H 95.287664"
|
||||||
|
id="path19357-6" />
|
||||||
|
<path
|
||||||
|
style="fill:#f6f7d7;fill-opacity:1;stroke:#f6f7d7;stroke-width:0.56;stroke-linecap:round;stroke-linejoin:round;stroke-dasharray:none;stroke-opacity:1"
|
||||||
|
d="M 82.1984,68.45114 H 95.287664"
|
||||||
|
id="path19357-5" />
|
||||||
|
</g>
|
||||||
|
</svg>
|
After Width: | Height: | Size: 4.8 KiB |
|
@ -1,3 +1,4 @@
|
||||||
[directories]
|
[directories]
|
||||||
dataset = '/home/gnunux/git/risotto/dataset/seed'
|
dataset = '/home/gnunux/git/risotto/dataset/seed'
|
||||||
dest = 'installations'
|
dest = 'installations'
|
||||||
|
dest_templates = 'templates'
|
||||||
|
|
368
sbin/risotto_auto_doc
Executable file
|
@ -0,0 +1,368 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
from os import listdir
|
||||||
|
from os.path import isdir, join
|
||||||
|
from tabulate import tabulate
|
||||||
|
from sys import argv
|
||||||
|
|
||||||
|
from rougail import RougailConfig
|
||||||
|
from rougail.convert import RougailConvert
|
||||||
|
from rougail.objspace import RootRougailObject
|
||||||
|
from risotto.utils import EXTRA_ANNOTATORS, ROUGAIL_NAMESPACE, ROUGAIL_NAMESPACE_DESCRIPTION
|
||||||
|
from risotto.image import load_application_service
|
||||||
|
|
||||||
|
|
||||||
|
rougailconfig = RougailConfig
|
||||||
|
rougailconfig['variable_namespace'] = ROUGAIL_NAMESPACE
|
||||||
|
rougailconfig['variable_namespace_description'] = ROUGAIL_NAMESPACE_DESCRIPTION
|
||||||
|
|
||||||
|
|
||||||
|
DEFAULT_TYPE = 'string'
|
||||||
|
ROUGAIL_VARIABLE_TYPE = 'https://forge.cloud.silique.fr/risotto/rougail/src/branch/main/doc/variable/README.md#le-type-de-la-variable'
|
||||||
|
|
||||||
|
|
||||||
|
def add_title_family(elts, dico):
|
||||||
|
for idx, elt in enumerate(elts):
|
||||||
|
description = elt.doc
|
||||||
|
if not idx:
|
||||||
|
description = description.capitalize()
|
||||||
|
space = idx + 3
|
||||||
|
title = '#' * space + f' {description} (*{elt.path}*)'
|
||||||
|
if title not in dico:
|
||||||
|
dico[title] = {'variables': [], 'help': '', 'type': ''}
|
||||||
|
if hasattr(elt, 'information') and hasattr(elt.information, 'help'):
|
||||||
|
dico[title]['help'] = elt.information.help
|
||||||
|
if hasattr(elt, 'suffixes') and elt.suffixes:
|
||||||
|
dico[title]['type'] = 'dynamic'
|
||||||
|
dico[title]['suffixes'] = elt.suffixes.path
|
||||||
|
if hasattr(elt, 'leadership') and elt.leadership:
|
||||||
|
dico[title]['type'] = 'leadership'
|
||||||
|
return title
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
def parse(applicationservice, elts, dico, providers_suppliers, hidden, objectspace):
|
||||||
|
elt = elts[-1]
|
||||||
|
first_variable = True
|
||||||
|
if not hidden:
|
||||||
|
hidden = hasattr(elt, 'properties') and ('hidden' in elt.properties or 'disabled' in elt.properties)
|
||||||
|
is_leadership = hasattr(elt, 'leadership') and elt.leadership
|
||||||
|
for children in vars(elt).values():
|
||||||
|
if isinstance(children, dict):
|
||||||
|
children = list(children.values())
|
||||||
|
if not isinstance(children, list):
|
||||||
|
continue
|
||||||
|
for idx, child in enumerate(children):
|
||||||
|
if isinstance(child, objectspace.property_) or \
|
||||||
|
not isinstance(child, RootRougailObject):
|
||||||
|
continue
|
||||||
|
if isinstance(child, objectspace.variable):
|
||||||
|
if not hidden and (not hasattr(child, 'properties') or ('hidden' not in child.properties and not 'disabled' in child.properties)):
|
||||||
|
if first_variable:
|
||||||
|
title = add_title_family(elts, dico)
|
||||||
|
first_variable = False
|
||||||
|
var_title = child.doc
|
||||||
|
if hasattr(child, 'properties') and 'mandatory' in child.properties:
|
||||||
|
var_title = '**' + var_title + '**'
|
||||||
|
var_path = child.xmlfiles[-1].split('/', 2)[-1]
|
||||||
|
if child.doc != child.name:
|
||||||
|
var_title += f' (*[{child.name}]({var_path})*)'
|
||||||
|
else:
|
||||||
|
var_title = f'*[{var_title}]({var_path})*'
|
||||||
|
if ((idx == 0 or not is_leadership) and child.multi is True) or (idx != 0 and is_leadership and child.multi == 'submulti'):
|
||||||
|
var_title += ' [+]'
|
||||||
|
values = {'description': var_title,
|
||||||
|
}
|
||||||
|
if hasattr(child, 'information') and hasattr(child.information, 'help'):
|
||||||
|
values['help'] = child.information.help
|
||||||
|
if child.type != DEFAULT_TYPE:
|
||||||
|
values['type'] = child.type
|
||||||
|
if hasattr(child, 'default'):
|
||||||
|
default = child.default
|
||||||
|
if isinstance(default, objectspace.value):
|
||||||
|
default = '<calculated>'
|
||||||
|
if isinstance(default, list):
|
||||||
|
default = '<br />'.join(default)
|
||||||
|
values['values'] = default
|
||||||
|
if hasattr(child, 'choice'):
|
||||||
|
values['choices'] = '<br />'.join([choice.name for choice in child.choice])
|
||||||
|
if hasattr(child, 'provider'):
|
||||||
|
provider = child.provider
|
||||||
|
values['provider'] = provider
|
||||||
|
if ':' not in provider:
|
||||||
|
providers_suppliers['providers'].setdefault(provider, []).append(applicationservice)
|
||||||
|
if hasattr(child, 'supplier'):
|
||||||
|
supplier = child.supplier
|
||||||
|
values['supplier'] = supplier
|
||||||
|
if ':' not in supplier:
|
||||||
|
providers_suppliers['suppliers'].setdefault(supplier, []).append(applicationservice)
|
||||||
|
dico[title]['variables'].append(values)
|
||||||
|
else:
|
||||||
|
if hasattr(child, 'provider'):
|
||||||
|
provider = child.provider
|
||||||
|
if ':' not in provider:
|
||||||
|
providers_suppliers['providers'].setdefault(provider, []).append(applicationservice)
|
||||||
|
if hasattr(child, 'supplier'):
|
||||||
|
supplier = child.supplier
|
||||||
|
if ':' not in supplier:
|
||||||
|
providers_suppliers['suppliers'].setdefault(supplier, []).append(applicationservice)
|
||||||
|
else:
|
||||||
|
parse(applicationservice, elts + [child], dico, providers_suppliers, hidden, objectspace)
|
||||||
|
|
||||||
|
|
||||||
|
def build_dependencies_tree(applicationservice, applicationservice_data, applicationservices_data, applicationservices_data_ext, space):
|
||||||
|
depends = []
|
||||||
|
if applicationservice_data['depends']:
|
||||||
|
if applicationservice in applicationservices_data:
|
||||||
|
app_data = applicationservices_data[applicationservice]
|
||||||
|
else:
|
||||||
|
for url, apps_data in applicationservices_data_ext.items():
|
||||||
|
if applicationservice in apps_data:
|
||||||
|
app_data = apps_data[applicationservice]
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
raise Exception(f'cannot find applicationservice "{applicationservice}"')
|
||||||
|
for idx, depend in enumerate(app_data['depends']):
|
||||||
|
if depend in applicationservices_data:
|
||||||
|
url = '..'
|
||||||
|
ext = False
|
||||||
|
else:
|
||||||
|
for url, apps_data in applicationservices_data_ext.items():
|
||||||
|
if depend in apps_data:
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
raise Exception(f'cannot find applicationservice "{applicationservice}"')
|
||||||
|
ext = True
|
||||||
|
subdepends = build_dependencies_tree(depend, applicationservice_data, applicationservices_data, applicationservices_data_ext, space + 2)
|
||||||
|
if not idx or subdepends:
|
||||||
|
title = '\n'
|
||||||
|
else:
|
||||||
|
title = ''
|
||||||
|
depend_desc = depend
|
||||||
|
if ext:
|
||||||
|
depend_desc += ' (in external dataset)'
|
||||||
|
title = ' ' * space + f'- [{depend_desc}]({url}/{depend}/README.md)'
|
||||||
|
depends.append(title)
|
||||||
|
depends.extend(subdepends)
|
||||||
|
return depends
|
||||||
|
|
||||||
|
|
||||||
|
def load_data(url, directory, applicationservices_data, global_data={}):
|
||||||
|
root_path = join(directory, 'seed')
|
||||||
|
applicationservices = listdir(root_path)
|
||||||
|
tmps = {}
|
||||||
|
for applicationservice in applicationservices:
|
||||||
|
as_dir = join(root_path, applicationservice)
|
||||||
|
if not isdir(as_dir):
|
||||||
|
continue
|
||||||
|
applicationservice_data = load_application_service(as_dir)
|
||||||
|
if not applicationservice_data.get('documentation', True):
|
||||||
|
continue
|
||||||
|
applicationservices_data[applicationservice] = {'description': applicationservice_data['description'],
|
||||||
|
'website': applicationservice_data.get('website'),
|
||||||
|
'as_dir': as_dir,
|
||||||
|
'depends': [],
|
||||||
|
'used_by': [],
|
||||||
|
}
|
||||||
|
if applicationservice in tmps:
|
||||||
|
for app in tmps.pop(applicationservice):
|
||||||
|
used_by = f'[{app}](../{app}/README.md)'
|
||||||
|
applicationservices_data[applicationservice]['used_by'].append(used_by)
|
||||||
|
if 'depends' in applicationservice_data:
|
||||||
|
for depend in applicationservice_data['depends']:
|
||||||
|
applicationservices_data[applicationservice]['depends'].append(depend)
|
||||||
|
if depend in applicationservices_data:
|
||||||
|
used_by = f'[{applicationservice}](../{applicationservice}/README.md)'
|
||||||
|
applicationservices_data[depend]['used_by'].append(used_by)
|
||||||
|
else:
|
||||||
|
tmps.setdefault(depend, []).append(applicationservice)
|
||||||
|
if tmps and global_data:
|
||||||
|
for depend, applications in tmps.items():
|
||||||
|
for app in applications:
|
||||||
|
used_by = f'[{app} (in external dataset)]({url}/{app}/README.md)'
|
||||||
|
global_data[depend]['used_by'].append(used_by)
|
||||||
|
|
||||||
|
|
||||||
|
def write_data(applicationservices_data, applicationservices_data_ext):
|
||||||
|
dico = {}
|
||||||
|
providers_suppliers = {'providers': {}, 'suppliers': {}}
|
||||||
|
for applicationservice, applicationservice_data in applicationservices_data.items():
|
||||||
|
as_dir = applicationservice_data['as_dir']
|
||||||
|
dirname = join(as_dir, 'dictionaries')
|
||||||
|
if isdir(dirname):
|
||||||
|
rougailconfig['dictionaries_dir'] = [dirname]
|
||||||
|
else:
|
||||||
|
rougailconfig['dictionaries_dir'] = []
|
||||||
|
dirname_extras = join(as_dir, 'extras')
|
||||||
|
extra_dictionaries = {}
|
||||||
|
if isdir(dirname_extras):
|
||||||
|
for extra in listdir(dirname_extras):
|
||||||
|
extra_dir = join(dirname_extras, extra)
|
||||||
|
if isdir(extra_dir):
|
||||||
|
extra_dictionaries.setdefault(extra, []).append(extra_dir)
|
||||||
|
if not isdir(dirname) and not extra_dictionaries:
|
||||||
|
continue
|
||||||
|
rougailconfig['extra_dictionaries'] = extra_dictionaries
|
||||||
|
converted = RougailConvert(rougailconfig, just_doc=True)
|
||||||
|
converted.load_dictionaries()
|
||||||
|
converted.annotate()
|
||||||
|
objectspace = converted.rougailobjspace
|
||||||
|
if hasattr(objectspace.space, 'variables'):
|
||||||
|
dico[applicationservice] = {}
|
||||||
|
for name, elt in objectspace.space.variables.items():
|
||||||
|
parse(applicationservice, [elt], dico[applicationservice], providers_suppliers, False, objectspace)
|
||||||
|
for applicationservice, applicationservice_data in applicationservices_data.items():
|
||||||
|
as_dir = applicationservice_data['as_dir']
|
||||||
|
with open(join(as_dir, 'README.md'), 'w') as as_fh:
|
||||||
|
as_fh.write(f'---\ngitea: none\ninclude_toc: true\n---\n\n')
|
||||||
|
as_fh.write(f'# {applicationservice}\n\n')
|
||||||
|
as_fh.write(f'## Description\n\n')
|
||||||
|
description = applicationservice_data['description'] + '.\n'
|
||||||
|
if applicationservice_data['website']:
|
||||||
|
description += f'\n[For more informations]({applicationservice_data["website"]})\n'
|
||||||
|
as_fh.write(description)
|
||||||
|
if applicationservice_data['depends']:
|
||||||
|
as_fh.write(f'\n## Dependances\n\n')
|
||||||
|
for depend in build_dependencies_tree(applicationservice, applicationservice_data, applicationservices_data, applicationservices_data_ext, 0):
|
||||||
|
as_fh.write(f'{depend}\n')
|
||||||
|
if applicationservice in dico and dico[applicationservice]:
|
||||||
|
as_fh.write('\n## Variables\n\n')
|
||||||
|
for title, data in dico[applicationservice].items():
|
||||||
|
as_fh.write(f'{title}\n')
|
||||||
|
if data['type'] == 'leadership':
|
||||||
|
as_fh.write('\nThis a family is a leadership.\n')
|
||||||
|
if data['type'] == 'dynamic':
|
||||||
|
as_fh.write(f'\nThis a dynamic family generated from the variable "{data["suffixes"]}".\n')
|
||||||
|
if data['help']:
|
||||||
|
as_fh.write(f'\n{data["help"]}\n')
|
||||||
|
keys = []
|
||||||
|
if data['variables']:
|
||||||
|
variables = data['variables']
|
||||||
|
for variable in variables:
|
||||||
|
for key in variable:
|
||||||
|
if key not in keys:
|
||||||
|
keys.append(key)
|
||||||
|
values = []
|
||||||
|
for variable in variables:
|
||||||
|
value = []
|
||||||
|
for key in keys:
|
||||||
|
if key in variable:
|
||||||
|
val = variable[key]
|
||||||
|
elif key == 'type':
|
||||||
|
val = DEFAULT_TYPE
|
||||||
|
else:
|
||||||
|
val = ''
|
||||||
|
if key == 'type':
|
||||||
|
val = f'[{val}]({ROUGAIL_VARIABLE_TYPE})'
|
||||||
|
value.append(val)
|
||||||
|
values.append(value)
|
||||||
|
as_fh.write('\n')
|
||||||
|
as_fh.write(tabulate(values, headers=[key.capitalize() for key in keys], tablefmt="github"))
|
||||||
|
as_fh.write('\n')
|
||||||
|
as_fh.write('\n')
|
||||||
|
# FIXME if not applicationservice_data['used_by']:
|
||||||
|
# FIXME as_fh.write('\n## Variables with dependencies\n\n')
|
||||||
|
as_fh.write('\n- [+]: variable is multiple\n- **bold**: variable is mandatory\n')
|
||||||
|
if applicationservice_data['used_by']:
|
||||||
|
as_fh.write('\n## Used by\n\n')
|
||||||
|
if len(applicationservice_data['used_by']) == 1:
|
||||||
|
link = applicationservice_data['used_by'][0]
|
||||||
|
as_fh.write(f'{link}\n')
|
||||||
|
else:
|
||||||
|
for link in applicationservice_data['used_by']:
|
||||||
|
as_fh.write(f'- {link}\n')
|
||||||
|
linked = []
|
||||||
|
for provider, provider_as in providers_suppliers['providers'].items():
|
||||||
|
if not applicationservice in provider_as:
|
||||||
|
continue
|
||||||
|
for supplier in providers_suppliers['suppliers'][provider]:
|
||||||
|
if supplier in linked:
|
||||||
|
continue
|
||||||
|
linked.append(supplier)
|
||||||
|
linked.sort()
|
||||||
|
if linked:
|
||||||
|
if len(linked) == 1:
|
||||||
|
as_fh.write('\n## Supplier\n\n')
|
||||||
|
as_fh.write(f'[{linked[0]}](../{linked[0]}/README.md)\n')
|
||||||
|
else:
|
||||||
|
as_fh.write('\n## Suppliers\n\n')
|
||||||
|
for supplier in linked:
|
||||||
|
as_fh.write(f'- [{supplier}](../{supplier}/README.md)\n')
|
||||||
|
linked = []
|
||||||
|
for supplier, supplier_as in providers_suppliers['suppliers'].items():
|
||||||
|
if not applicationservice in supplier_as:
|
||||||
|
continue
|
||||||
|
for provider in providers_suppliers['providers'][supplier]:
|
||||||
|
if provider in linked:
|
||||||
|
continue
|
||||||
|
linked.append(provider)
|
||||||
|
linked.sort()
|
||||||
|
if linked:
|
||||||
|
if len(linked) == 1:
|
||||||
|
as_fh.write('\n## Provider\n\n')
|
||||||
|
as_fh.write(f'[{linked[0]}](../{linked[0]}/README.md)\n')
|
||||||
|
else:
|
||||||
|
as_fh.write('\n## Providers\n\n')
|
||||||
|
for provider in linked:
|
||||||
|
as_fh.write(f'- [{provider}](../{provider}/README.md)\n')
|
||||||
|
as_fh.write(f'\n[All applications services for this dataset.](../README.md)\n')
|
||||||
|
|
||||||
|
with open('seed/README.md', 'w') as as_fh:
|
||||||
|
as_fh.write('# Application services\n\n')
|
||||||
|
applicationservices = {}
|
||||||
|
for applicationservice in applicationservices_data:
|
||||||
|
applicationservices.setdefault(applicationservice.split('-')[0], []).append(applicationservice)
|
||||||
|
applicationservice_categories = list(applicationservices.keys())
|
||||||
|
applicationservice_categories.sort()
|
||||||
|
for category in applicationservice_categories:
|
||||||
|
applicationservices_ = applicationservices[category]
|
||||||
|
if len(applicationservices_) == 1:
|
||||||
|
applicationservice = applicationservices_[0]
|
||||||
|
applicationservice_data = applicationservices_data[applicationservice]
|
||||||
|
as_fh.write(f'- [{applicationservice}]({applicationservice}/README.md): {applicationservice_data["description"]}\n')
|
||||||
|
else:
|
||||||
|
as_fh.write(f'- {category}:\n')
|
||||||
|
applicationservices_.sort()
|
||||||
|
for applicationservice in applicationservices_:
|
||||||
|
applicationservice_data = applicationservices_data[applicationservice]
|
||||||
|
as_fh.write(f' - [{applicationservice}]({applicationservice}/README.md): {applicationservice_data["description"]}\n')
|
||||||
|
providers = list(providers_suppliers['providers'].keys())
|
||||||
|
providers.sort()
|
||||||
|
if providers:
|
||||||
|
as_fh.write('\n# Providers and suppliers\n\n')
|
||||||
|
for provider in providers:
|
||||||
|
as_fh.write(f'- {provider}:\n')
|
||||||
|
if providers_suppliers['providers'][provider]:
|
||||||
|
if len(providers_suppliers['providers'][provider]) == 1:
|
||||||
|
applicationservice = providers_suppliers['providers'][provider][0]
|
||||||
|
as_fh.write(f' - Provider: [{applicationservice}]({applicationservice}/README.md)\n')
|
||||||
|
else:
|
||||||
|
as_fh.write(f' - Providers:\n')
|
||||||
|
for applicationservice in providers_suppliers['providers'][provider]:
|
||||||
|
as_fh.write(f' - [{applicationservice}]({applicationservice}/README.md)\n')
|
||||||
|
if providers_suppliers['suppliers']:
|
||||||
|
if len(providers_suppliers['suppliers'][provider]) == 1:
|
||||||
|
applicationservice = providers_suppliers['suppliers'][provider][0]
|
||||||
|
as_fh.write(f' - Supplier: [{applicationservice}]({applicationservice}/README.md)\n')
|
||||||
|
else:
|
||||||
|
as_fh.write(f' - Suppliers:\n')
|
||||||
|
for applicationservice in providers_suppliers['suppliers'][provider]:
|
||||||
|
as_fh.write(f' - [{applicationservice}]({applicationservice}/README.md)\n')
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
applicationservices_data = {}
|
||||||
|
load_data('..', '', applicationservices_data)
|
||||||
|
applicationservices_data_ext = {}
|
||||||
|
for arg in argv[1:]:
|
||||||
|
if '=' not in arg:
|
||||||
|
raise Exception(f'cannot parse argument "{arg}", should be dataset_path=url')
|
||||||
|
path, url = arg.split('=', 1)
|
||||||
|
if url in applicationservices_data_ext:
|
||||||
|
raise Exception(f'duplicate url "{url}" in arguments')
|
||||||
|
applicationservices_data_ext[url] = {}
|
||||||
|
load_data(url, path, applicationservices_data_ext[url], applicationservices_data)
|
||||||
|
write_data(applicationservices_data, applicationservices_data_ext)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
main()
|
28
sbin/risotto_check_certificates
Executable file
|
@ -0,0 +1,28 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
from os import walk
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
|
||||||
|
week_day = datetime.now().isocalendar().week
|
||||||
|
week_cert = f'certificate_{week_day}.crt'
|
||||||
|
|
||||||
|
|
||||||
|
for p, d, f in walk('pki/x509'):
|
||||||
|
if not d and not f:
|
||||||
|
print('empty dir, you can remove it: ', p)
|
||||||
|
if not f:
|
||||||
|
continue
|
||||||
|
if f == ['serial_number']:
|
||||||
|
continue
|
||||||
|
if not p.endswith('/ca') and not p.endswith('/server') and not p.endswith('/client'):
|
||||||
|
print('unknown directory: ', p)
|
||||||
|
continue
|
||||||
|
if week_cert in f:
|
||||||
|
continue
|
||||||
|
for ff in f:
|
||||||
|
if ff.startswith('certificate_') and ff.endswith('.crt'):
|
||||||
|
print(f'old certificat in: ', p)
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
print('cannot find certificat in: ', p)
|
235
sbin/risotto_display
Executable file
|
@ -0,0 +1,235 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
from tabulate import tabulate
|
||||||
|
from argparse import ArgumentParser
|
||||||
|
|
||||||
|
from rougail.utils import normalize_family
|
||||||
|
from tiramisu.error import PropertiesOptionError
|
||||||
|
from risotto.machine import load, remove_cache, ROUGAIL_NAMESPACE
|
||||||
|
|
||||||
|
|
||||||
|
HIDE_SECRET = True
|
||||||
|
|
||||||
|
|
||||||
|
def list_to_string(lst):
|
||||||
|
if isinstance(lst, list):
|
||||||
|
return "\n".join([str(val) for val in lst])
|
||||||
|
return lst
|
||||||
|
|
||||||
|
|
||||||
|
def get_files_subelements(type_name, element, files_subelement, files_cols):
|
||||||
|
data = {}
|
||||||
|
if not element.option('activate').value.get():
|
||||||
|
return data
|
||||||
|
for subelement in files_subelement.values():
|
||||||
|
if subelement['type'] == 'subelement':
|
||||||
|
try:
|
||||||
|
value = list_to_string(element.option(subelement['key']).value.get())
|
||||||
|
# FIXME except AttributeError:
|
||||||
|
except Exception:
|
||||||
|
value = ''
|
||||||
|
elif subelement['type'] == 'information':
|
||||||
|
value = element.information.get(subelement['key'], '')
|
||||||
|
elif subelement['type'] == 'none':
|
||||||
|
value = subelement['value']
|
||||||
|
else:
|
||||||
|
raise Exception('unknown subelement')
|
||||||
|
if value != '':
|
||||||
|
files_cols.add(subelement['key'])
|
||||||
|
data[subelement['key']] = value
|
||||||
|
if type_name == 'overrides':
|
||||||
|
data['name'] = f'/systemd/system/{data["source"]}.d/rougail.conf'
|
||||||
|
if not data['engine']:
|
||||||
|
data['engine'] = 'none'
|
||||||
|
elif not data['engine']:
|
||||||
|
data['engine'] = 'cheetah'
|
||||||
|
return data
|
||||||
|
|
||||||
|
|
||||||
|
def services(config, values):
|
||||||
|
files_subelement = {'Source': {'key': 'source', 'type': 'information'},
|
||||||
|
'Nom': {'key': 'name', 'type': 'subelement'},
|
||||||
|
'Variable': {'key': 'variable', 'type': 'subelement'},
|
||||||
|
'Propriétaire': {'key': 'owner', 'type': 'subelement'},
|
||||||
|
'Groupe': {'key': 'group', 'type': 'subelement'},
|
||||||
|
'Mode': {'key': 'mode', 'type': 'subelement'},
|
||||||
|
'Moteur': {'key': 'engine', 'type': 'information'},
|
||||||
|
}
|
||||||
|
disabled_services = []
|
||||||
|
for service in config.option.list(type="all"):
|
||||||
|
doc = service.option.doc()
|
||||||
|
files_lst = []
|
||||||
|
files_cols = set()
|
||||||
|
if not service.option('manage').value.get():
|
||||||
|
doc += " - unmanaged"
|
||||||
|
if not service.option('activate').value.get():
|
||||||
|
disabled_services.append([doc])
|
||||||
|
else:
|
||||||
|
for type in service.list(type="all"):
|
||||||
|
type_name = type.option.doc()
|
||||||
|
if type_name in ['files', 'overrides']:
|
||||||
|
for element in type.list(type="all"):
|
||||||
|
data = get_files_subelements(type_name, element, files_subelement, files_cols)
|
||||||
|
if data:
|
||||||
|
files_lst.append(data)
|
||||||
|
elif type_name == 'manage':
|
||||||
|
pass
|
||||||
|
elif type_name == 'activate':
|
||||||
|
if not type.value.get():
|
||||||
|
doc += " - unactivated"
|
||||||
|
else:
|
||||||
|
print("FIXME " + type_name)
|
||||||
|
if files_lst:
|
||||||
|
keys = [key for key, val in files_subelement.items() if val['key'] in files_cols]
|
||||||
|
values[doc] = {'keys': keys, 'lst': []}
|
||||||
|
for lst in files_lst:
|
||||||
|
values[doc]['lst'].append([val for key, val in lst.items() if key in files_cols])
|
||||||
|
if disabled_services:
|
||||||
|
values["Services désactivés"] = {'keys': ['Nom'], 'lst': disabled_services}
|
||||||
|
|
||||||
|
|
||||||
|
def table_leader(config, read_only):
|
||||||
|
keys = ['Description']
|
||||||
|
if read_only:
|
||||||
|
keys.append('Cachée')
|
||||||
|
leadership_lst = config.list(type="all")
|
||||||
|
leader = leadership_lst.pop(0)
|
||||||
|
leader_owner = leader.owner.get()
|
||||||
|
follower_names = [follower.option.name() for follower in leadership_lst]
|
||||||
|
doc = leader.option.doc()
|
||||||
|
properties = leader.property.get()
|
||||||
|
if 'mandatory' in properties:
|
||||||
|
doc += '*'
|
||||||
|
name = leader.option.name()
|
||||||
|
lst = [[f'{doc} ({name})']]
|
||||||
|
if read_only:
|
||||||
|
if 'hidden' in properties:
|
||||||
|
hidden = 'oui'
|
||||||
|
else:
|
||||||
|
hidden = ''
|
||||||
|
lst[0].append(hidden)
|
||||||
|
for idx, leader_value in enumerate(leader.value.get()):
|
||||||
|
keys.append(f'Valeur {idx}')
|
||||||
|
keys.append(f'Utilisateur {idx}')
|
||||||
|
lst[0].append(leader_value)
|
||||||
|
lst[0].append(leader_owner)
|
||||||
|
for follower_idx, follower_name in enumerate(follower_names):
|
||||||
|
follower_option = config.option(follower_name, idx)
|
||||||
|
if idx == 0:
|
||||||
|
doc = follower_option.option.doc()
|
||||||
|
properties = follower_option.property.get()
|
||||||
|
if 'mandatory' in properties:
|
||||||
|
doc += '*'
|
||||||
|
name = follower_option.option.name()
|
||||||
|
lst.append([f'{doc} ({name})'])
|
||||||
|
if read_only:
|
||||||
|
if 'hidden' in properties:
|
||||||
|
hidden = 'oui'
|
||||||
|
else:
|
||||||
|
hidden = ''
|
||||||
|
lst[-1].append(hidden)
|
||||||
|
try:
|
||||||
|
lst[follower_idx + 1].append(list_to_string(follower_option.value.get()))
|
||||||
|
lst[follower_idx + 1].append(follower_option.owner.get())
|
||||||
|
except PropertiesOptionError:
|
||||||
|
pass
|
||||||
|
# leader = next leader_iter
|
||||||
|
# if master_values is None:
|
||||||
|
# master_values = subconfig.value.get()
|
||||||
|
return {'keys': keys, 'lst': lst}
|
||||||
|
|
||||||
|
|
||||||
|
def table(config, prefix_len, values, read_only):
|
||||||
|
lst = []
|
||||||
|
for subconfig in config.option.list(type="all"):
|
||||||
|
# prefix = prefix_len * 2 * ' '
|
||||||
|
# if subconfig.option.isoptiondescription():
|
||||||
|
# prefix += '=>'
|
||||||
|
# else:
|
||||||
|
# prefix += '-'
|
||||||
|
# display_str = f'{prefix} {description}'
|
||||||
|
# if name != description:
|
||||||
|
# display_str = f'{display_str} ({name})'
|
||||||
|
name = subconfig.option.name()
|
||||||
|
doc = subconfig.option.doc()
|
||||||
|
if prefix_len == 0 and ROUGAIL_NAMESPACE != name:
|
||||||
|
doc = doc.capitalize()
|
||||||
|
if prefix_len == 0 and name == 'services':
|
||||||
|
values['Services'] = {}
|
||||||
|
services(subconfig, values['Services'])
|
||||||
|
elif subconfig.option.isoptiondescription():
|
||||||
|
od_name = f'{doc} ({(subconfig.option.path()).split(".", 1)[1]})'
|
||||||
|
values[od_name] = None
|
||||||
|
if subconfig.option.isleadership():
|
||||||
|
values[od_name] = table_leader(subconfig, read_only)
|
||||||
|
else:
|
||||||
|
values[od_name] = table(subconfig, prefix_len + 1, values, read_only)
|
||||||
|
else:
|
||||||
|
value = list_to_string(subconfig.value.get())
|
||||||
|
doc = subconfig.option.doc()
|
||||||
|
properties = subconfig.property.get()
|
||||||
|
if 'mandatory' in properties:
|
||||||
|
doc += '*'
|
||||||
|
name = subconfig.option.name()
|
||||||
|
lst.append([f'{doc} ({name})', value])
|
||||||
|
if read_only:
|
||||||
|
if 'hidden' in properties:
|
||||||
|
hidden = 'oui'
|
||||||
|
else:
|
||||||
|
hidden = ''
|
||||||
|
lst[-1].append(hidden)
|
||||||
|
lst[-1].append(subconfig.owner.get())
|
||||||
|
keys = ['Description', 'Valeur']
|
||||||
|
if read_only:
|
||||||
|
keys.append('Cachée')
|
||||||
|
keys.append('Utilisateur')
|
||||||
|
return {'keys': keys, 'lst': lst}
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
parser = ArgumentParser()
|
||||||
|
parser.add_argument('server_name')
|
||||||
|
parser.add_argument('--read_only', action='store_true')
|
||||||
|
parser.add_argument('--nocache', action='store_true')
|
||||||
|
parser.add_argument('--debug', action='store_true')
|
||||||
|
args = parser.parse_args()
|
||||||
|
if args.nocache:
|
||||||
|
remove_cache()
|
||||||
|
|
||||||
|
values = {}
|
||||||
|
server_name = args.server_name
|
||||||
|
config = load(hide_secret=HIDE_SECRET,
|
||||||
|
original_display_name=True,
|
||||||
|
valid_mandatories=args.read_only,
|
||||||
|
)
|
||||||
|
if not args.read_only:
|
||||||
|
config.property.read_write()
|
||||||
|
root_option = config.option(normalize_family(server_name))
|
||||||
|
try:
|
||||||
|
root_option.option.get()
|
||||||
|
except AttributeError:
|
||||||
|
exit(f'Unable to find {server_name} configuration: {[o.option.description() for o in config.option.list(type="optiondescription")]}')
|
||||||
|
table(root_option, 0, values, args.read_only)
|
||||||
|
for title, dico in values.items():
|
||||||
|
if title == 'Services':
|
||||||
|
if not dico:
|
||||||
|
continue
|
||||||
|
print()
|
||||||
|
print(title)
|
||||||
|
print('=' * len(title))
|
||||||
|
print()
|
||||||
|
for subtitle, dic in dico.items():
|
||||||
|
print()
|
||||||
|
print(' ' + subtitle)
|
||||||
|
print(' ' + '-' * len(subtitle))
|
||||||
|
print()
|
||||||
|
print(tabulate(dic['lst'], headers=dic['keys'], tablefmt="fancy_grid"))
|
||||||
|
elif dico['lst']:
|
||||||
|
print()
|
||||||
|
print(title)
|
||||||
|
print('=' * len(title))
|
||||||
|
print()
|
||||||
|
print(tabulate(dico['lst'], headers=dico['keys'], tablefmt="fancy_grid"))
|
||||||
|
|
||||||
|
|
||||||
|
main()
|
34
sbin/risotto_templates
Executable file
|
@ -0,0 +1,34 @@
|
||||||
|
#!/usr/bin/env python3
|
||||||
|
|
||||||
|
from argparse import ArgumentParser
|
||||||
|
from traceback import print_exc
|
||||||
|
|
||||||
|
from risotto.machine import remove_cache, build_files, INSTALL_DIR
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
parser = ArgumentParser()
|
||||||
|
parser.add_argument('server_name', nargs='?')
|
||||||
|
parser.add_argument('--nocache', action='store_true')
|
||||||
|
parser.add_argument('--debug', action='store_true')
|
||||||
|
parser.add_argument('--copy_tests', action='store_true')
|
||||||
|
parser.add_argument('--template')
|
||||||
|
args = parser.parse_args()
|
||||||
|
if args.nocache:
|
||||||
|
remove_cache()
|
||||||
|
|
||||||
|
try:
|
||||||
|
build_files(None,
|
||||||
|
args.server_name,
|
||||||
|
False,
|
||||||
|
args.copy_tests,
|
||||||
|
template=args.template,
|
||||||
|
)
|
||||||
|
except Exception as err:
|
||||||
|
if args.debug:
|
||||||
|
print_exc()
|
||||||
|
exit(err)
|
||||||
|
print(f'templates generated in "{INSTALL_DIR}" directory')
|
||||||
|
|
||||||
|
|
||||||
|
main()
|
273
src/risotto/image.py
Normal file
|
@ -0,0 +1,273 @@
|
||||||
|
from shutil import copy2, copytree
|
||||||
|
from os import listdir, makedirs
|
||||||
|
from os.path import join, isdir, isfile, dirname
|
||||||
|
from yaml import load as yaml_load, SafeLoader
|
||||||
|
from tiramisu.error import PropertiesOptionError
|
||||||
|
from .rougail import func
|
||||||
|
#
|
||||||
|
from .utils import RISOTTO_CONFIG
|
||||||
|
|
||||||
|
|
||||||
|
class ModuleCfg():
|
||||||
|
def __init__(self, module_name):
|
||||||
|
self.module_name = module_name
|
||||||
|
self.dictionaries_dir = []
|
||||||
|
self.functions_file = [func.__file__]
|
||||||
|
self.templates_dir = []
|
||||||
|
self.patches_dir = []
|
||||||
|
self.extra_dictionaries = {}
|
||||||
|
self.servers = []
|
||||||
|
self.depends = []
|
||||||
|
self.manuals = []
|
||||||
|
self.tests = []
|
||||||
|
#self.providers = []
|
||||||
|
#self.suppliers = []
|
||||||
|
|
||||||
|
def __repr__(self):
|
||||||
|
return str(vars(self))
|
||||||
|
|
||||||
|
def load_application_service(as_dir: str) -> str:
|
||||||
|
with open(join(as_dir, 'applicationservice.yml')) as yaml:
|
||||||
|
return yaml_load(yaml, Loader=SafeLoader)
|
||||||
|
|
||||||
|
class Applications:
|
||||||
|
def __init__(self) -> None:
|
||||||
|
self.datasets = RISOTTO_CONFIG.get('directories', {}).get('datasets', ['dataset'])
|
||||||
|
self.application_directories = self._load_application_directories()
|
||||||
|
|
||||||
|
def _load_application_directories(self) -> dict:
|
||||||
|
"""List all service applications in datasets
|
||||||
|
Returns something link:
|
||||||
|
{<applicationservice>: seed/<applicationservice>}
|
||||||
|
"""
|
||||||
|
applications = {'host': None}
|
||||||
|
for dataset_directory in self.datasets:
|
||||||
|
for applicationservice in listdir(dataset_directory):
|
||||||
|
applicationservice_dir = join(dataset_directory, applicationservice)
|
||||||
|
if not isdir(applicationservice_dir) or \
|
||||||
|
not isfile(join(applicationservice_dir, 'applicationservice.yml')):
|
||||||
|
continue
|
||||||
|
if applicationservice in applications:
|
||||||
|
raise Exception(f'multi applicationservice: {applicationservice} ({applicationservice_dir} <=> {applications[applicationservice]})')
|
||||||
|
applications[applicationservice] = applicationservice_dir
|
||||||
|
return applications
|
||||||
|
|
||||||
|
|
||||||
|
class Modules:
|
||||||
|
"""Modules are defined by the end user
|
||||||
|
A module is the a list of service applications
|
||||||
|
The class collects all the useful information for the module
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self,
|
||||||
|
applicationservices: Applications,
|
||||||
|
applicationservice_provider: str,
|
||||||
|
modules_name: list,
|
||||||
|
host_applicationsservice: str,
|
||||||
|
) -> None:
|
||||||
|
self.application_directories = applicationservices.application_directories
|
||||||
|
self.module_infos = {}
|
||||||
|
self.module_infos['host'] = self._load_module_informations('host',
|
||||||
|
['host', host_applicationsservice],
|
||||||
|
is_host=True,
|
||||||
|
)
|
||||||
|
for module_name in modules_name:
|
||||||
|
if module_name == 'host':
|
||||||
|
raise Exception('forbidden module name: "host"')
|
||||||
|
self.module_infos[module_name] = self._load_module_informations(module_name,
|
||||||
|
[applicationservice_provider, module_name],
|
||||||
|
is_host=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
def get(self,
|
||||||
|
module_name: str,
|
||||||
|
) -> ModuleCfg:
|
||||||
|
return self.module_infos[module_name]
|
||||||
|
|
||||||
|
def _load_module_informations(self,
|
||||||
|
module_name: str,
|
||||||
|
applicationservices: list,
|
||||||
|
is_host: bool,
|
||||||
|
) -> ModuleCfg:
|
||||||
|
"""Create a ModuleCfg object and collect informations
|
||||||
|
A module must depend to an unique distribution
|
||||||
|
"""
|
||||||
|
cfg = ModuleCfg(module_name)
|
||||||
|
distribution = None
|
||||||
|
|
||||||
|
for applicationservice in applicationservices:
|
||||||
|
ret = self._load_applicationservice(applicationservice,
|
||||||
|
cfg,
|
||||||
|
)
|
||||||
|
if ret:
|
||||||
|
if distribution:
|
||||||
|
raise Exception(f'duplicate distribution for {cfg.module_name}: {distribution} and {ret} (dependencies: {cfg.depends}) ')
|
||||||
|
distribution = ret
|
||||||
|
if not is_host and not distribution:
|
||||||
|
raise Exception(f'cannot found any linux distribution for {module_name}')
|
||||||
|
return cfg
|
||||||
|
|
||||||
|
def _load_applicationservice(self,
|
||||||
|
appname: str,
|
||||||
|
cfg: ModuleCfg,
|
||||||
|
) -> str:
|
||||||
|
"""extract informations from an application service and load it's dependency
|
||||||
|
informations collected is store to the module
|
||||||
|
|
||||||
|
returns the name of current distribution, if found
|
||||||
|
"""
|
||||||
|
if appname not in self.application_directories:
|
||||||
|
raise Exception(f'cannot find application dependency "{appname}"')
|
||||||
|
cfg.depends.append(appname)
|
||||||
|
as_dir = self.application_directories[appname]
|
||||||
|
if not as_dir:
|
||||||
|
return
|
||||||
|
self._load_applicationservice_directories(as_dir,
|
||||||
|
cfg,
|
||||||
|
)
|
||||||
|
app = load_application_service(as_dir)
|
||||||
|
provider = app.get('provider')
|
||||||
|
if provider:
|
||||||
|
cfg.providers.setdefault(provider, [])
|
||||||
|
if appname not in cfg.providers[provider]:
|
||||||
|
cfg.providers[provider].append(appname)
|
||||||
|
supplier = app.get('supplier')
|
||||||
|
#if supplier:
|
||||||
|
# self.suppliers.setdefault(supplier, [])
|
||||||
|
# if appname not in self.suppliers[supplier]:
|
||||||
|
# self.suppliers[supplier].append(appname)
|
||||||
|
if 'distribution' in app and app['distribution']:
|
||||||
|
distribution = appname
|
||||||
|
else:
|
||||||
|
distribution = None
|
||||||
|
for depend in app.get('depends', []):
|
||||||
|
if depend in cfg.depends:
|
||||||
|
#this dependancy is already loaded for this module
|
||||||
|
continue
|
||||||
|
ret = self._load_applicationservice(depend,
|
||||||
|
cfg,
|
||||||
|
)
|
||||||
|
if ret:
|
||||||
|
if distribution:
|
||||||
|
raise Exception(f'duplicate distribution for {cfg.module_name}: {distribution} and {ret} (dependencies: {cfg.depends}) ')
|
||||||
|
distribution = ret
|
||||||
|
return distribution
|
||||||
|
|
||||||
|
def _load_applicationservice_directories(self,
|
||||||
|
as_dir: str,
|
||||||
|
cfg: ModuleCfg,
|
||||||
|
) -> None:
|
||||||
|
# dictionaries
|
||||||
|
dictionaries_dir = join(as_dir, 'dictionaries')
|
||||||
|
if isdir(dictionaries_dir):
|
||||||
|
cfg.dictionaries_dir.append(dictionaries_dir)
|
||||||
|
# funcs
|
||||||
|
funcs_dir = join(as_dir, 'funcs')
|
||||||
|
if isdir(funcs_dir):
|
||||||
|
for f in listdir(funcs_dir):
|
||||||
|
if f.startswith('__'):
|
||||||
|
continue
|
||||||
|
cfg.functions_file.append(join(funcs_dir, f))
|
||||||
|
# templates
|
||||||
|
templates_dir = join(as_dir, 'templates')
|
||||||
|
if isdir(templates_dir):
|
||||||
|
cfg.templates_dir.append(templates_dir)
|
||||||
|
# patches
|
||||||
|
patches_dir = join(as_dir, 'patches')
|
||||||
|
if isdir(patches_dir):
|
||||||
|
cfg.patches_dir.append(patches_dir)
|
||||||
|
# extras
|
||||||
|
extras_dir = join(as_dir, 'extras')
|
||||||
|
if isdir(extras_dir):
|
||||||
|
for extra in listdir(extras_dir):
|
||||||
|
extra_dir = join(extras_dir, extra)
|
||||||
|
if isdir(extra_dir):
|
||||||
|
cfg.extra_dictionaries.setdefault(extra, []).append(extra_dir)
|
||||||
|
# manual
|
||||||
|
manual_dir = join(as_dir, 'manual', 'image')
|
||||||
|
if isdir(manual_dir):
|
||||||
|
cfg.manuals.append(manual_dir)
|
||||||
|
# tests
|
||||||
|
tests_dir = join(as_dir, 'tests')
|
||||||
|
if isdir(tests_dir):
|
||||||
|
cfg.tests.append(tests_dir)
|
||||||
|
|
||||||
|
|
||||||
|
def applicationservice_copy(src_file: str,
|
||||||
|
dst_file: str,
|
||||||
|
) -> None:
|
||||||
|
if isdir(src_file):
|
||||||
|
if not isdir(dst_file):
|
||||||
|
makedirs(dst_file)
|
||||||
|
for subfilename in listdir(src_file):
|
||||||
|
#if not copy_if_not_exists or not isfile(dst_file):
|
||||||
|
src = join(src_file, subfilename)
|
||||||
|
dst = join(dst_file, subfilename)
|
||||||
|
if isfile(src):
|
||||||
|
copy2(src, dst)
|
||||||
|
else:
|
||||||
|
copytree(src, dst)
|
||||||
|
else:
|
||||||
|
dst = dirname(dst_file)
|
||||||
|
if not isdir(dst):
|
||||||
|
makedirs(dst)
|
||||||
|
if isfile(src_file):
|
||||||
|
copy2(src_file, dst_file)
|
||||||
|
else:
|
||||||
|
copytree(src_file, dst_file)
|
||||||
|
|
||||||
|
|
||||||
|
def valid_mandatories(config):
|
||||||
|
mandatories = config.value.mandatory()
|
||||||
|
config.property.remove('mandatory')
|
||||||
|
hidden = {}
|
||||||
|
variables = {}
|
||||||
|
title = None
|
||||||
|
if mandatories:
|
||||||
|
server_name = None
|
||||||
|
for mandatory_option in mandatories:
|
||||||
|
path_server_name, path = mandatory_option.path().split('.', 1)
|
||||||
|
var_server_name = config.option(path_server_name).description()
|
||||||
|
if server_name != var_server_name:
|
||||||
|
server_name = var_server_name
|
||||||
|
title = f'=== Missing variables for "{server_name.split(".", 1)[0]}" ==='
|
||||||
|
text = mandatory_option.doc()
|
||||||
|
msg = f' - {text} ({path})'
|
||||||
|
supplier = mandatory_option.information.get('supplier', None)
|
||||||
|
if supplier and not ':' in supplier:
|
||||||
|
msg += f' you should add a service that provides "{supplier}"'
|
||||||
|
if mandatory_option.isfollower():
|
||||||
|
leader = mandatory_option.leader()
|
||||||
|
try:
|
||||||
|
leader_value = leader.value.get()
|
||||||
|
except PropertiesOptionError as err:
|
||||||
|
if 'hidden' not in err.proptype:
|
||||||
|
raise err from err
|
||||||
|
hidden.setdefault(title, []).append(msg)
|
||||||
|
else:
|
||||||
|
config.property.add('mandatory')
|
||||||
|
for idx in range(mandatory_option.value.len()):
|
||||||
|
try:
|
||||||
|
config.option(mandatory_option.path(), idx).value.get()
|
||||||
|
except PropertiesOptionError as err:
|
||||||
|
path = leader.path()
|
||||||
|
spath = path.split('.', 1)[1]
|
||||||
|
submsg = f'{msg} at index {idx} (value of leader "{leader.doc()}" ({spath}) is "{leader_value[idx]}")'
|
||||||
|
if 'hidden' in err.proptype:
|
||||||
|
hidden.setdefault(title, []).append(submsg)
|
||||||
|
elif 'mandatory' in err.proptype:
|
||||||
|
variables.setdefault(title, []).append(submsg)
|
||||||
|
else:
|
||||||
|
raise err from err
|
||||||
|
config.property.remove('mandatory')
|
||||||
|
else:
|
||||||
|
try:
|
||||||
|
mandatory_option.value.get()
|
||||||
|
variables.setdefault(title, []).append(msg)
|
||||||
|
except PropertiesOptionError as err:
|
||||||
|
if 'hidden' not in err.proptype:
|
||||||
|
raise err from err
|
||||||
|
hidden.setdefault(title, []).append(msg)
|
||||||
|
if not variables:
|
||||||
|
variables = hidden
|
||||||
|
return variables
|
601
src/risotto/machine.py
Normal file
|
@ -0,0 +1,601 @@
|
||||||
|
from .utils import MULTI_FUNCTIONS, load_zones, value_pprint, RISOTTO_CONFIG, EXTRA_ANNOTATORS, ROUGAIL_NAMESPACE, ROUGAIL_NAMESPACE_DESCRIPTION
|
||||||
|
from .image import Applications, Modules, valid_mandatories, applicationservice_copy
|
||||||
|
|
||||||
|
from rougail import RougailConfig, RougailConvert
|
||||||
|
from os import remove, makedirs, listdir, chmod
|
||||||
|
from os.path import isfile, isdir, abspath, join, dirname
|
||||||
|
from pickle import dump as pickle_dump, load as pickle_load
|
||||||
|
from yaml import load as yaml_load, SafeLoader
|
||||||
|
from ipaddress import IPv4Interface, ip_network
|
||||||
|
#
|
||||||
|
from tiramisu import Config
|
||||||
|
from rougail.utils import normalize_family
|
||||||
|
from rougail import RougailSystemdTemplate
|
||||||
|
from shutil import copy2, copytree, rmtree
|
||||||
|
|
||||||
|
|
||||||
|
def tiramisu_display_name(kls,
|
||||||
|
dyn_name: 'Base'=None,
|
||||||
|
suffix: str=None,
|
||||||
|
) -> str:
|
||||||
|
# FIXME
|
||||||
|
if dyn_name is not None:
|
||||||
|
name = kls.impl_getpath() + str(suffix)
|
||||||
|
else:
|
||||||
|
name = kls.impl_getpath()
|
||||||
|
return name
|
||||||
|
|
||||||
|
|
||||||
|
CONFIG_FILE = 'servers.yml'
|
||||||
|
TIRAMISU_CACHE = 'tiramisu_cache.py'
|
||||||
|
VALUES_CACHE = 'values_cache.pickle'
|
||||||
|
INFORMATIONS_CACHE = 'informations_cache.pickle'
|
||||||
|
INSTALL_DIR = RISOTTO_CONFIG['directories']['dest']
|
||||||
|
INSTALL_CONFIG_DIR = 'configurations'
|
||||||
|
INSTALL_TMPL_DIR= 'templates'
|
||||||
|
INSTALL_IMAGES_DIR = 'images_files'
|
||||||
|
INSTALL_TESTS_DIR = 'tests'
|
||||||
|
|
||||||
|
|
||||||
|
def copy(src_file, dst_file):
|
||||||
|
if isdir(src_file):
|
||||||
|
if not isdir(dst_file):
|
||||||
|
makedirs(dst_file)
|
||||||
|
for subfilename in listdir(src_file):
|
||||||
|
if not isfile(dst_file):
|
||||||
|
src = join(src_file, subfilename)
|
||||||
|
dst = join(dst_file, subfilename)
|
||||||
|
if isfile(src):
|
||||||
|
copy2(src, dst)
|
||||||
|
else:
|
||||||
|
copytree(src, dst)
|
||||||
|
elif not isfile(dst_file):
|
||||||
|
dst = dirname(dst_file)
|
||||||
|
if not isdir(dst):
|
||||||
|
makedirs(dst)
|
||||||
|
if isfile(src_file):
|
||||||
|
copy2(src_file, dst_file)
|
||||||
|
else:
|
||||||
|
copytree(src_file, dst_file)
|
||||||
|
|
||||||
|
|
||||||
|
def re_create(dir_name):
|
||||||
|
if isdir(dir_name):
|
||||||
|
rmtree(dir_name)
|
||||||
|
makedirs(dir_name)
|
||||||
|
|
||||||
|
|
||||||
|
def remove_cache():
|
||||||
|
if isfile(TIRAMISU_CACHE):
|
||||||
|
remove(TIRAMISU_CACHE)
|
||||||
|
if isfile(VALUES_CACHE):
|
||||||
|
remove(VALUES_CACHE)
|
||||||
|
if isfile(INFORMATIONS_CACHE):
|
||||||
|
remove(INFORMATIONS_CACHE)
|
||||||
|
|
||||||
|
|
||||||
|
def templates(server_name,
|
||||||
|
config,
|
||||||
|
just_copy=False,
|
||||||
|
copy_manuals=False,
|
||||||
|
template=None,
|
||||||
|
extra_variables=None,
|
||||||
|
):
|
||||||
|
subconfig = config.option(normalize_family(server_name))
|
||||||
|
try:
|
||||||
|
subconfig.get()
|
||||||
|
except:
|
||||||
|
servers = [server.description() for server in config.list('optiondescription')]
|
||||||
|
raise Exception(f'cannot find server name "{server_name}": {servers}')
|
||||||
|
|
||||||
|
rougailconfig = RougailConfig.copy()
|
||||||
|
rougailconfig['variable_namespace'] = ROUGAIL_NAMESPACE
|
||||||
|
rougailconfig['variable_namespace_description'] = ROUGAIL_NAMESPACE_DESCRIPTION
|
||||||
|
rougailconfig['tmp_dir'] = 'tmp'
|
||||||
|
rougailconfig['templates_dir'] = subconfig.information.get('templates_dir')
|
||||||
|
rougailconfig['patches_dir'] = subconfig.information.get('patches_dir')
|
||||||
|
rougailconfig['functions_file'] = subconfig.information.get('functions_files')
|
||||||
|
module = subconfig.information.get('module')
|
||||||
|
is_host = module == 'host'
|
||||||
|
if is_host:
|
||||||
|
rougailconfig['systemd_tmpfile_delete_before_create'] = True
|
||||||
|
if just_copy:
|
||||||
|
raise Exception('cannot generate template with option just_copy for a host')
|
||||||
|
else:
|
||||||
|
rougailconfig['systemd_tmpfile_delete_before_create'] = False
|
||||||
|
#rougailconfig['systemd_tmpfile_factory_dir'] = '/usr/local/lib'
|
||||||
|
if not just_copy:
|
||||||
|
rougailconfig['destinations_dir'] = join(INSTALL_DIR, INSTALL_CONFIG_DIR, server_name)
|
||||||
|
else:
|
||||||
|
rougailconfig['destinations_dir'] = join(INSTALL_DIR, INSTALL_TMPL_DIR, server_name)
|
||||||
|
re_create(rougailconfig['destinations_dir'])
|
||||||
|
re_create(rougailconfig['tmp_dir'])
|
||||||
|
|
||||||
|
engine = RougailSystemdTemplate(subconfig,
|
||||||
|
rougailconfig,
|
||||||
|
)
|
||||||
|
if just_copy:
|
||||||
|
# for all engine to none
|
||||||
|
ori_engines = {}
|
||||||
|
for eng in engine.engines:
|
||||||
|
if eng == 'none':
|
||||||
|
continue
|
||||||
|
ori_engines[eng] = engine.engines[eng]
|
||||||
|
engine.engines[eng] = engine.engines['none']
|
||||||
|
try:
|
||||||
|
if not template:
|
||||||
|
engine.instance_files(extra_variables=extra_variables)
|
||||||
|
else:
|
||||||
|
engine.instance_file(template, extra_variables=extra_variables)
|
||||||
|
except Exception as err:
|
||||||
|
print()
|
||||||
|
print(f'=== Configuration: {server_name} ===')
|
||||||
|
try:
|
||||||
|
values = subconfig.value.dict()
|
||||||
|
value_pprint(values, subconfig)
|
||||||
|
except:
|
||||||
|
pass
|
||||||
|
raise err from err
|
||||||
|
if just_copy:
|
||||||
|
for eng, old_engine in ori_engines.items():
|
||||||
|
engine.engines[eng] = old_engine
|
||||||
|
secrets_dir = join(rougailconfig['destinations_dir'], 'secrets')
|
||||||
|
if isdir(secrets_dir):
|
||||||
|
chmod(secrets_dir, 0o700)
|
||||||
|
if copy_manuals and not is_host:
|
||||||
|
dest_dir = join(INSTALL_DIR, INSTALL_IMAGES_DIR, module)
|
||||||
|
if not isdir(dest_dir):
|
||||||
|
for manual in subconfig.information.get('manuals_dirs'):
|
||||||
|
for filename in listdir(manual):
|
||||||
|
src_file = join(manual, filename)
|
||||||
|
dst_file = join(dest_dir, filename)
|
||||||
|
copy(src_file, dst_file)
|
||||||
|
copy_tests = config.information.get('copy_tests')
|
||||||
|
|
||||||
|
if copy_tests and not is_host:
|
||||||
|
dest_dir = join(INSTALL_DIR, INSTALL_TESTS_DIR, module)
|
||||||
|
if not isdir(dest_dir):
|
||||||
|
for tests in subconfig.information.get('tests_dirs'):
|
||||||
|
for filename in listdir(tests):
|
||||||
|
src_file = join(tests, filename)
|
||||||
|
dst_file = join(dest_dir, filename)
|
||||||
|
copy(src_file, dst_file)
|
||||||
|
|
||||||
|
|
||||||
|
class Loader:
|
||||||
|
def __init__(self,
|
||||||
|
hide_secret,
|
||||||
|
original_display_name,
|
||||||
|
valid_mandatories,
|
||||||
|
config_file=CONFIG_FILE,
|
||||||
|
):
|
||||||
|
self.hide_secret = hide_secret
|
||||||
|
self.original_display_name = original_display_name
|
||||||
|
self.valid_mandatories = valid_mandatories
|
||||||
|
self.config_file = config_file
|
||||||
|
|
||||||
|
def load_tiramisu_file(self):
|
||||||
|
"""Load config file (servers.yml) and build tiramisu file with dataset informations
|
||||||
|
"""
|
||||||
|
with open(self.config_file, 'r') as server_fh:
|
||||||
|
self.servers_json = yaml_load(server_fh, Loader=SafeLoader)
|
||||||
|
self.add_tls()
|
||||||
|
# set global rougail configuration
|
||||||
|
cfg = RougailConfig.copy()
|
||||||
|
cfg['variable_namespace'] = ROUGAIL_NAMESPACE
|
||||||
|
cfg['variable_namespace_description'] = ROUGAIL_NAMESPACE_DESCRIPTION
|
||||||
|
cfg['multi_functions'] = MULTI_FUNCTIONS
|
||||||
|
cfg['extra_annotators'] = EXTRA_ANNOTATORS
|
||||||
|
cfg['force_convert_dyn_option_description'] = True
|
||||||
|
cfg['risotto_globals'] = {}
|
||||||
|
|
||||||
|
# initialise variables to store useful informations
|
||||||
|
# those variables are use during templating
|
||||||
|
self.templates_dir = {}
|
||||||
|
self.patches_dir = {}
|
||||||
|
self.functions_files = {}
|
||||||
|
self.manuals_dirs = {}
|
||||||
|
self.tests_dirs = {}
|
||||||
|
self.modules = {}
|
||||||
|
|
||||||
|
functions_files = set()
|
||||||
|
applicationservices = Applications()
|
||||||
|
zones_name = {}
|
||||||
|
rougail = RougailConvert(cfg)
|
||||||
|
for host_name, datas in self.servers_json['hosts'].items():
|
||||||
|
for server_name, server_datas in datas['servers'].items():
|
||||||
|
if 'provider_zone' not in server_datas and 'zones_name' not in server_datas:
|
||||||
|
raise Exception(f'cannot find "zones_name" attribute for server "{server_name}"')
|
||||||
|
if 'provider_zone' in server_datas:
|
||||||
|
zones_name.setdefault(server_datas['provider_zone'], []).append(server_name)
|
||||||
|
if 'zones_name' not in server_datas:
|
||||||
|
server_datas['zones_name'] = []
|
||||||
|
if server_datas['provider_zone'] in server_datas['zones_name']:
|
||||||
|
raise Exception(_('provider_zone "{server_datas["provider_zone"]}" must not be in "zones" "{server_datas["zones_name"]}"'))
|
||||||
|
# external zone is better in first place
|
||||||
|
if server_datas['zones_name'] and self.servers_json['zones']['external_zone'] == server_datas['zones_name'][0]:
|
||||||
|
server_datas['zones_name'].append(server_datas['provider_zone'])
|
||||||
|
else:
|
||||||
|
server_datas['zones_name'].insert(0, server_datas['provider_zone'])
|
||||||
|
# if server_datas['zones_name'] and server_datas['provider_zone'] == self.servers_json['zones']['external_zone']:
|
||||||
|
# server_datas['zones_name'].insert(0, server_datas['provider_zone'])
|
||||||
|
# else:
|
||||||
|
# server_datas['zones_name'].append(server_datas['provider_zone'])
|
||||||
|
for zone in server_datas['zones_name']:
|
||||||
|
zones_name.setdefault(zone, []).append(server_name)
|
||||||
|
self.zones = {}
|
||||||
|
zones_network = ip_network(self.servers_json['zones']['network'])
|
||||||
|
zone_start_ip = zones_network.network_address
|
||||||
|
domain_name = self.servers_json['zones']['prefix_domain_name']
|
||||||
|
for zone_name in zones_name:
|
||||||
|
len_zone = len(zones_name[zone_name])
|
||||||
|
for zone_cidr in [29, 28, 27, 26]:
|
||||||
|
try:
|
||||||
|
sub_network = ip_network(f'{zone_start_ip}/{zone_cidr}')
|
||||||
|
except ValueError:
|
||||||
|
# calc network address for this mask
|
||||||
|
zone_start_ip = IPv4Interface(f'{zone_start_ip}/{zone_cidr}').network.broadcast_address + 1
|
||||||
|
sub_network = ip_network(f'{zone_start_ip}/{zone_cidr}')
|
||||||
|
if not sub_network.subnet_of(zones_network):
|
||||||
|
raise Exception('not enough IP available')
|
||||||
|
length = sub_network.num_addresses - 3 # network + broadcast + host
|
||||||
|
if length >= len_zone:
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
raise Exception(f'network too small for zone "{zone_name}" ({sub_network.num_addresses - 2} < {len_zone})')
|
||||||
|
if self.servers_json['zones']['external_zone'] == zone_name:
|
||||||
|
zone_domaine_name = domain_name
|
||||||
|
else:
|
||||||
|
zone_domaine_name = zone_name + '.' + domain_name
|
||||||
|
network = sub_network.network_address
|
||||||
|
self.zones[zone_name] = {'domain_name': zone_domaine_name,
|
||||||
|
'network': str(sub_network),
|
||||||
|
'host_ip': str(network + 1),
|
||||||
|
'host_name': host_name.split('.', 1)[0],
|
||||||
|
'length': length,
|
||||||
|
'start_ip': str(network + 2)
|
||||||
|
}
|
||||||
|
zone_start_ip = str(sub_network.broadcast_address + 1)
|
||||||
|
|
||||||
|
for host_name, datas in self.servers_json['hosts'].items():
|
||||||
|
# load modules associate to this host
|
||||||
|
modules_name = set()
|
||||||
|
for name, mod_datas in datas['servers'].items():
|
||||||
|
if not 'applicationservice' in mod_datas:
|
||||||
|
raise Exception(f'applicationservice is mandatory for "{name}"')
|
||||||
|
modules_name.add(mod_datas['applicationservice'])
|
||||||
|
# load modules informations from config files
|
||||||
|
modules = Modules(applicationservices,
|
||||||
|
datas['applicationservice_provider'],
|
||||||
|
modules_name,
|
||||||
|
datas['applicationservice'],
|
||||||
|
)
|
||||||
|
|
||||||
|
# load host
|
||||||
|
module_info = modules.get('host')
|
||||||
|
tls_host_name = f'{server_name}.{self.zones[list(self.zones)[0]]["domain_name"]}'
|
||||||
|
short_host_name = host_name.split('.', 1)[0]
|
||||||
|
values = [f'{short_host_name}.{self.zones[zone_name]["domain_name"]}' for zone_name in zones_name]
|
||||||
|
cfg['risotto_globals'][host_name] = {'global:server_name': host_name,
|
||||||
|
'global:server_names': values,
|
||||||
|
'global:zones_name': list(self.zones),
|
||||||
|
'global:module_name': 'host',
|
||||||
|
'global:host_install_dir': abspath(INSTALL_DIR),
|
||||||
|
'global:tls_server': tls_host_name,
|
||||||
|
}
|
||||||
|
functions_files |= set(module_info.functions_file)
|
||||||
|
self.load_dictionaries(cfg,
|
||||||
|
module_info,
|
||||||
|
host_name,
|
||||||
|
rougail,
|
||||||
|
)
|
||||||
|
# load servers
|
||||||
|
modules_info = {}
|
||||||
|
for server_name, server_datas in datas['servers'].items():
|
||||||
|
module_info = modules.get(server_datas['applicationservice'])
|
||||||
|
zones_name = server_datas['zones_name']
|
||||||
|
values = [f'{server_name}.{self.zones[zone_name]["domain_name"]}' for zone_name in zones_name]
|
||||||
|
if server_datas['applicationservice'] == 'tls':
|
||||||
|
true_host_name = f'{server_name}.{self.zones[server_datas["zones_name"][0]]["domain_name"]}'
|
||||||
|
else:
|
||||||
|
true_host_name = values[0]
|
||||||
|
cfg['risotto_globals'][true_host_name] = {'global:host_name': host_name,
|
||||||
|
'global:server_name': true_host_name,
|
||||||
|
'global:server_names': values,
|
||||||
|
'global:zones_name': zones_name,
|
||||||
|
'global:zones_list': list(range(len(zones_name))),
|
||||||
|
'global:module_name': server_datas['applicationservice'],
|
||||||
|
'global:prefix_domain_name': self.servers_json['zones']['prefix_domain_name']
|
||||||
|
}
|
||||||
|
if 'provider_zone' in server_datas:
|
||||||
|
cfg['risotto_globals'][true_host_name]['global:provider_zone'] = server_datas['provider_zone']
|
||||||
|
server_datas['server_name'] = true_host_name
|
||||||
|
functions_files |= set(module_info.functions_file)
|
||||||
|
self.load_dictionaries(cfg,
|
||||||
|
module_info,
|
||||||
|
true_host_name,
|
||||||
|
rougail,
|
||||||
|
)
|
||||||
|
modules_info[module_info.module_name] = module_info.depends
|
||||||
|
self.modules[host_name] = modules_info
|
||||||
|
cfg['functions_file'] = list(functions_files)
|
||||||
|
self.tiram_obj = rougail.save(TIRAMISU_CACHE)
|
||||||
|
with open(TIRAMISU_CACHE, 'a') as cache:
|
||||||
|
cache.write(f"""#from pickle import load
|
||||||
|
#config = Config(option_0)
|
||||||
|
#config.property.read_only()
|
||||||
|
#with open('{VALUES_CACHE}', 'rb') as fh:
|
||||||
|
# config.value.importation(load(fh))
|
||||||
|
#with open('{INFORMATIONS_CACHE}', 'rb') as fh:
|
||||||
|
# config.information.importation(load(fh))
|
||||||
|
#print(config.value.mandatory())
|
||||||
|
""")
|
||||||
|
|
||||||
|
def add_tls(self):
|
||||||
|
dns_module_name = None
|
||||||
|
for host in self.servers_json['hosts'].values():
|
||||||
|
zones = [self.servers_json['zones']['external_zone'], None]
|
||||||
|
for server_name, datas in host['servers'].items():
|
||||||
|
if not 'applicationservice' in datas:
|
||||||
|
raise Exception(f'cannot find applicationservice for "{server_name}"')
|
||||||
|
if datas['applicationservice'] == 'tls':
|
||||||
|
raise Exception(f'forbidden module name "tls" for server "{server_name}"')
|
||||||
|
#FIXME use provider!
|
||||||
|
if datas['applicationservice'] == 'nginx-reverse-proxy' and len(datas['zones_name']) > 0:
|
||||||
|
if dns_module_name:
|
||||||
|
break
|
||||||
|
zones[1] = datas['provider_zone']
|
||||||
|
if zones[0] == zones[1] or not zones[1]:
|
||||||
|
zones.pop(1)
|
||||||
|
host['servers']['tls'] = {'applicationservice': 'tls',
|
||||||
|
'zones_name': zones,
|
||||||
|
}
|
||||||
|
|
||||||
|
def load_dictionaries(self, cfg, module_info, server_name, rougail):
|
||||||
|
if not module_info.dictionaries_dir:
|
||||||
|
raise Exception(f'server "{server_name}" has any dictionaries!')
|
||||||
|
cfg['dictionaries_dir'] = module_info.dictionaries_dir
|
||||||
|
cfg['extra_dictionaries'] = module_info.extra_dictionaries
|
||||||
|
cfg['functions_file'] = module_info.functions_file
|
||||||
|
rougail.load_dictionaries(path_prefix=server_name)
|
||||||
|
self.templates_dir[server_name] = module_info.templates_dir
|
||||||
|
self.patches_dir[server_name] = module_info.patches_dir
|
||||||
|
self.functions_files[server_name] = module_info.functions_file
|
||||||
|
self.manuals_dirs[server_name] = module_info.manuals
|
||||||
|
self.tests_dirs[server_name] = module_info.tests
|
||||||
|
|
||||||
|
def tiramisu_file_to_tiramisu(self):
|
||||||
|
# l
|
||||||
|
tiramisu_space = {}
|
||||||
|
try:
|
||||||
|
exec(self.tiram_obj, None, tiramisu_space)
|
||||||
|
except Exception as err:
|
||||||
|
raise Exception(f'unknown error when load tiramisu object: "{err}" see the file "{TIRAMISU_CACHE}" for more details') from err
|
||||||
|
if self.original_display_name:
|
||||||
|
display_name = None
|
||||||
|
else:
|
||||||
|
display_name = tiramisu_display_name
|
||||||
|
self.config = Config(tiramisu_space['option_0'],
|
||||||
|
display_name=display_name,
|
||||||
|
)
|
||||||
|
|
||||||
|
def load_values_and_informations(self):
|
||||||
|
config = self.config
|
||||||
|
config.property.read_write()
|
||||||
|
config.property.remove('validator')
|
||||||
|
config.property.remove('cache')
|
||||||
|
load_zones(self.zones, self.servers_json['hosts'])
|
||||||
|
config.information.set('zones', self.zones)
|
||||||
|
for host_name, hosts_datas in self.servers_json['hosts'].items():
|
||||||
|
information = config.option(normalize_family(host_name)).information
|
||||||
|
information.set('module', 'host')
|
||||||
|
information.set('templates_dir', self.templates_dir[host_name])
|
||||||
|
information.set('patches_dir', self.patches_dir[host_name])
|
||||||
|
information.set('functions_files', self.functions_files[host_name])
|
||||||
|
self.set_values(host_name, config, hosts_datas)
|
||||||
|
for datas in hosts_datas['servers'].values():
|
||||||
|
server_name = datas['server_name']
|
||||||
|
information = config.option(normalize_family(server_name)).information
|
||||||
|
information.set('module', datas['applicationservice'])
|
||||||
|
information.set('templates_dir', self.templates_dir[server_name])
|
||||||
|
information.set('patches_dir', self.patches_dir[server_name])
|
||||||
|
information.set('functions_files', self.functions_files[server_name])
|
||||||
|
information.set('manuals_dirs', self.manuals_dirs[server_name])
|
||||||
|
information.set('tests_dirs', self.tests_dirs[server_name])
|
||||||
|
self.set_values(server_name, config, datas)
|
||||||
|
|
||||||
|
config.information.set('copy_tests', False)
|
||||||
|
# FIXME only one host_name is supported
|
||||||
|
config.information.set('modules', self.modules[host_name])
|
||||||
|
# config.information.set('modules', {module_name: module_info.depends for module_name, module_info in self.module_infos.items() if module_name in modules})
|
||||||
|
with open(VALUES_CACHE, 'wb') as fh:
|
||||||
|
pickle_dump(config.value.exportation(), fh)
|
||||||
|
with open(INFORMATIONS_CACHE, 'wb') as fh:
|
||||||
|
pickle_dump(config.information.exportation(), fh)
|
||||||
|
config.property.add('cache')
|
||||||
|
if self.valid_mandatories:
|
||||||
|
messages = valid_mandatories(config)
|
||||||
|
if messages:
|
||||||
|
msg = ''
|
||||||
|
for title, variables in messages.items():
|
||||||
|
msg += '\n' + title + '\n'
|
||||||
|
msg += '\n'.join(variables)
|
||||||
|
raise Exception(msg)
|
||||||
|
config.property.read_only()
|
||||||
|
|
||||||
|
def set_values(self,
|
||||||
|
server_name,
|
||||||
|
config,
|
||||||
|
datas,
|
||||||
|
):
|
||||||
|
if 'values' not in datas:
|
||||||
|
return
|
||||||
|
if not isinstance(datas['values'], dict):
|
||||||
|
raise Exception(f'Values of "{server_name}" are not a dict: {datas["values"]}')
|
||||||
|
server_path = normalize_family(server_name)
|
||||||
|
config.owner.set(self.config_file)
|
||||||
|
for vpath, value in datas['values'].items():
|
||||||
|
path = f'{server_path}.{vpath}'
|
||||||
|
try:
|
||||||
|
if isinstance(value, dict):
|
||||||
|
for idx, val in value.items():
|
||||||
|
config.option(path, int(idx)).value.set(val)
|
||||||
|
else:
|
||||||
|
config.option(path).value.set(value)
|
||||||
|
except Exception as err:
|
||||||
|
value_pprint(config.value.dict(), config)
|
||||||
|
error_msg = f'cannot configure variable {vpath} for server "{server_name}": {err}'
|
||||||
|
raise Exception(error_msg) from err
|
||||||
|
config.owner.set('user')
|
||||||
|
|
||||||
|
|
||||||
|
class LoaderCache(Loader):
|
||||||
|
def load_tiramisu_file(self):
|
||||||
|
with open(TIRAMISU_CACHE) as fh:
|
||||||
|
self.tiram_obj = fh.read()
|
||||||
|
|
||||||
|
def load_values_and_informations(self):
|
||||||
|
with open(VALUES_CACHE, 'rb') as fh:
|
||||||
|
self.config.value.importation(pickle_load(fh))
|
||||||
|
with open(INFORMATIONS_CACHE, 'rb') as fh:
|
||||||
|
self.config.information.importation(pickle_load(fh))
|
||||||
|
|
||||||
|
|
||||||
|
def load(hide_secret=False,
|
||||||
|
original_display_name: bool=False,
|
||||||
|
valid_mandatories: bool=True,
|
||||||
|
copy_tests: bool=False,
|
||||||
|
):
|
||||||
|
if isfile(TIRAMISU_CACHE) and isfile(VALUES_CACHE) and isfile(INFORMATIONS_CACHE):
|
||||||
|
loader_obj = LoaderCache
|
||||||
|
else:
|
||||||
|
loader_obj = Loader
|
||||||
|
loader = loader_obj(hide_secret,
|
||||||
|
original_display_name,
|
||||||
|
valid_mandatories,
|
||||||
|
)
|
||||||
|
loader.load_tiramisu_file()
|
||||||
|
loader.tiramisu_file_to_tiramisu()
|
||||||
|
loader.load_values_and_informations()
|
||||||
|
config = loader.config
|
||||||
|
config.property.read_only()
|
||||||
|
config.information.set('copy_tests', copy_tests)
|
||||||
|
config.cache.reset()
|
||||||
|
return config
|
||||||
|
|
||||||
|
|
||||||
|
def build_files(hostname: str,
|
||||||
|
only_machine: str,
|
||||||
|
just_copy: bool,
|
||||||
|
copy_tests: bool,
|
||||||
|
template: str=None,
|
||||||
|
) -> None:
|
||||||
|
if isdir(INSTALL_DIR):
|
||||||
|
rmtree(INSTALL_DIR)
|
||||||
|
makedirs(INSTALL_DIR)
|
||||||
|
with open(CONFIG_FILE, 'r') as server_fh:
|
||||||
|
servers_json = yaml_load(server_fh, Loader=SafeLoader)
|
||||||
|
config = load(copy_tests=copy_tests)
|
||||||
|
machines = [subconfig.description() for subconfig in config.option.list(type='optiondescription')]
|
||||||
|
certificates = {'certificates': {},
|
||||||
|
'configuration': servers_json['certificates'],
|
||||||
|
}
|
||||||
|
# get certificates informations
|
||||||
|
tls_machine = config.option(f'{normalize_family(hostname)}.general.tls_server').value.get()
|
||||||
|
for machine in machines:
|
||||||
|
if machine == tls_machine:
|
||||||
|
continue
|
||||||
|
if hostname is None:
|
||||||
|
# FIXME multi host!
|
||||||
|
hostname = config.option(normalize_family(machine)).option('general.host_name').value.get()
|
||||||
|
if just_copy:
|
||||||
|
continue
|
||||||
|
is_host = machine == hostname
|
||||||
|
if is_host:
|
||||||
|
continue
|
||||||
|
machine_config = config.option(normalize_family(machine))
|
||||||
|
certificate_names = []
|
||||||
|
private_names = []
|
||||||
|
for service in machine_config.option('services').list('optiondescription'):
|
||||||
|
if not service.option('activate').value.get():
|
||||||
|
continue
|
||||||
|
# if service.option('manage').value.get():
|
||||||
|
# certificate_type = 'server'
|
||||||
|
# else:
|
||||||
|
# certificate_type = 'client'
|
||||||
|
tls_ca_directory = machine_config.option('general.tls_ca_directory').value.get()
|
||||||
|
tls_cert_directory = machine_config.option('general.tls_cert_directory').value.get()
|
||||||
|
tls_key_directory = machine_config.option('general.tls_key_directory').value.get()
|
||||||
|
try:
|
||||||
|
for certificate in service.option('certificates').list('all'):
|
||||||
|
if not certificate.option('activate').value.get():
|
||||||
|
continue
|
||||||
|
certificate_data = {key.rsplit('.', 1)[1]: value for key, value in certificate.value.dict().items()}
|
||||||
|
certificate_data['type'] = certificate.information.get('type')
|
||||||
|
certificate_data['authority'] = join(tls_ca_directory, certificate.information.get('authority') + '.crt')
|
||||||
|
certificate_data['format'] = certificate.information.get('format')
|
||||||
|
is_list_name = isinstance(certificate_data['name'], list)
|
||||||
|
is_list_domain = isinstance(certificate_data['domain'], list)
|
||||||
|
if is_list_name != is_list_domain:
|
||||||
|
raise Exception('certificate name and domain name must be a list together')
|
||||||
|
if 'provider' not in certificate_data:
|
||||||
|
certificate_data['provider'] = 'autosigne'
|
||||||
|
if is_list_name:
|
||||||
|
if len(certificate_data['name']) != len(certificate_data['domain']):
|
||||||
|
raise Exception('certificate name and domain name must have same length')
|
||||||
|
for idx, certificate_name in enumerate(certificate_data['name']):
|
||||||
|
cert_data = certificate_data.copy()
|
||||||
|
if certificate_data['format'] == 'cert_key':
|
||||||
|
cert_data['name'] = join(tls_cert_directory, certificate_name + '.crt')
|
||||||
|
private = join(tls_key_directory, certificate_name + '.key')
|
||||||
|
if private in private_names:
|
||||||
|
raise Exception(f'duplicate private key {private} for {machine}')
|
||||||
|
cert_data['private'] = private
|
||||||
|
private_names.append(private)
|
||||||
|
else:
|
||||||
|
cert_data['name'] = join(tls_key_directory, certificate_name + '.pem')
|
||||||
|
cert_data['domain'] = certificate_data['domain'][idx]
|
||||||
|
if cert_data['name'] in certificate_names:
|
||||||
|
raise Exception(f'duplicate certificate {cert_data["name"]} for {machine}')
|
||||||
|
certificates['certificates'].setdefault(machine, []).append(cert_data)
|
||||||
|
certificate_names.append(cert_data['name'])
|
||||||
|
else:
|
||||||
|
name = certificate_data['name']
|
||||||
|
if certificate_data['format'] == 'cert_key':
|
||||||
|
certificate_data['name'] = join(tls_cert_directory, name + '.crt')
|
||||||
|
private = join(tls_key_directory, name + '.key')
|
||||||
|
if private in private_names:
|
||||||
|
raise Exception(f'duplicate private key {private} for {machine}')
|
||||||
|
certificate_data['private'] = private
|
||||||
|
else:
|
||||||
|
certificate_data['name'] = join(tls_key_directory, name + '.pem')
|
||||||
|
if certificate_data['name'] in certificate_names:
|
||||||
|
raise Exception(f'duplicate certificate {certificate_data["name"]} for {machine}')
|
||||||
|
certificate_names.append(certificate_data['name'])
|
||||||
|
certificates['certificates'].setdefault(machine, []).append(certificate_data)
|
||||||
|
except AttributeError:
|
||||||
|
pass
|
||||||
|
directories = {}
|
||||||
|
for machine in machines:
|
||||||
|
if just_copy and hostname == machine:
|
||||||
|
continue
|
||||||
|
if only_machine and only_machine != machine:
|
||||||
|
continue
|
||||||
|
templates(machine,
|
||||||
|
config,
|
||||||
|
just_copy=just_copy,
|
||||||
|
copy_manuals=True,
|
||||||
|
template=template,
|
||||||
|
extra_variables=certificates,
|
||||||
|
)
|
||||||
|
is_host = machine == hostname
|
||||||
|
if is_host:
|
||||||
|
directories[machine] = '/usr/local/lib'
|
||||||
|
elif not just_copy:
|
||||||
|
machine_config = config.option(normalize_family(machine))
|
||||||
|
directories[machine] = machine_config.option('general.config_dir').value.get()
|
||||||
|
if only_machine:
|
||||||
|
return directories
|
||||||
|
if only_machine:
|
||||||
|
raise Exception(f'cannot find machine {only_machine}: {machines}')
|
||||||
|
return directories, certificates
|
|
@ -1,5 +1,9 @@
|
||||||
from rougail.annotator.variable import Walk
|
from rougail.annotator.variable import Walk
|
||||||
|
from rougail.error import DictConsistencyError
|
||||||
|
from rougail.utils import normalize_family
|
||||||
from risotto.utils import _
|
from risotto.utils import _
|
||||||
|
from warnings import warn
|
||||||
|
from typing import List, Tuple
|
||||||
|
|
||||||
|
|
||||||
class Annotator(Walk):
|
class Annotator(Walk):
|
||||||
|
@ -7,51 +11,394 @@ class Annotator(Walk):
|
||||||
def __init__(self,
|
def __init__(self,
|
||||||
objectspace: 'RougailObjSpace',
|
objectspace: 'RougailObjSpace',
|
||||||
*args):
|
*args):
|
||||||
|
self.providers = {}
|
||||||
|
self.suppliers = {}
|
||||||
|
self.globals = {}
|
||||||
|
self.provider_links = {}
|
||||||
|
self.providers_zone = {}
|
||||||
|
self.provider_maps = {}
|
||||||
self.objectspace = objectspace
|
self.objectspace = objectspace
|
||||||
# self.convert_get_linked_information()
|
#
|
||||||
# self.convert_provider()
|
self.get_providers_suppliers()
|
||||||
|
self.get_provider_links()
|
||||||
|
self.get_provider_maps()
|
||||||
|
self.convert_providers()
|
||||||
|
self.convert_globals()
|
||||||
|
|
||||||
def convert_get_linked_information(self):
|
def get_providers_suppliers(self) -> None:
|
||||||
if not hasattr(self.objectspace.space, 'constraints') or \
|
"""parse all variable and get provider and supplier informations
|
||||||
not hasattr(self.objectspace.space.constraints, 'fill'):
|
"""
|
||||||
return
|
|
||||||
for fill in self.objectspace.space.constraints.fill:
|
|
||||||
if fill.name == 'get_linked_configuration':
|
|
||||||
# add server_name
|
|
||||||
param = self.objectspace.param(fill.xmlfiles)
|
|
||||||
param.name = 'server_name'
|
|
||||||
param.type = 'information'
|
|
||||||
param.text = 'server_name'
|
|
||||||
fill.param.append(param)
|
|
||||||
# add current_user
|
|
||||||
param = self.objectspace.param(fill.xmlfiles)
|
|
||||||
param.name = 'current_user'
|
|
||||||
param.type = 'information'
|
|
||||||
param.text = 'current_user'
|
|
||||||
fill.param.append(param)
|
|
||||||
# add test
|
|
||||||
param = self.objectspace.param(fill.xmlfiles)
|
|
||||||
param.name = 'test'
|
|
||||||
param.type = 'target_information'
|
|
||||||
param.text = 'test'
|
|
||||||
fill.param.append(param)
|
|
||||||
|
|
||||||
def convert_provider(self):
|
|
||||||
if not hasattr(self.objectspace.space, 'variables'):
|
|
||||||
return
|
|
||||||
for family in self.get_families():
|
|
||||||
if not hasattr(family, 'provider'):
|
|
||||||
continue
|
|
||||||
if 'dynamic' not in vars(family):
|
|
||||||
raise Exception(_(f'{family.name} is not a dynamic family so cannot have provider attribute'))
|
|
||||||
if not hasattr(family, 'information'):
|
|
||||||
family.information = self.objectspace.information(family.xmlfiles)
|
|
||||||
family.information.provider = family.provider
|
|
||||||
del family.provider
|
|
||||||
for variable in self.get_variables():
|
for variable in self.get_variables():
|
||||||
if not hasattr(variable, 'provider'):
|
for type_ in 'provider', 'supplier':
|
||||||
continue
|
if hasattr(variable, type_):
|
||||||
|
provider_name = getattr(variable, type_)
|
||||||
|
# add information with provider/supplier name
|
||||||
if not hasattr(variable, 'information'):
|
if not hasattr(variable, 'information'):
|
||||||
variable.information = self.objectspace.information(variable.xmlfiles)
|
variable.information = self.objectspace.information(variable.xmlfiles)
|
||||||
variable.information.provider = variable.provider
|
setattr(variable.information, type_, provider_name)
|
||||||
del variable.provider
|
|
||||||
|
# construct self.globals, self.suppliers and self.providers dictionnaries
|
||||||
|
provider_prefix, provider_suffix = self._cut_out_provider_name(provider_name)
|
||||||
|
dns = self.objectspace.space.variables[variable.path_prefix].doc
|
||||||
|
if provider_prefix == 'global':
|
||||||
|
if type_ == 'supplier':
|
||||||
|
raise DictConsistencyError(f'{type_} {provider_name} in {dns} not allowed', 0, variable.xmlfiles)
|
||||||
|
obj = self.globals
|
||||||
|
elif type_ == 'supplier':
|
||||||
|
obj = self.suppliers
|
||||||
|
else:
|
||||||
|
obj = self.providers
|
||||||
|
sub_obj = obj.setdefault(provider_prefix, {}).setdefault(dns, {})
|
||||||
|
if provider_suffix in sub_obj:
|
||||||
|
raise DictConsistencyError(f'multiple {type_} {provider_name} in {dns}', 0, sub_obj[provider_suffix].xmlfiles + variable.xmlfiles)
|
||||||
|
sub_obj[provider_suffix] = variable
|
||||||
|
|
||||||
|
def _cut_out_provider_name(self,
|
||||||
|
provider_name: str,
|
||||||
|
) -> Tuple[str, str]:
|
||||||
|
"""get provider_name and return provider_prefix and provider_suffix
|
||||||
|
"""
|
||||||
|
if ':' in provider_name:
|
||||||
|
provider_prefix, provider_suffix = provider_name.split(':', 1)
|
||||||
|
else:
|
||||||
|
provider_prefix = provider_name
|
||||||
|
provider_suffix = None
|
||||||
|
return provider_prefix, provider_suffix
|
||||||
|
|
||||||
|
def get_provider_links(self):
|
||||||
|
"""Search link between providers
|
||||||
|
'ProviderPrefix': {'provider_dns': ['supplier_dns_1',
|
||||||
|
'supplier_dns_2',
|
||||||
|
'supplier_dns_3']}
|
||||||
|
"""
|
||||||
|
for provider_prefix, providers_dns in self.providers.items():
|
||||||
|
self.provider_links[provider_prefix] = {}
|
||||||
|
for provider_dns, providers_suffix in providers_dns.items():
|
||||||
|
if None not in providers_suffix:
|
||||||
|
# it's a reverse provider!
|
||||||
|
continue
|
||||||
|
if provider_prefix != 'Host':
|
||||||
|
provider_zone = self.objectspace.rougailconfig['risotto_globals'][provider_dns]['global:provider_zone']
|
||||||
|
if provider_prefix not in self.suppliers:
|
||||||
|
continue
|
||||||
|
for supplier_dns, suppliers_suffix in self.suppliers[provider_prefix].items():
|
||||||
|
if provider_dns == supplier_dns:
|
||||||
|
continue
|
||||||
|
if provider_prefix == 'Host':
|
||||||
|
provider_zone = self.objectspace.rougailconfig['risotto_globals'][supplier_dns]['global:zones_name'][0]
|
||||||
|
if provider_zone not in self.objectspace.rougailconfig['risotto_globals'][supplier_dns]['global:zones_name']:
|
||||||
|
continue
|
||||||
|
self.provider_links[provider_prefix].setdefault(provider_dns, []).append(supplier_dns)
|
||||||
|
|
||||||
|
def get_provider_maps(self):
|
||||||
|
"""relation between provider_prefix and provider_suffix
|
||||||
|
'provider_prefix': {'normal': {None, 'provider_suffix_1', 'provider_suffix_2'},
|
||||||
|
'reverse': {'provider_suffix_3', 'provider_suffix_4'}}
|
||||||
|
"""
|
||||||
|
for provider_prefix, providers_dns in self.provider_links.items():
|
||||||
|
self.provider_maps[provider_prefix] = {'normal': set(), 'reverse': set()}
|
||||||
|
for provider_dns, suppliers_dns in providers_dns.items():
|
||||||
|
for supplier_dns in suppliers_dns:
|
||||||
|
if supplier_dns in self.providers[provider_prefix] and None in self.providers[provider_prefix][supplier_dns]:
|
||||||
|
#exists?
|
||||||
|
continue
|
||||||
|
# get prefixes
|
||||||
|
prefixes = set(self.providers[provider_prefix][provider_dns]) & set(self.suppliers[provider_prefix][supplier_dns])
|
||||||
|
self.provider_maps[provider_prefix]['normal'] |= prefixes
|
||||||
|
# get suffixes
|
||||||
|
if supplier_dns not in self.providers[provider_prefix]:
|
||||||
|
continue
|
||||||
|
suffixes = set(self.providers[provider_prefix][supplier_dns]) & set(self.suppliers[provider_prefix][provider_dns])
|
||||||
|
self.provider_maps[provider_prefix]['reverse'] |= suffixes
|
||||||
|
|
||||||
|
def convert_providers(self) -> None:
|
||||||
|
"""Convert providers informations to default values or fills
|
||||||
|
"""
|
||||||
|
for provider_prefix, providers_dns in self.provider_links.items():
|
||||||
|
for provider_dns, suppliers_dns in providers_dns.items():
|
||||||
|
for provider_suffix in self.provider_maps[provider_prefix]['normal']:
|
||||||
|
self._convert_providers_normal(provider_prefix,
|
||||||
|
provider_suffix,
|
||||||
|
provider_dns,
|
||||||
|
suppliers_dns,
|
||||||
|
)
|
||||||
|
for provider_suffix in self.provider_maps[provider_prefix]['reverse']:
|
||||||
|
self._convert_providers_reverse(provider_prefix,
|
||||||
|
provider_suffix,
|
||||||
|
provider_dns,
|
||||||
|
suppliers_dns,
|
||||||
|
)
|
||||||
|
|
||||||
|
def _convert_providers_normal(self,
|
||||||
|
provider_prefix: str,
|
||||||
|
provider_suffix: str,
|
||||||
|
provider_dns: str,
|
||||||
|
suppliers_dns: dict,
|
||||||
|
) -> None:
|
||||||
|
if provider_prefix != 'Host':
|
||||||
|
provider_zone = self.objectspace.rougailconfig['risotto_globals'][provider_dns]['global:provider_zone']
|
||||||
|
provider_option_dns = self._get_dns_from_provider_zone(provider_dns, provider_zone)
|
||||||
|
variable = self.providers[provider_prefix][provider_dns][provider_suffix]
|
||||||
|
if hasattr(variable, 'value'):
|
||||||
|
raise DictConsistencyError(f'variable {variable.path} has a provider and a value', 0, variable.xmlfiles)
|
||||||
|
suppliers_var = {}
|
||||||
|
for supplier_dns in suppliers_dns:
|
||||||
|
if provider_suffix not in self.suppliers[provider_prefix][supplier_dns]:
|
||||||
|
continue
|
||||||
|
if provider_prefix == 'Host':
|
||||||
|
provider_zone = self.objectspace.rougailconfig['risotto_globals'][supplier_dns]['global:zones_name'][0]
|
||||||
|
provider_option_dns = self._get_dns_from_provider_zone(provider_dns, provider_zone)
|
||||||
|
supplier_variable = self.suppliers[provider_prefix][supplier_dns][provider_suffix]
|
||||||
|
supplier_option_dns = self._get_dns_from_provider_zone(supplier_dns, provider_zone)
|
||||||
|
suppliers_var[supplier_option_dns] = supplier_variable
|
||||||
|
if provider_suffix:
|
||||||
|
AddFill(self.objectspace,
|
||||||
|
provider_prefix,
|
||||||
|
provider_suffix,
|
||||||
|
provider_option_dns,
|
||||||
|
variable,
|
||||||
|
suppliers_var,
|
||||||
|
False,
|
||||||
|
supplier_variable.path_prefix,
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
self._set_provider_supplier(provider_option_dns,
|
||||||
|
variable,
|
||||||
|
suppliers_var,
|
||||||
|
)
|
||||||
|
|
||||||
|
def _convert_providers_reverse(self,
|
||||||
|
provider_prefix: str,
|
||||||
|
provider_suffix: str,
|
||||||
|
provider_dns: str,
|
||||||
|
suppliers_dns: dict,
|
||||||
|
) -> None:
|
||||||
|
if provider_prefix != 'Host':
|
||||||
|
provider_zone = self.objectspace.rougailconfig['risotto_globals'][provider_dns]['global:provider_zone']
|
||||||
|
provider_option_dns = self._get_dns_from_provider_zone(provider_dns, provider_zone)
|
||||||
|
variable = self.suppliers[provider_prefix][provider_dns][provider_suffix]
|
||||||
|
if hasattr(variable, 'value'):
|
||||||
|
raise DictConsistencyError(f'variable {variable.path} has a provider and a value', 0, variable.xmlfiles)
|
||||||
|
for supplier_dns in suppliers_dns:
|
||||||
|
if provider_prefix == 'Host':
|
||||||
|
provider_zone = self.objectspace.rougailconfig['risotto_globals'][supplier_dns]['global:zones_name'][0]
|
||||||
|
provider_option_dns = self._get_dns_from_provider_zone(provider_dns, provider_zone)
|
||||||
|
supplier_variable = self.providers[provider_prefix][supplier_dns][provider_suffix]
|
||||||
|
supplier_option_dns = self._get_dns_from_provider_zone(supplier_dns, provider_zone)
|
||||||
|
AddFill(self.objectspace,
|
||||||
|
provider_prefix,
|
||||||
|
provider_suffix,
|
||||||
|
supplier_option_dns,
|
||||||
|
supplier_variable,
|
||||||
|
{provider_option_dns: variable},
|
||||||
|
True,
|
||||||
|
supplier_variable.path_prefix,
|
||||||
|
)
|
||||||
|
|
||||||
|
def _get_dns_from_provider_zone(self,
|
||||||
|
dns,
|
||||||
|
zone,
|
||||||
|
) -> str:
|
||||||
|
risotto_global = self.objectspace.rougailconfig['risotto_globals'][dns]
|
||||||
|
index = risotto_global['global:zones_name'].index(zone)
|
||||||
|
return risotto_global['global:server_names'][index]
|
||||||
|
|
||||||
|
def _set_provider_supplier(self,
|
||||||
|
provider_dns: str,
|
||||||
|
variable,
|
||||||
|
suppliers,
|
||||||
|
) -> None:
|
||||||
|
# suffix is None so not a client and there only one provider
|
||||||
|
# the value of this variable is the list of suppliers DNS name
|
||||||
|
if not variable.multi:
|
||||||
|
raise DictConsistencyError(f'"{variable.name}" is a provider and must be a multi', 0, variable.xmlfiles)
|
||||||
|
variable.default = list(suppliers)
|
||||||
|
# suffix is None so the supplier values are provider DNS
|
||||||
|
for sub_variable in suppliers.values():
|
||||||
|
#FIXME
|
||||||
|
#if hasattr(option, 'value'):
|
||||||
|
# raise DictConsistencyError(f'"{option.name}" is a supplier and cannot have value', 0, option.xmlfiles)
|
||||||
|
if sub_variable.multi:
|
||||||
|
raise DictConsistencyError(f'"{sub_variable.name}" is a supplier and mustnot be a multi', 0, sub_variable.xmlfiles)
|
||||||
|
sub_variable.default = provider_dns
|
||||||
|
|
||||||
|
def convert_globals(self):
|
||||||
|
"""Convert providers global informations to default values or fills
|
||||||
|
"""
|
||||||
|
provider_prefix = 'global'
|
||||||
|
for provider_dns, providers_suffix in self.globals[provider_prefix].items():
|
||||||
|
for provider_suffix, variable in providers_suffix.items():
|
||||||
|
provider_name = f'{provider_prefix}:{provider_suffix}'
|
||||||
|
if provider_name not in self.objectspace.rougailconfig['risotto_globals'][provider_dns]:
|
||||||
|
raise DictConsistencyError(f'cannot find {provider_name} for variable {variable.path}, should be in {list(self.objectspace.rougailconfig["risotto_globals"][provider_dns])}', 0, variable.xmlfiles)
|
||||||
|
provider_values = self.objectspace.rougailconfig['risotto_globals'][provider_dns][provider_name]
|
||||||
|
if isinstance(provider_values, list) and self.objectspace.paths.is_dynamic(variable):
|
||||||
|
if variable.multi:
|
||||||
|
raise DictConsistencyError(f'variable {variable.path} has provider {provider_name} and is in dynamic family so must not be a multi', 0, variable.xmlfiles)
|
||||||
|
self._set_global_dynamic_option(variable, provider_values)
|
||||||
|
else:
|
||||||
|
if isinstance(provider_values, list) and not variable.multi:
|
||||||
|
raise DictConsistencyError(f'variable {variable.path} has provider {provider_name} which is a multi', 0, variable.xmlfiles)
|
||||||
|
if not isinstance(provider_values, list):
|
||||||
|
if variable.multi:
|
||||||
|
raise DictConsistencyError(f'variable {variable.path} has provider {provider_name} which is not a multi', 0, variable.xmlfiles)
|
||||||
|
provider_values = [provider_values]
|
||||||
|
|
||||||
|
variable.value = []
|
||||||
|
for provider_value in provider_values:
|
||||||
|
value = self.objectspace.value(variable.xmlfiles)
|
||||||
|
value.name = provider_value
|
||||||
|
if isinstance(provider_value, bool):
|
||||||
|
value.type = 'boolean'
|
||||||
|
variable.value.append(value)
|
||||||
|
|
||||||
|
def _set_global_dynamic_option(self,
|
||||||
|
variable: 'self.objectspace.Variable',
|
||||||
|
values: List[str],
|
||||||
|
):
|
||||||
|
fill = self.objectspace.fill(variable.xmlfiles)
|
||||||
|
new_target = self.objectspace.target(variable.xmlfiles)
|
||||||
|
new_target.name = variable.name
|
||||||
|
fill.target = [new_target]
|
||||||
|
fill.namespace = variable.namespace
|
||||||
|
fill.index = 0
|
||||||
|
fill.name = 'risotto_providers_global'
|
||||||
|
param1 = self.objectspace.param(variable.xmlfiles)
|
||||||
|
param1.text = values
|
||||||
|
param2 = self.objectspace.param(variable.xmlfiles)
|
||||||
|
param2.type = 'suffix'
|
||||||
|
fill.param = [param1, param2]
|
||||||
|
if not hasattr(self.objectspace.space.variables[variable.path_prefix].constraints, 'fill'):
|
||||||
|
self.objectspace.space.variables[variable.path_prefix].constraints.fill = []
|
||||||
|
self.objectspace.space.variables[variable.path_prefix].constraints.fill.append(fill)
|
||||||
|
|
||||||
|
|
||||||
|
class AddFill:
|
||||||
|
"""Add fill for variable
|
||||||
|
"""
|
||||||
|
def __init__(self,
|
||||||
|
objectspace,
|
||||||
|
provider_prefix,
|
||||||
|
provider_suffix,
|
||||||
|
provider_dns,
|
||||||
|
variable,
|
||||||
|
suppliers,
|
||||||
|
reverse,
|
||||||
|
path_prefix,
|
||||||
|
) -> None:
|
||||||
|
self.objectspace = objectspace
|
||||||
|
self.provider_dns = provider_dns
|
||||||
|
self.variable = variable
|
||||||
|
self.path_prefix = path_prefix
|
||||||
|
self.suppliers = suppliers
|
||||||
|
self.create_fill()
|
||||||
|
if reverse:
|
||||||
|
self.param_reverse()
|
||||||
|
else:
|
||||||
|
if self.objectspace.paths.is_dynamic(self.variable):
|
||||||
|
self.param_dynamic()
|
||||||
|
elif self.variable.multi:
|
||||||
|
self.param_multi()
|
||||||
|
else:
|
||||||
|
provider_name = f'{provider_prefix}:{provider_suffix}'
|
||||||
|
raise DictConsistencyError(f'provider "{provider_name}" options must be in dynamic option or must be a multiple', 0, self.variable.xmlfiles)
|
||||||
|
if self.objectspace.paths.is_follower(self.variable):
|
||||||
|
self.param_follower()
|
||||||
|
self.end()
|
||||||
|
|
||||||
|
def create_fill(self) -> None:
|
||||||
|
self.fill = self.objectspace.fill(self.variable.xmlfiles)
|
||||||
|
new_target = self.objectspace.target(self.variable.xmlfiles)
|
||||||
|
new_target.name = self.variable
|
||||||
|
self.fill.target = [new_target]
|
||||||
|
self.fill.namespace = self.variable.namespace
|
||||||
|
self.fill.index = 0
|
||||||
|
self.fill.param = []
|
||||||
|
|
||||||
|
def param_reverse(self) -> None:
|
||||||
|
self.fill.name = 'risotto_flatten_values_client'
|
||||||
|
#
|
||||||
|
if self.objectspace.paths.is_follower(self.variable):
|
||||||
|
multi = self.variable.multi is True
|
||||||
|
else:
|
||||||
|
multi = self.variable.multi is not False
|
||||||
|
param = self.objectspace.param(self.variable.xmlfiles)
|
||||||
|
param.text = multi
|
||||||
|
param.type = 'boolean'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
for dns, variable in self.suppliers.items():
|
||||||
|
param = self.objectspace.param(variable.xmlfiles)
|
||||||
|
param.text = variable
|
||||||
|
param.propertyerror = False
|
||||||
|
param.type = 'variable'
|
||||||
|
param.suffix = normalize_family(self.provider_dns)
|
||||||
|
namespace = variable.namespace
|
||||||
|
family_path = self.objectspace.paths.get_variable_family_path(param.text.path,
|
||||||
|
namespace,
|
||||||
|
force_path_prefix=self.variable.path_prefix,
|
||||||
|
)
|
||||||
|
param.family = self.objectspace.paths.get_family(family_path,
|
||||||
|
namespace,
|
||||||
|
self.variable.path_prefix,
|
||||||
|
)
|
||||||
|
self.fill.param.append(param)
|
||||||
|
|
||||||
|
def param_dynamic(self) -> None:
|
||||||
|
self.fill.name = 'risotto_dyn_values'
|
||||||
|
#
|
||||||
|
param = self.objectspace.param(self.variable.xmlfiles)
|
||||||
|
param.type = 'suffix'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
#
|
||||||
|
param = self.objectspace.param(self.variable.xmlfiles)
|
||||||
|
param.text = self.variable.unique != "False"
|
||||||
|
param.type = 'boolean'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
#
|
||||||
|
if self.objectspace.paths.is_follower(self.variable):
|
||||||
|
multi = self.variable.multi is True
|
||||||
|
else:
|
||||||
|
multi = self.variable.multi is not False
|
||||||
|
param = self.objectspace.param(self.variable.xmlfiles)
|
||||||
|
param.text = multi
|
||||||
|
param.type = 'boolean'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
for dns, variable in self.suppliers.items():
|
||||||
|
#
|
||||||
|
param = self.objectspace.param(variable.xmlfiles)
|
||||||
|
param.text = variable
|
||||||
|
param.name = normalize_family(dns)
|
||||||
|
param.propertyerror = False
|
||||||
|
param.type = 'variable'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
|
||||||
|
def param_multi(self) -> None:
|
||||||
|
self.fill.name = 'risotto_flatten_values'
|
||||||
|
#
|
||||||
|
if self.objectspace.paths.is_follower(self.variable):
|
||||||
|
multi = self.variable.multi is True
|
||||||
|
else:
|
||||||
|
multi = self.variable.multi is not False
|
||||||
|
param = self.objectspace.param(self.variable.xmlfiles)
|
||||||
|
param.text = multi
|
||||||
|
param.type = 'boolean'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
for dns, variable in self.suppliers.items():
|
||||||
|
param = self.objectspace.param(variable.xmlfiles)
|
||||||
|
param.text = variable
|
||||||
|
param.propertyerror = False
|
||||||
|
param.type = 'variable'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
|
||||||
|
def param_follower(self):
|
||||||
|
param = self.objectspace.param(self.variable.xmlfiles)
|
||||||
|
param.name = 'follower_index'
|
||||||
|
param.type = 'index'
|
||||||
|
self.fill.param.append(param)
|
||||||
|
|
||||||
|
def end(self):
|
||||||
|
if not hasattr(self.objectspace.space.variables[self.path_prefix], 'constraints'):
|
||||||
|
self.objectspace.space.variables[self.path_prefix].constraints = self.objectspace.constraints(None)
|
||||||
|
if not hasattr(self.objectspace.space.variables[self.path_prefix].constraints, 'fill'):
|
||||||
|
self.objectspace.space.variables[self.path_prefix].constraints.fill = []
|
||||||
|
self.objectspace.space.variables[self.path_prefix].constraints.fill.append(self.fill)
|
||||||
|
|
61
src/risotto/rougail/func.py
Normal file
|
@ -0,0 +1,61 @@
|
||||||
|
from risotto.utils import multi_function as _multi_function
|
||||||
|
from rougail.utils import normalize_family
|
||||||
|
from tiramisu import valid_network_netmask, valid_ip_netmask, valid_broadcast, valid_in_network, valid_not_equal, calc_value, calc_value_property_help
|
||||||
|
|
||||||
|
|
||||||
|
@_multi_function
|
||||||
|
def risotto_providers_global(value, suffix=None):
|
||||||
|
if suffix is not None:
|
||||||
|
return value[int(suffix)]
|
||||||
|
return value
|
||||||
|
|
||||||
|
|
||||||
|
@_multi_function
|
||||||
|
def risotto_flatten_values(multi, *args, follower_index=None):
|
||||||
|
values = []
|
||||||
|
#if follower_index is None:
|
||||||
|
# for arg in args:
|
||||||
|
# if isinstance(arg, list):
|
||||||
|
# values.extend(arg)
|
||||||
|
# else:
|
||||||
|
# values.append(arg)
|
||||||
|
#else:
|
||||||
|
values = args
|
||||||
|
if follower_index is not None and len(values) > follower_index:
|
||||||
|
values = values[follower_index]
|
||||||
|
elif not multi:
|
||||||
|
if not values:
|
||||||
|
values = None
|
||||||
|
if len(values) == 1:
|
||||||
|
values = values[0]
|
||||||
|
return values
|
||||||
|
|
||||||
|
|
||||||
|
@_multi_function
|
||||||
|
def risotto_flatten_values_client(multi, *args):
|
||||||
|
values = []
|
||||||
|
for arg in args:
|
||||||
|
if isinstance(arg, list):
|
||||||
|
values.extend(arg)
|
||||||
|
else:
|
||||||
|
values.append(arg)
|
||||||
|
if not multi:
|
||||||
|
if not values:
|
||||||
|
values = None
|
||||||
|
if len(values) == 1:
|
||||||
|
values = values[0]
|
||||||
|
return values
|
||||||
|
|
||||||
|
|
||||||
|
@_multi_function
|
||||||
|
def risotto_dyn_values(suffix, unique, multi, follower_index=None, **kwargs):
|
||||||
|
values = kwargs.get(normalize_family(suffix), [] if multi else None)
|
||||||
|
if not multi and follower_index is not None and isinstance(values, list) and len(values) > follower_index:
|
||||||
|
values = values[follower_index]
|
||||||
|
if isinstance(values, list) and unique:
|
||||||
|
values_ = []
|
||||||
|
for val in values:
|
||||||
|
if val not in values_:
|
||||||
|
values_.append(val)
|
||||||
|
values = values_
|
||||||
|
return values
|
|
@ -1,5 +1,31 @@
|
||||||
|
from os import environ, makedirs
|
||||||
|
from os.path import isfile, join, isdir
|
||||||
|
from typing import List
|
||||||
|
from ipaddress import ip_address
|
||||||
|
from toml import load as toml_load
|
||||||
|
from json import load, dump
|
||||||
|
from json.decoder import JSONDecodeError
|
||||||
|
from pprint import pprint
|
||||||
|
|
||||||
|
|
||||||
MULTI_FUNCTIONS = []
|
MULTI_FUNCTIONS = []
|
||||||
CONFIGS = {}
|
EXTRA_ANNOTATORS = ['risotto.rougail']
|
||||||
|
ROUGAIL_NAMESPACE = 'general'
|
||||||
|
ROUGAIL_NAMESPACE_DESCRIPTION = 'Général'
|
||||||
|
HERE = environ['PWD']
|
||||||
|
IP_DIR = join(HERE, 'ip')
|
||||||
|
|
||||||
|
|
||||||
|
# custom filters from dataset
|
||||||
|
custom_filters = {}
|
||||||
|
|
||||||
|
|
||||||
|
config_file = environ.get('CONFIG_FILE', 'risotto.conf')
|
||||||
|
if isfile(config_file):
|
||||||
|
with open(config_file, 'r') as fh:
|
||||||
|
RISOTTO_CONFIG = toml_load(fh)
|
||||||
|
else:
|
||||||
|
RISOTTO_CONFIG = {}
|
||||||
|
|
||||||
|
|
||||||
def _(s):
|
def _(s):
|
||||||
|
@ -12,3 +38,61 @@ def multi_function(function):
|
||||||
if name not in MULTI_FUNCTIONS:
|
if name not in MULTI_FUNCTIONS:
|
||||||
MULTI_FUNCTIONS.append(name)
|
MULTI_FUNCTIONS.append(name)
|
||||||
return function
|
return function
|
||||||
|
|
||||||
|
|
||||||
|
def value_pprint(dico, config):
|
||||||
|
pprint_dict = {}
|
||||||
|
for path, value in dico.items():
|
||||||
|
if config.option(path).type() == 'password' and value:
|
||||||
|
value = 'X' * len(value)
|
||||||
|
pprint_dict[path] = value
|
||||||
|
pprint(pprint_dict)
|
||||||
|
|
||||||
|
|
||||||
|
def load_zones(zones, hosts):
|
||||||
|
if not isdir(IP_DIR):
|
||||||
|
makedirs(IP_DIR)
|
||||||
|
json_file = join(IP_DIR, 'zones.json')
|
||||||
|
if isfile(json_file):
|
||||||
|
try:
|
||||||
|
with open(json_file, 'r') as fh:
|
||||||
|
ori_zones_ip = load(fh)
|
||||||
|
except JSONDecodeError:
|
||||||
|
ori_zones_ip = {}
|
||||||
|
else:
|
||||||
|
ori_zones_ip = {}
|
||||||
|
new_zones_ip = {}
|
||||||
|
# cache, machine should not change IP
|
||||||
|
for host_name, dhosts in hosts.items():
|
||||||
|
for server_name, server in dhosts['servers'].items():
|
||||||
|
server_zones = server['zones_name']
|
||||||
|
for idx, zone_name in enumerate(server_zones):
|
||||||
|
zone = zones[zone_name]
|
||||||
|
zone.setdefault('hosts', {})
|
||||||
|
if zone_name not in new_zones_ip:
|
||||||
|
new_zones_ip[zone_name] = {}
|
||||||
|
if zone_name in ori_zones_ip and server_name in ori_zones_ip[zone_name]:
|
||||||
|
server_index = ori_zones_ip[zone_name][server_name]
|
||||||
|
if server_index >= zone['length']:
|
||||||
|
server_index = None
|
||||||
|
elif server_index in new_zones_ip[zone_name].values():
|
||||||
|
server_index = None
|
||||||
|
else:
|
||||||
|
server_index = None
|
||||||
|
new_zones_ip[zone_name][server_name] = server_index
|
||||||
|
for zone_name, servers in new_zones_ip.items():
|
||||||
|
for server_name, server_idx in servers.items():
|
||||||
|
if server_idx is not None:
|
||||||
|
continue
|
||||||
|
for new_idx in range(zones[zone_name]['length']):
|
||||||
|
if new_idx not in new_zones_ip[zone_name].values():
|
||||||
|
new_zones_ip[zone_name][server_name] = new_idx
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
raise Exception(f'cannot find free IP in zone "{zone_name}" for "{server_name}"')
|
||||||
|
for zone_name, servers in new_zones_ip.items():
|
||||||
|
start_ip = ip_address(zones[zone_name]['start_ip'])
|
||||||
|
for server_name, server_index in servers.items():
|
||||||
|
zones[zone_name]['hosts'][server_name] = str(start_ip + server_index)
|
||||||
|
with open(json_file, 'w') as fh:
|
||||||
|
dump(new_zones_ip, fh)
|
||||||
|
|
|
@ -1,254 +0,0 @@
|
||||||
from OpenSSL.crypto import load_certificate, load_privatekey, dump_certificate, dump_privatekey, dump_publickey, PKey, X509, X509Extension, TYPE_RSA, FILETYPE_PEM
|
|
||||||
from os import makedirs, symlink
|
|
||||||
from os.path import join, isdir, isfile, exists
|
|
||||||
#from shutil import rmtree
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
|
|
||||||
PKI_DIR = 'pki/x509'
|
|
||||||
#FIXME
|
|
||||||
EMAIL = 'gnunux@gnunux.info'
|
|
||||||
COUNTRY = 'FR'
|
|
||||||
LOCALITY = 'Dijon'
|
|
||||||
STATE = 'France'
|
|
||||||
ORG_NAME = 'Cadoles'
|
|
||||||
ORG_UNIT_NAME = 'CSS'
|
|
||||||
|
|
||||||
|
|
||||||
def _gen_key_pair():
|
|
||||||
key = PKey()
|
|
||||||
key.generate_key(TYPE_RSA, 4096)
|
|
||||||
return key
|
|
||||||
|
|
||||||
|
|
||||||
def _gen_cert(is_ca,
|
|
||||||
common_names,
|
|
||||||
serial_number,
|
|
||||||
validity_end_in_seconds,
|
|
||||||
key_file,
|
|
||||||
cert_file,
|
|
||||||
type=None,
|
|
||||||
ca_cert=None,
|
|
||||||
ca_key=None,
|
|
||||||
email_address=None,
|
|
||||||
country_name=None,
|
|
||||||
locality_name=None,
|
|
||||||
state_or_province_name=None,
|
|
||||||
organization_name=None,
|
|
||||||
organization_unit_name=None,
|
|
||||||
):
|
|
||||||
#can look at generated file using openssl:
|
|
||||||
#openssl x509 -inform pem -in selfsigned.crt -noout -text
|
|
||||||
# create a key pair
|
|
||||||
if isfile(key_file):
|
|
||||||
with open(key_file) as fh:
|
|
||||||
filecontent = bytes(fh.read(), 'utf-8')
|
|
||||||
key = load_privatekey(FILETYPE_PEM, filecontent)
|
|
||||||
else:
|
|
||||||
key = _gen_key_pair()
|
|
||||||
cert = X509()
|
|
||||||
cert.set_version(2)
|
|
||||||
cert.get_subject().C = country_name
|
|
||||||
cert.get_subject().ST = state_or_province_name
|
|
||||||
cert.get_subject().L = locality_name
|
|
||||||
cert.get_subject().O = organization_name
|
|
||||||
cert.get_subject().OU = organization_unit_name
|
|
||||||
cert.get_subject().CN = common_names[0]
|
|
||||||
cert.get_subject().emailAddress = email_address
|
|
||||||
cert_ext = []
|
|
||||||
if not is_ca:
|
|
||||||
cert_ext.append(X509Extension(b'basicConstraints', False, b'CA:FALSE'))
|
|
||||||
cert_ext.append(X509Extension(b'keyUsage', True, b'digitalSignature, keyEncipherment'))
|
|
||||||
cert_ext.append(X509Extension(b'subjectAltName', False, ", ".join([f'DNS:{common_name}' for common_name in common_names]).encode('ascii')))
|
|
||||||
if type == 'server':
|
|
||||||
cert_ext.append(X509Extension(b'extendedKeyUsage', True, b'serverAuth'))
|
|
||||||
else:
|
|
||||||
cert_ext.append(X509Extension(b'extendedKeyUsage', True, b'clientAuth'))
|
|
||||||
else:
|
|
||||||
cert_ext.append(X509Extension(b'basicConstraints', False, b'CA:TRUE'))
|
|
||||||
cert_ext.append(X509Extension(b"keyUsage", True, b'keyCertSign, cRLSign'))
|
|
||||||
cert_ext.append(X509Extension(b'subjectAltName', False, f'email:{email_address}'.encode()))
|
|
||||||
cert_ext.append(X509Extension(b'subjectKeyIdentifier', False, b"hash", subject=cert))
|
|
||||||
cert.add_extensions(cert_ext)
|
|
||||||
cert.set_serial_number(serial_number)
|
|
||||||
cert.gmtime_adj_notBefore(0)
|
|
||||||
cert.gmtime_adj_notAfter(validity_end_in_seconds)
|
|
||||||
if is_ca:
|
|
||||||
ca_cert = cert
|
|
||||||
ca_key = key
|
|
||||||
else:
|
|
||||||
with open(ca_cert) as fh:
|
|
||||||
filecontent = bytes(fh.read(), 'utf-8')
|
|
||||||
ca_cert = load_certificate(FILETYPE_PEM, filecontent)
|
|
||||||
with open(ca_key) as fh:
|
|
||||||
filecontent = bytes(fh.read(), 'utf-8')
|
|
||||||
ca_key = load_privatekey(FILETYPE_PEM, filecontent)
|
|
||||||
cert.set_issuer(ca_cert.get_subject())
|
|
||||||
cert.add_extensions([X509Extension(b"authorityKeyIdentifier", False, b'keyid:always', issuer=ca_cert)])
|
|
||||||
cert.set_pubkey(key)
|
|
||||||
cert.sign(ca_key, "sha512")
|
|
||||||
|
|
||||||
with open(cert_file, "wt") as f:
|
|
||||||
f.write(dump_certificate(FILETYPE_PEM, cert).decode("utf-8"))
|
|
||||||
with open(key_file, "wt") as f:
|
|
||||||
f.write(dump_privatekey(FILETYPE_PEM, key).decode("utf-8"))
|
|
||||||
|
|
||||||
|
|
||||||
def gen_ca(authority_dns,
|
|
||||||
authority_name,
|
|
||||||
base_dir,
|
|
||||||
):
|
|
||||||
authority_cn = authority_name + '+' + authority_dns
|
|
||||||
week_number = datetime.now().isocalendar().week
|
|
||||||
root_dir_name = join(base_dir, PKI_DIR, authority_cn)
|
|
||||||
ca_dir_name = join(root_dir_name, 'ca')
|
|
||||||
sn_ca_name = join(ca_dir_name, 'serial_number')
|
|
||||||
key_ca_name = join(ca_dir_name, 'private.key')
|
|
||||||
cert_ca_name = join(ca_dir_name, f'certificate_{week_number}.crt')
|
|
||||||
if not isfile(cert_ca_name):
|
|
||||||
if not isdir(ca_dir_name):
|
|
||||||
# rmtree(ca_dir_name)
|
|
||||||
makedirs(ca_dir_name)
|
|
||||||
if isfile(sn_ca_name):
|
|
||||||
with open(sn_ca_name, 'r') as fh:
|
|
||||||
serial_number = int(fh.read().strip()) + 1
|
|
||||||
else:
|
|
||||||
serial_number = 0
|
|
||||||
_gen_cert(True,
|
|
||||||
[authority_cn],
|
|
||||||
serial_number,
|
|
||||||
10*24*60*60,
|
|
||||||
key_ca_name,
|
|
||||||
cert_ca_name,
|
|
||||||
email_address=EMAIL,
|
|
||||||
country_name=COUNTRY,
|
|
||||||
locality_name=LOCALITY,
|
|
||||||
state_or_province_name=STATE,
|
|
||||||
organization_name=ORG_NAME,
|
|
||||||
organization_unit_name=ORG_UNIT_NAME,
|
|
||||||
)
|
|
||||||
with open(sn_ca_name, 'w') as fh:
|
|
||||||
fh.write(str(serial_number))
|
|
||||||
with open(cert_ca_name, 'r') as fh:
|
|
||||||
return fh.read().strip()
|
|
||||||
|
|
||||||
|
|
||||||
def gen_cert_iter(cn,
|
|
||||||
extra_domainnames,
|
|
||||||
authority_cn,
|
|
||||||
authority_name,
|
|
||||||
type,
|
|
||||||
base_dir,
|
|
||||||
dir_name,
|
|
||||||
):
|
|
||||||
week_number = datetime.now().isocalendar().week
|
|
||||||
root_dir_name = join(base_dir, PKI_DIR, authority_cn)
|
|
||||||
ca_dir_name = join(root_dir_name, 'ca')
|
|
||||||
key_ca_name = join(ca_dir_name, 'private.key')
|
|
||||||
cert_ca_name = join(ca_dir_name, f'certificate_{week_number}.crt')
|
|
||||||
sn_name = join(dir_name, f'serial_number')
|
|
||||||
key_name = join(dir_name, f'private.key')
|
|
||||||
cert_name = join(dir_name, f'certificate_{week_number}.crt')
|
|
||||||
if not isfile(cert_ca_name):
|
|
||||||
raise Exception(f'cannot find CA file "{cert_ca_name}"')
|
|
||||||
if not isfile(cert_name):
|
|
||||||
if not isdir(dir_name):
|
|
||||||
makedirs(dir_name)
|
|
||||||
if isfile(sn_name):
|
|
||||||
with open(sn_name, 'r') as fh:
|
|
||||||
serial_number = int(fh.read().strip()) + 1
|
|
||||||
else:
|
|
||||||
serial_number = 0
|
|
||||||
common_names = [cn]
|
|
||||||
common_names.extend(extra_domainnames)
|
|
||||||
_gen_cert(False,
|
|
||||||
common_names,
|
|
||||||
serial_number,
|
|
||||||
10*24*60*60,
|
|
||||||
key_name,
|
|
||||||
cert_name,
|
|
||||||
ca_cert=cert_ca_name,
|
|
||||||
ca_key=key_ca_name,
|
|
||||||
type=type,
|
|
||||||
email_address=EMAIL,
|
|
||||||
country_name=COUNTRY,
|
|
||||||
locality_name=LOCALITY,
|
|
||||||
state_or_province_name=STATE,
|
|
||||||
organization_name=ORG_NAME,
|
|
||||||
organization_unit_name=ORG_UNIT_NAME,
|
|
||||||
)
|
|
||||||
with open(sn_name, 'w') as fh:
|
|
||||||
fh.write(str(serial_number))
|
|
||||||
for extra in extra_domainnames:
|
|
||||||
extra_dir_name = join(base_dir, PKI_DIR, authority_name + '+' + extra)
|
|
||||||
if not exists(extra_dir_name):
|
|
||||||
symlink(root_dir_name, extra_dir_name)
|
|
||||||
for extra in extra_domainnames:
|
|
||||||
extra_dir_name = join(base_dir, PKI_DIR, authority_name + '+' + extra)
|
|
||||||
if not exists(extra_dir_name):
|
|
||||||
raise Exception(f'file {extra_dir_name} not already exists that means subjectAltName is not set in certificat, please remove {cert_name}')
|
|
||||||
return cert_name
|
|
||||||
|
|
||||||
|
|
||||||
def gen_cert(cn,
|
|
||||||
extra_domainnames,
|
|
||||||
authority_cn,
|
|
||||||
authority_name,
|
|
||||||
type,
|
|
||||||
file_type,
|
|
||||||
base_dir,
|
|
||||||
):
|
|
||||||
if '.' in authority_name:
|
|
||||||
raise Exception(f'dot is not allowed in authority_name "{authority_name}"')
|
|
||||||
if type == 'server' and authority_cn is None:
|
|
||||||
authority_cn = cn
|
|
||||||
if authority_cn is None:
|
|
||||||
raise Exception(f'authority_cn is mandatory when authority type is client')
|
|
||||||
if extra_domainnames is None:
|
|
||||||
extra_domainnames = []
|
|
||||||
auth_cn = authority_name + '+' + authority_cn
|
|
||||||
dir_name = join(base_dir, PKI_DIR, auth_cn, 'certificats', cn, type)
|
|
||||||
if file_type == 'crt':
|
|
||||||
filename = gen_cert_iter(cn,
|
|
||||||
extra_domainnames,
|
|
||||||
auth_cn,
|
|
||||||
authority_name,
|
|
||||||
type,
|
|
||||||
base_dir,
|
|
||||||
dir_name,
|
|
||||||
)
|
|
||||||
else:
|
|
||||||
filename = join(dir_name, f'private.key')
|
|
||||||
with open(filename, 'r') as fh:
|
|
||||||
return fh.read().strip()
|
|
||||||
|
|
||||||
|
|
||||||
def has_pub(cn,
|
|
||||||
base_dir,
|
|
||||||
):
|
|
||||||
dir_name = join(base_dir, PKI_DIR, 'public', cn)
|
|
||||||
cert_name = join(dir_name, f'public.pub')
|
|
||||||
return isfile(cert_name)
|
|
||||||
|
|
||||||
|
|
||||||
def gen_pub(cn,
|
|
||||||
file_type,
|
|
||||||
base_dir,
|
|
||||||
):
|
|
||||||
dir_name = join(base_dir, PKI_DIR, 'public', cn)
|
|
||||||
key_name = join(dir_name, f'private.key')
|
|
||||||
if file_type == 'pub':
|
|
||||||
pub_name = join(dir_name, f'public.pub')
|
|
||||||
if not isfile(pub_name):
|
|
||||||
if not isdir(dir_name):
|
|
||||||
makedirs(dir_name)
|
|
||||||
key = _gen_key_pair()
|
|
||||||
with open(pub_name, "wt") as f:
|
|
||||||
f.write(dump_publickey(FILETYPE_PEM, key).decode("utf-8"))
|
|
||||||
with open(key_name, "wt") as f:
|
|
||||||
f.write(dump_privatekey(FILETYPE_PEM, key).decode("utf-8"))
|
|
||||||
filename = pub_name
|
|
||||||
else:
|
|
||||||
filename = key_name
|
|
||||||
with open(filename, 'r') as fh:
|
|
||||||
return fh.read().strip()
|
|
468
test.py
|
@ -1,468 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
|
|
||||||
from asyncio import run
|
|
||||||
from os import listdir, link, makedirs, environ
|
|
||||||
from os.path import isdir, isfile, join
|
|
||||||
from shutil import rmtree, copy2, copytree
|
|
||||||
from json import load as json_load
|
|
||||||
from yaml import load, SafeLoader
|
|
||||||
from toml import load as toml_load
|
|
||||||
from pprint import pprint
|
|
||||||
from typing import Any
|
|
||||||
from warnings import warn_explicit
|
|
||||||
from copy import copy
|
|
||||||
|
|
||||||
from tiramisu import Config
|
|
||||||
from tiramisu.error import ValueWarning
|
|
||||||
from rougail import RougailConfig, RougailConvert, RougailSystemdTemplate
|
|
||||||
from rougail.utils import normalize_family
|
|
||||||
|
|
||||||
from risotto.utils import MULTI_FUNCTIONS, CONFIGS
|
|
||||||
|
|
||||||
|
|
||||||
with open(environ.get('CONFIG_FILE', 'risotto.conf'), 'r') as fh:
|
|
||||||
config = toml_load(fh)
|
|
||||||
|
|
||||||
|
|
||||||
DATASET_DIRECTORY = config['directories']['dataset']
|
|
||||||
INSTALL_DIR = config['directories']['dest']
|
|
||||||
FUNCTIONS = 'funcs.py'
|
|
||||||
CONFIG_DEST_DIR = 'configurations'
|
|
||||||
SRV_DEST_DIR = 'srv'
|
|
||||||
|
|
||||||
|
|
||||||
with open('servers.json', 'r') as server_fh:
|
|
||||||
jsonfile = json_load(server_fh)
|
|
||||||
SERVERS = jsonfile['servers']
|
|
||||||
MODULES = jsonfile['modules']
|
|
||||||
|
|
||||||
|
|
||||||
async def set_linked(linked_server: str,
|
|
||||||
linked_provider: str,
|
|
||||||
linked_value: str,
|
|
||||||
linked_returns: str=None,
|
|
||||||
dynamic: str=None,
|
|
||||||
):
|
|
||||||
if None in (linked_server, linked_provider, linked_value):
|
|
||||||
return
|
|
||||||
if linked_server not in CONFIGS:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
0,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
config = CONFIGS[linked_server][0]
|
|
||||||
path = await config.information.get('provider:' + linked_provider, None)
|
|
||||||
if not path:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find provider "{linked_provider}" in linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
0,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
await config.property.read_write()
|
|
||||||
try:
|
|
||||||
option = config.forcepermissive.option(path)
|
|
||||||
if await option.option.ismulti():
|
|
||||||
values = await option.value.get()
|
|
||||||
if linked_value not in values:
|
|
||||||
values.append(linked_value)
|
|
||||||
await option.value.set(values)
|
|
||||||
else:
|
|
||||||
await option.value.set(linked_value)
|
|
||||||
except Exception as err:
|
|
||||||
await config.property.read_only()
|
|
||||||
raise err from err
|
|
||||||
await config.property.read_only()
|
|
||||||
if linked_returns is not None:
|
|
||||||
linked_variable = await config.information.get('provider:' + linked_returns, None)
|
|
||||||
if not linked_variable:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find linked variable "{linked_returns}" in linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
0,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
else:
|
|
||||||
linked_variable = None
|
|
||||||
if linked_variable is not None:
|
|
||||||
if dynamic:
|
|
||||||
linked_variable = linked_variable.replace('{suffix}', normalize_family(dynamic))
|
|
||||||
elif '{suffix}' in linked_variable:
|
|
||||||
idx = CONFIGS[linked_server][3]
|
|
||||||
linked_variable = linked_variable.replace('{suffix}', str(idx))
|
|
||||||
ret = await config.forcepermissive.option(linked_variable).value.get()
|
|
||||||
else:
|
|
||||||
ret = normalize_family(linked_value)
|
|
||||||
return ret
|
|
||||||
|
|
||||||
|
|
||||||
async def get_linked_configuration(linked_server: str,
|
|
||||||
linked_provider: str,
|
|
||||||
dynamic: str=None,
|
|
||||||
):
|
|
||||||
if linked_server not in CONFIGS:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
1,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
config = CONFIGS[linked_server][0]
|
|
||||||
path = await config.information.get('provider:' + linked_provider, None)
|
|
||||||
if not path:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find variable "{path}" in linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
1,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
if dynamic:
|
|
||||||
path = path.replace('{suffix}', normalize_family(dynamic))
|
|
||||||
try:
|
|
||||||
return await config.forcepermissive.option(path).value.get()
|
|
||||||
except AttributeError as err:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find get value of "{path}" in linked server "{linked_server}": {err}'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
1,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class Empty:
|
|
||||||
pass
|
|
||||||
empty = Empty()
|
|
||||||
|
|
||||||
|
|
||||||
async def set_linked_configuration(_linked_value: Any,
|
|
||||||
linked_server: str,
|
|
||||||
linked_provider: str,
|
|
||||||
linked_value: Any=empty,
|
|
||||||
dynamic: str=None,
|
|
||||||
leader_provider: str=None,
|
|
||||||
leader_value: Any=None,
|
|
||||||
leader_index: int=None,
|
|
||||||
):
|
|
||||||
if linked_value is not empty:
|
|
||||||
_linked_value = linked_value
|
|
||||||
linked_value = _linked_value
|
|
||||||
if linked_server is None:
|
|
||||||
return
|
|
||||||
if linked_value is None or linked_server not in CONFIGS:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
2,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
config = CONFIGS[linked_server][0]
|
|
||||||
path = await config.information.get('provider:' + linked_provider, None)
|
|
||||||
if not path:
|
|
||||||
warn_explicit(ValueWarning(f'cannot find variable "{path}" in linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
2,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
if dynamic:
|
|
||||||
path = path.replace('{suffix}', normalize_family(dynamic))
|
|
||||||
await config.property.read_write()
|
|
||||||
try:
|
|
||||||
if leader_provider is not None:
|
|
||||||
leader_path = await config.information.get('provider:' + leader_provider, None)
|
|
||||||
if not leader_path:
|
|
||||||
await config.property.read_only()
|
|
||||||
warn_explicit(ValueWarning(f'cannot find leader variable with leader_provider "{leader_provider}" in linked server "{linked_server}"'),
|
|
||||||
ValueWarning,
|
|
||||||
__file__,
|
|
||||||
2,
|
|
||||||
)
|
|
||||||
return
|
|
||||||
if dynamic:
|
|
||||||
leader_path = leader_path.replace('{suffix}', normalize_family(dynamic))
|
|
||||||
values = await config.forcepermissive.option(leader_path).value.get()
|
|
||||||
if not isinstance(leader_value, list):
|
|
||||||
leader_value = [leader_value]
|
|
||||||
for lv in leader_value:
|
|
||||||
if lv in values:
|
|
||||||
slave_idx = values.index(lv)
|
|
||||||
slave_option = config.forcepermissive.option(path, slave_idx)
|
|
||||||
if await slave_option.option.issubmulti():
|
|
||||||
slave_values = await slave_option.value.get()
|
|
||||||
if linked_value not in slave_values:
|
|
||||||
slave_values.append(linked_value)
|
|
||||||
await slave_option.value.set(slave_values)
|
|
||||||
else:
|
|
||||||
await slave_option.value.set(linked_value)
|
|
||||||
else:
|
|
||||||
option = config.forcepermissive.option(path, leader_index)
|
|
||||||
if leader_index is None and await option.option.ismulti() and not isinstance(linked_value, list):
|
|
||||||
values = await option.value.get()
|
|
||||||
if linked_value not in values:
|
|
||||||
values.append(linked_value)
|
|
||||||
await option.value.set(values)
|
|
||||||
else:
|
|
||||||
await option.value.set(linked_value)
|
|
||||||
except AttributeError as err:
|
|
||||||
#raise ValueError(str(err)) from err
|
|
||||||
pass
|
|
||||||
except Exception as err:
|
|
||||||
await config.property.read_only()
|
|
||||||
raise err from err
|
|
||||||
await config.property.read_only()
|
|
||||||
|
|
||||||
|
|
||||||
def tiramisu_display_name(kls,
|
|
||||||
dyn_name: 'Base'=None,
|
|
||||||
suffix: str=None,
|
|
||||||
) -> str:
|
|
||||||
if dyn_name is not None:
|
|
||||||
name = kls.impl_getpath() + suffix
|
|
||||||
else:
|
|
||||||
name = kls.impl_getpath()
|
|
||||||
return name
|
|
||||||
|
|
||||||
|
|
||||||
def load_applications():
|
|
||||||
applications = {}
|
|
||||||
distrib_dir = join(DATASET_DIRECTORY, 'applicationservice')
|
|
||||||
for release in listdir(distrib_dir):
|
|
||||||
release_dir = join(distrib_dir, release)
|
|
||||||
if not isdir(release_dir):
|
|
||||||
continue
|
|
||||||
for applicationservice in listdir(release_dir):
|
|
||||||
applicationservice_dir = join(release_dir, applicationservice)
|
|
||||||
if not isdir(applicationservice_dir):
|
|
||||||
continue
|
|
||||||
if applicationservice in applications:
|
|
||||||
raise Exception(f'multi applicationservice: {applicationservice} ({applicationservice_dir} <=> {applications[applicationservice]})')
|
|
||||||
applications[applicationservice] = applicationservice_dir
|
|
||||||
return applications
|
|
||||||
|
|
||||||
|
|
||||||
class ModuleCfg():
|
|
||||||
def __init__(self):
|
|
||||||
self.dictionaries_dir = []
|
|
||||||
self.modules = []
|
|
||||||
self.functions_file = [FUNCTIONS]
|
|
||||||
self.templates_dir = []
|
|
||||||
self.extra_dictionaries = {}
|
|
||||||
self.servers = []
|
|
||||||
|
|
||||||
|
|
||||||
def build_module(module_name, datas, module_infos):
|
|
||||||
install_dir = join(INSTALL_DIR, module_name)
|
|
||||||
makedirs(install_dir)
|
|
||||||
applications = load_applications()
|
|
||||||
cfg = ModuleCfg()
|
|
||||||
module_infos[module_name] = cfg
|
|
||||||
def calc_depends(appname, added):
|
|
||||||
if appname in added:
|
|
||||||
return
|
|
||||||
as_dir = applications[appname]
|
|
||||||
cfg.modules.append(appname)
|
|
||||||
dictionaries_dir = join(as_dir, 'dictionaries')
|
|
||||||
if isdir(dictionaries_dir):
|
|
||||||
cfg.dictionaries_dir.append(dictionaries_dir)
|
|
||||||
funcs_dir = join(as_dir, 'funcs')
|
|
||||||
if isdir(funcs_dir):
|
|
||||||
for f in listdir(funcs_dir):
|
|
||||||
if f.startswith('__'):
|
|
||||||
continue
|
|
||||||
cfg.functions_file.append(join(funcs_dir, f))
|
|
||||||
templates_dir = join(as_dir, 'templates')
|
|
||||||
if isdir(templates_dir):
|
|
||||||
cfg.templates_dir.append(templates_dir)
|
|
||||||
extras_dir = join(as_dir, 'extras')
|
|
||||||
if isdir(extras_dir):
|
|
||||||
for extra in listdir(extras_dir):
|
|
||||||
extra_dir = join(extras_dir, extra)
|
|
||||||
if isdir(extra_dir):
|
|
||||||
cfg.extra_dictionaries.setdefault(extra, []).append(extra_dir)
|
|
||||||
for type in ['image', 'install']:
|
|
||||||
manual_dir = join(as_dir, 'manual', type)
|
|
||||||
if isdir(manual_dir):
|
|
||||||
for filename in listdir(manual_dir):
|
|
||||||
src_file = join(manual_dir, filename)
|
|
||||||
if type == 'image':
|
|
||||||
dst_file = join(install_dir, filename)
|
|
||||||
verify = False
|
|
||||||
else:
|
|
||||||
dst_file= join(INSTALL_DIR, filename)
|
|
||||||
verify = True
|
|
||||||
if isdir(src_file):
|
|
||||||
if not isdir(dst_file):
|
|
||||||
makedirs(dst_file)
|
|
||||||
for subfilename in listdir(src_file):
|
|
||||||
if not verify or not isfile(dst_file):
|
|
||||||
src = join(src_file, subfilename)
|
|
||||||
dst = join(dst_file, subfilename)
|
|
||||||
if isfile(src):
|
|
||||||
copy2(src, dst)
|
|
||||||
else:
|
|
||||||
copytree(src, dst)
|
|
||||||
elif not verify or not isfile(dst_file):
|
|
||||||
src = join(manual_dir, filename)
|
|
||||||
dst = dst_file
|
|
||||||
if isfile(src):
|
|
||||||
copy2(src, dst)
|
|
||||||
else:
|
|
||||||
copytree(src, dst)
|
|
||||||
added.append(appname)
|
|
||||||
with open(join(as_dir, 'applicationservice.yml')) as yaml:
|
|
||||||
app = load(yaml, Loader=SafeLoader)
|
|
||||||
|
|
||||||
for xml in app.get('depends', []):
|
|
||||||
calc_depends(xml, added)
|
|
||||||
added = []
|
|
||||||
for applicationservice in datas['applicationservices']:
|
|
||||||
calc_depends(applicationservice, added)
|
|
||||||
|
|
||||||
|
|
||||||
async def build(server_name, datas, module_infos):
|
|
||||||
if server_name in CONFIGS:
|
|
||||||
raise Exception(f'server "{server_name}" is duplicate')
|
|
||||||
cfg = RougailConfig.copy()
|
|
||||||
module_info = module_infos[datas['module']]
|
|
||||||
module_info.servers.append(server_name)
|
|
||||||
if datas['module'] == 'host':
|
|
||||||
cfg['tmpfile_dest_dir'] = datas['values']['rougail.host_install_dir'] + '/host/configurations/' + server_name
|
|
||||||
cfg['templates_dir'] = module_info.templates_dir
|
|
||||||
cfg['dictionaries_dir'] = module_info.dictionaries_dir
|
|
||||||
cfg['functions_file'] = module_info.functions_file
|
|
||||||
cfg['multi_functions'] = MULTI_FUNCTIONS
|
|
||||||
cfg['extra_dictionaries'] = module_info.extra_dictionaries
|
|
||||||
cfg['extra_annotators'].append('risotto.rougail')
|
|
||||||
optiondescription = {'set_linked': set_linked,
|
|
||||||
'get_linked_configuration': get_linked_configuration,
|
|
||||||
'set_linked_configuration': set_linked_configuration,
|
|
||||||
}
|
|
||||||
cfg['internal_functions'] = list(optiondescription.keys())
|
|
||||||
try:
|
|
||||||
eolobj = RougailConvert(cfg)
|
|
||||||
except Exception as err:
|
|
||||||
print(f'Try to load {module_info.modules}')
|
|
||||||
raise err from err
|
|
||||||
tiram_obj = eolobj.save(None)
|
|
||||||
# if server_name == 'revprox.in.silique.fr':
|
|
||||||
# print(tiram_obj)
|
|
||||||
#cfg['patches_dir'] = join(test_dir, 'patches')
|
|
||||||
cfg['tmp_dir'] = 'tmp'
|
|
||||||
cfg['destinations_dir'] = join(INSTALL_DIR, datas['module'], CONFIG_DEST_DIR, server_name)
|
|
||||||
if isdir('tmp'):
|
|
||||||
rmtree('tmp')
|
|
||||||
makedirs('tmp')
|
|
||||||
makedirs(cfg['destinations_dir'])
|
|
||||||
try:
|
|
||||||
exec(tiram_obj, None, optiondescription)
|
|
||||||
except Exception as err:
|
|
||||||
print(tiram_obj)
|
|
||||||
raise Exception(f'unknown error when load tiramisu object {err}') from err
|
|
||||||
config = await Config(optiondescription['option_0'], display_name=tiramisu_display_name)
|
|
||||||
await config.property.read_write()
|
|
||||||
try:
|
|
||||||
if await config.option('machine.add_srv').value.get():
|
|
||||||
srv = join(INSTALL_DIR, SRV_DEST_DIR, server_name)
|
|
||||||
else:
|
|
||||||
srv = None
|
|
||||||
except AttributeError:
|
|
||||||
srv = None
|
|
||||||
await config.property.read_write()
|
|
||||||
CONFIGS[server_name] = (config, cfg, srv, 0)
|
|
||||||
|
|
||||||
|
|
||||||
async def value_pprint(dico, config):
|
|
||||||
pprint_dict = {}
|
|
||||||
for path, value in dico.items():
|
|
||||||
if await config.option(path).option.type() == 'password' and value:
|
|
||||||
value = 'X' * len(value)
|
|
||||||
pprint_dict[path] = value
|
|
||||||
pprint(pprint_dict)
|
|
||||||
|
|
||||||
|
|
||||||
async def set_values(server_name, config, datas):
|
|
||||||
if 'informations' in datas:
|
|
||||||
for information, value in datas['informations'].items():
|
|
||||||
await config.information.set(information, value)
|
|
||||||
if 'extra_domainnames' in datas['informations']:
|
|
||||||
for idx, extra_domainname in enumerate(datas['informations']['extra_domainnames']):
|
|
||||||
if extra_domainname in CONFIGS:
|
|
||||||
raise Exception(f'server "{server_name}" is duplicate')
|
|
||||||
value = list(CONFIGS[server_name])
|
|
||||||
value[3] = idx + 1
|
|
||||||
CONFIGS[extra_domainname] = tuple(value)
|
|
||||||
await config.information.set('server_name', server_name)
|
|
||||||
await config.property.read_write()
|
|
||||||
try:
|
|
||||||
if 'values' in datas:
|
|
||||||
for path, value in datas['values'].items():
|
|
||||||
if isinstance(value, dict):
|
|
||||||
for idx, val in value.items():
|
|
||||||
await config.option(path, int(idx)).value.set(val)
|
|
||||||
else:
|
|
||||||
await config.option(path).value.set(value)
|
|
||||||
except Exception as err:
|
|
||||||
await value_pprint(await config.value.dict(), config)
|
|
||||||
error_msg = f'cannot configure server "{server_name}": {err}'
|
|
||||||
raise Exception(error_msg) from err
|
|
||||||
await config.property.read_only()
|
|
||||||
#await config.value.dict()
|
|
||||||
|
|
||||||
|
|
||||||
async def valid_mandatories(server_name, config):
|
|
||||||
mandatories = await config.value.mandatory()
|
|
||||||
if mandatories:
|
|
||||||
print()
|
|
||||||
print(f'=== Configuration: {server_name} ===')
|
|
||||||
await config.property.pop('mandatory')
|
|
||||||
await value_pprint(await config.value.dict(), config)
|
|
||||||
raise Exception(f'server "{server_name}" has mandatories variables without values "{", ".join(mandatories)}"')
|
|
||||||
|
|
||||||
|
|
||||||
async def templates(server_name, config, cfg, srv, int_idx):
|
|
||||||
values = await config.value.dict()
|
|
||||||
engine = RougailSystemdTemplate(config, cfg)
|
|
||||||
# if server_name == 'revprox.in.silique.fr':
|
|
||||||
# print()
|
|
||||||
# print(f'=== Configuration: {server_name} ===')
|
|
||||||
# pprint(values)
|
|
||||||
try:
|
|
||||||
await engine.instance_files()
|
|
||||||
except Exception as err:
|
|
||||||
print()
|
|
||||||
print(f'=== Configuration: {server_name} ===')
|
|
||||||
await value_pprint(values, config)
|
|
||||||
raise err from err
|
|
||||||
if srv:
|
|
||||||
makedirs(srv)
|
|
||||||
|
|
||||||
|
|
||||||
async def main():
|
|
||||||
if isdir(INSTALL_DIR):
|
|
||||||
rmtree(INSTALL_DIR)
|
|
||||||
makedirs(INSTALL_DIR)
|
|
||||||
module_infos = {}
|
|
||||||
for module_name, datas in MODULES.items():
|
|
||||||
build_module(module_name, datas, module_infos)
|
|
||||||
for server_name, datas in SERVERS.items():
|
|
||||||
await build(server_name, datas, module_infos)
|
|
||||||
for module_name, cfg in module_infos.items():
|
|
||||||
with open(join(INSTALL_DIR, module_name, 'install_machines'), 'w') as fh:
|
|
||||||
for server_name in cfg.servers:
|
|
||||||
fh.write(f'./install_machine {module_name} {server_name}\n')
|
|
||||||
for server_name, datas in SERVERS.items():
|
|
||||||
await set_values(server_name, CONFIGS[server_name][0], datas)
|
|
||||||
for server_name in SERVERS:
|
|
||||||
config = CONFIGS[server_name][0]
|
|
||||||
await config.property.pop('mandatory')
|
|
||||||
await config.value.dict()
|
|
||||||
await config.property.add('mandatory')
|
|
||||||
for server_name in SERVERS:
|
|
||||||
await valid_mandatories(server_name, CONFIGS[server_name][0])
|
|
||||||
# print(await CONFIGS['revprox.in.gnunux.info'][0].option('nginx.reverse_proxy_for_netbox_in_gnunux_info.reverse_proxy_netbox_in_gnunux_info.revprox_url_netbox_in_gnunux_info', 0).value.get())
|
|
||||||
for server_name in SERVERS:
|
|
||||||
await templates(server_name, *CONFIGS[server_name])
|
|
||||||
|
|
||||||
|
|
||||||
run(main())
|
|