From: Christian Pointner Date: Sat, 21 Apr 2018 21:56:18 +0000 (+0200) Subject: add initial vm installer role X-Git-Url: https://git.realraum.at/?a=commitdiff_plain;h=cada7766cb94b291b214784fa717f874db922d57;p=noc.git add initial vm installer role --- diff --git a/ansible/ansible.cfg b/ansible/ansible.cfg index 3be1971..4c457eb 100644 --- a/ansible/ansible.cfg +++ b/ansible/ansible.cfg @@ -1,5 +1,6 @@ [defaults] inventory = ./hosts.ini +roles_path = ./roles:../roles log_path = ./log nocows=1 vault_password_file = ./gpg/get-vault-pass.sh diff --git a/ansible/host_playbooks/testvm.yml b/ansible/host_playbooks/testvm.yml new file mode 100644 index 0000000..e79d3e3 --- /dev/null +++ b/ansible/host_playbooks/testvm.yml @@ -0,0 +1,7 @@ +--- +- name: Basic Setup for server: testvm + hosts: + roles: + - role: base + - role: vm-grub + - role: vm-network diff --git a/ansible/host_vars/alfred/main.yml b/ansible/host_vars/alfred/main.yml new file mode 100644 index 0000000..b6b9b0c --- /dev/null +++ b/ansible/host_vars/alfred/main.yml @@ -0,0 +1,27 @@ +--- +vm_host: + installer: + net_if: br-mgmt + preseed_path: /srv/preseed + path: /srv/installer + distros: + - distro: debian + codename: stretch + arch: + - amd64 + - i386 + - distro: ubuntu + codename: xenial + arch: + - amd64 + - i386 + network: + interface: br-mgmt + ip: 192.168.33.65 + mask: 255.255.255.0 + gateway: 192.168.33.1 + nameservers: + - 192.168.33.1 + - 10.12.0.10 + indices: + testvm: 200 diff --git a/ansible/host_vars/testvm/main.yml b/ansible/host_vars/testvm/main.yml new file mode 100644 index 0000000..479861c --- /dev/null +++ b/ansible/host_vars/testvm/main.yml @@ -0,0 +1,30 @@ +--- +vm_install_host: alfred.mgmt + +vm_install: + host: "{{ vm_install_host }}" + mem: 1024 + numcpu: 2 + disks: + primary: vda + virtio: + vda: + vg: storage + lv: "{{ inventory_hostname }}" + size: 10g + interfaces: + - idx: 1 + bridge: "{{ hostvars[vm_install_host].vm_host.network.interface }}" + name: primary0 + autostart: True + +vm_network: + systemd_link: + interfaces: "{{ vm_install.interfaces }}" + primary: + interface: primary0 + ip: "{{ (hostvars[vm_install_host].vm_host.network.ip+'/'+hostvars[vm_install_host].vm_host.network.mask) | ipaddr(hostvars[vm_install_host].vm_host.network.indices[inventory_hostname]) | ipaddr('address') }}" + mask: "{{ hostvars[vm_install_host].vm_host.network.mask }}" + gateway: "{{ hostvars[vm_install_host].vm_host.network.gateway | default(hostvars[vm_install_host].vm_host.network.ip) }}" + nameservers: "{{ hostvars[vm_install_host].vm_host.network.nameservers }}" + domain: realraum.at diff --git a/ansible/remove-known-host.sh b/ansible/remove-known-host.sh new file mode 100755 index 0000000..d2c8b8d --- /dev/null +++ b/ansible/remove-known-host.sh @@ -0,0 +1,15 @@ +#!/bin/bash + +if [ -z "$1" ]; then + echo "$0 " + exit 1 +fi + +SHORT="$1" +SSH_HOST=$(ssh -G "$1" | grep "^hostname " | awk '{ print($2) }' ) + +for name in $SHORT $SSH_HOST; do + ssh-keygen -f "$HOME/.ssh/known_hosts" -R "$name" +done + +exit 0 diff --git a/ansible/roles/reboot-and-wait/defaults/main.yml b/ansible/roles/reboot-and-wait/defaults/main.yml new file mode 100644 index 0000000..d25b709 --- /dev/null +++ b/ansible/roles/reboot-and-wait/defaults/main.yml @@ -0,0 +1,3 @@ +--- +reboot_delay: 60 +reboot_timeout: 300 diff --git a/ansible/roles/reboot-and-wait/tasks/main.yml b/ansible/roles/reboot-and-wait/tasks/main.yml new file mode 100644 index 0000000..e648f0a --- /dev/null +++ b/ansible/roles/reboot-and-wait/tasks/main.yml @@ -0,0 +1,10 @@ +- name: reboot machine + shell: sleep 2 && shutdown -r now + async: 1 + poll: 0 + ignore_errors: true + +- name: waiting for host to come back + wait_for_connection: + delay: "{{ reboot_delay }}" + timeout: "{{ reboot_timeout }}" diff --git a/ansible/roles/vm-grub/handlers/main.yml b/ansible/roles/vm-grub/handlers/main.yml new file mode 100644 index 0000000..4bddbb1 --- /dev/null +++ b/ansible/roles/vm-grub/handlers/main.yml @@ -0,0 +1,3 @@ +--- +- name: update grub + command: /usr/sbin/update-grub diff --git a/ansible/roles/vm-grub/tasks/main.yml b/ansible/roles/vm-grub/tasks/main.yml new file mode 100644 index 0000000..f751243 --- /dev/null +++ b/ansible/roles/vm-grub/tasks/main.yml @@ -0,0 +1,16 @@ +--- +- name: enable serial console in grub and for kernel + with_items: + - regexp: '^GRUB_TIMEOUT=' + line: 'GRUB_TIMEOUT=2' + - regexp: '^GRUB_CMDLINE_LINUX=' + line: 'GRUB_CMDLINE_LINUX="console=ttyS0,115200n8"' + - regexp: '^GRUB_TERMINAL=' + line: 'GRUB_TERMINAL=serial' + - regexp: '^GRUB_SERIAL_COMMAND=' + line: 'GRUB_SERIAL_COMMAND="serial --unit=0 --speed=115200 --word=8 --parity=no --stop=1"' + lineinfile: + dest: /etc/default/grub + regexp: "{{ item.regexp }}" + line: "{{ item.line }}" + notify: update grub diff --git a/ansible/roles/vm-host/defaults/main.yml b/ansible/roles/vm-host/defaults/main.yml new file mode 100644 index 0000000..deaa50a --- /dev/null +++ b/ansible/roles/vm-host/defaults/main.yml @@ -0,0 +1,5 @@ +--- +vm_host_force_download_installer: False +vm_host_installer_url: + debian: "http://debian.mur.at/debian" + ubuntu: "http://ubuntu.uni-klu.ac.at/ubuntu" diff --git a/ansible/roles/vm-host/handlers/main.yml b/ansible/roles/vm-host/handlers/main.yml new file mode 100644 index 0000000..158f4dc --- /dev/null +++ b/ansible/roles/vm-host/handlers/main.yml @@ -0,0 +1,5 @@ +--- +- name: restart inetd + service: + name: openbsd-inetd + state: restarted diff --git a/ansible/roles/vm-host/tasks/main.yml b/ansible/roles/vm-host/tasks/main.yml new file mode 100644 index 0000000..248f855 --- /dev/null +++ b/ansible/roles/vm-host/tasks/main.yml @@ -0,0 +1,53 @@ +--- +- name: install tftpd and python-libvirt + apt: + name: + - atftpd + - openbsd-inetd + - qemu-kvm + - libvirt-bin + - python-libvirt + state: present + +- name: configure tftpd via inetd + lineinfile: + regexp: "^#?({{ vm_host.network.ip }}:)?tftp" + line: "{{ vm_host.network.ip }}:tftp dgram udp4 wait nobody /usr/sbin/tcpd /usr/sbin/in.tftpd --tftpd-timeout 300 --retry-timeout 5 --maxthread 10 --verbose=5 {{ vm_host.installer.preseed_path }}" + path: /etc/inetd.conf + notify: restart inetd + +- name: make sure installer directories exists + with_items: + - "{{ vm_host.installer.path }}" + - "{{ vm_host.installer.preseed_path }}" + file: + name: "{{ item }}" + state: directory + +- name: prepare directories for installer images + with_subelements: + - "{{ vm_host.installer.distros }}" + - arch + file: + name: "{{ vm_host.installer.path }}/{{ item.0.distro }}-{{ item.0.codename }}/{{ item.1 }}" + state: directory + +- name: download installer kernel images + with_subelements: + - "{{ vm_host.installer.distros }}" + - arch + get_url: + url: "{{ vm_host_installer_url[item.0.distro] }}/dists/{{ item.0.codename }}/main/installer-{{ item.1 }}/current/images/netboot/{{ item.0.distro }}-installer/{{ item.1 }}/linux" + dest: "{{ vm_host.installer.path }}/{{ item.0.distro }}-{{ item.0.codename }}/{{ item.1 }}/linux" + mode: 0644 + force: "{{ vm_host_force_download_installer }}" + +- name: download installer initrd.gz + with_subelements: + - "{{ vm_host.installer.distros }}" + - arch + get_url: + url: "{{ vm_host_installer_url[item.0.distro] }}/dists/{{ item.0.codename }}/main/installer-{{ item.1 }}/current/images/netboot/{{ item.0.distro }}-installer/{{ item.1 }}/initrd.gz" + dest: "{{ vm_host.installer.path }}/{{ item.0.distro }}-{{ item.0.codename }}/{{ item.1 }}/initrd.gz" + mode: 0644 + force: "{{ vm_host_force_download_installer }}" diff --git a/ansible/roles/vm-install/library/wait_for_virt.py b/ansible/roles/vm-install/library/wait_for_virt.py new file mode 100644 index 0000000..6c49fae --- /dev/null +++ b/ansible/roles/vm-install/library/wait_for_virt.py @@ -0,0 +1,179 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +from __future__ import absolute_import, division, print_function +__metaclass__ = type + + +import traceback +import time + +try: + import libvirt +except ImportError: + HAS_VIRT = False +else: + HAS_VIRT = True + +from ansible.module_utils.basic import AnsibleModule +from ansible.module_utils._text import to_native + + +VIRT_FAILED = 1 +VIRT_SUCCESS = 0 +VIRT_UNAVAILABLE = 2 + +VIRT_STATE_NAME_MAP = { + 0: "running", + 1: "running", + 2: "running", + 3: "paused", + 4: "shutdown", + 5: "shutdown", + 6: "crashed" +} + + +class VMNotFound(Exception): + pass + + +class LibvirtConnection(object): + + def __init__(self, uri, module): + + self.module = module + + cmd = "uname -r" + rc, stdout, stderr = self.module.run_command(cmd) + + if "xen" in stdout: + conn = libvirt.open(None) + elif "esx" in uri: + auth = [[libvirt.VIR_CRED_AUTHNAME, libvirt.VIR_CRED_NOECHOPROMPT], [], None] + conn = libvirt.openAuth(uri, auth) + else: + conn = libvirt.open(uri) + + if not conn: + raise Exception("hypervisor connection failure") + + self.conn = conn + + def find_vm(self, vmid): + """ + Extra bonus feature: vmid = -1 returns a list of everything + """ + conn = self.conn + + vms = [] + + # this block of code borrowed from virt-manager: + # get working domain's name + ids = conn.listDomainsID() + for id in ids: + vm = conn.lookupByID(id) + vms.append(vm) + # get defined domain + names = conn.listDefinedDomains() + for name in names: + vm = conn.lookupByName(name) + vms.append(vm) + + if vmid == -1: + return vms + + for vm in vms: + if vm.name() == vmid: + return vm + + raise VMNotFound("virtual machine %s not found" % vmid) + + def get_status(self, vmid): + state = self.find_vm(vmid).info()[0] + return VIRT_STATE_NAME_MAP.get(state, "unknown") + + +class Virt(object): + + def __init__(self, uri, module): + self.module = module + self.uri = uri + + def __get_conn(self): + self.conn = LibvirtConnection(self.uri, self.module) + return self.conn + + def status(self, vmid): + """ + Return a state suitable for server consumption. Aka, codes.py values, not XM output. + """ + self.__get_conn() + return self.conn.get_status(vmid) + + +def core(module): + + states = module.params.get('states', None) + guest = module.params.get('name', None) + uri = module.params.get('uri', None) + delay = module.params.get('delay', None) + sleep = module.params.get('sleep', None) + timeout = module.params.get('timeout', None) + + v = Virt(uri, module) + res = {'changed': False, 'failed': True} + + if delay > 0: + time.sleep(delay) + + for _ in range(0, timeout, sleep): + state = v.status(guest) + if state in states: + res['state'] = state + res['failed'] = False + res['msg'] = "guest '%s' has reached state: %s" % (guest, state) + return VIRT_SUCCESS, res + + time.sleep(sleep) + + res['msg'] = "timeout waiting for guest '%s' to reach one of states: %s" % (guest, ', '.join(states)) + return VIRT_FAILED, res + + +def main(): + + module = AnsibleModule(argument_spec=dict( + name=dict(aliases=['guest'], required=True), + states=dict(type='list', required=True), + uri=dict(default='qemu:///system'), + delay=dict(type='int', default=0), + sleep=dict(type='int', default=1), + timeout=dict(type='int', default=300), + )) + + if not HAS_VIRT: + module.fail_json( + msg='The `libvirt` module is not importable. Check the requirements.' + ) + + for state in module.params.get('states', None): + if state not in set(VIRT_STATE_NAME_MAP.values()): + module.fail_json( + msg="states contains invalid state '%s', must be one of %s" % (state, ', '.join(set(VIRT_STATE_NAME_MAP.values()))) + ) + + rc = VIRT_SUCCESS + try: + rc, result = core(module) + except Exception as e: + module.fail_json(msg=to_native(e), exception=traceback.format_exc()) + + if rc != 0: # something went wrong emit the msg + module.fail_json(rc=rc, msg=result) + else: + module.exit_json(**result) + + +if __name__ == '__main__': + main() diff --git a/ansible/roles/vm-install/tasks/main.yml b/ansible/roles/vm-install/tasks/main.yml new file mode 100644 index 0000000..037632f --- /dev/null +++ b/ansible/roles/vm-install/tasks/main.yml @@ -0,0 +1,124 @@ +--- +- name: generate preseed file + template: + src: "preseed_{{ vmdistro }}-{{ vmdistcodename }}.cfg.j2" + dest: "{{ hostvars[vm_install.host].vm_host.installer.preseed_path }}/vm-{{ inventory_hostname }}-{{ vmdistro }}-{{ vmdistcodename }}.cfg" + delegate_to: "{{ vm_install.host }}" + +- name: create disks for vm + with_dict: "{{ vm_install.disks.virtio | default({}) | combine(vm_install.disks.scsi | default({})) }}" + lvol: + vg: "{{ item.value.vg }}" + lv: "{{ item.value.lv }}" + size: "{{ item.value.size }}" + delegate_to: "{{ vm_install.host }}" + +- name: check if vm already exists + virt: + name: "{{ inventory_hostname }}" + command: info + delegate_to: "{{ vm_install.host }}" + register: vmhost_info + +- name: destroy exisiting vm + virt: + name: "{{ inventory_hostname }}" + state: destroyed + delegate_to: "{{ vm_install.host }}" + when: inventory_hostname in vmhost_info + +- name: wait for vm to be destroyed + wait_for_virt: + name: "{{ inventory_hostname }}" + states: shutdown,crashed + timeout: 5 + delegate_to: "{{ vm_install.host }}" + when: inventory_hostname in vmhost_info + +- name: undefining exisiting vm + virt: + name: "{{ inventory_hostname }}" + command: undefine + delegate_to: "{{ vm_install.host }}" + when: inventory_hostname in vmhost_info + +- name: enable installer in VM config + set_fact: + run_installer: True + +- name: define new installer vm + virt: + name: "{{ inventory_hostname }}" + command: define + xml: "{{ lookup('template', 'libvirt-domain.xml.j2') }}" + delegate_to: "{{ vm_install.host }}" + +- name: start vm + virt: + name: "{{ inventory_hostname }}" + state: running + delegate_to: "{{ vm_install.host }}" + +- name: wait for installer to start + wait_for_virt: + name: "{{ inventory_hostname }}" + states: running + timeout: 10 + delegate_to: "{{ vm_install.host }}" + +- debug: + msg: "you can check on the status of the installer running this command 'virsh console {{ inventory_hostname }}' on host {{ vm_install.host }}." + +- name: wait for installer to finish or crash + wait_for_virt: + name: "{{ inventory_hostname }}" + states: shutdown,crashed + timeout: 900 + delegate_to: "{{ vm_install.host }}" + register: installer_result + failed_when: installer_result.failed or installer_result.state == "crashed" + +- name: undefining installer vm + virt: + name: "{{ inventory_hostname }}" + command: undefine + delegate_to: "{{ vm_install.host }}" + +- name: disable installer in VM config + set_fact: + run_installer: False + +- name: define new production vm + virt: + name: "{{ inventory_hostname }}" + command: define + xml: "{{ lookup('template', 'libvirt-domain.xml.j2') }}" + delegate_to: "{{ vm_install.host }}" + +- name: start vm + virt: + name: "{{ inventory_hostname }}" + state: running + delegate_to: "{{ vm_install.host }}" + +- name: mark vm as autostarted + virt: + name: "{{ inventory_hostname }}" + autostart: "{{ vm_install.autostart }}" + command: info ## virt module needs either command or state + delegate_to: "{{ vm_install.host }}" + when: vm_install.autostart is defined + + ## TODO: find a better way to fetch host key of new VMs +- name: disable ssh StrictHostKeyChecking for the next step + set_fact: + ansible_ssh_extra_args: -o StrictHostKeyChecking=no + +- name: wait for vm to start up + wait_for_connection: + delay: 5 + timeout: 120 + +- name: reenable StrictHostKeyChecking + set_fact: + ansible_ssh_extra_args: "" diff --git a/ansible/roles/vm-install/templates/libvirt-domain.xml.j2 b/ansible/roles/vm-install/templates/libvirt-domain.xml.j2 new file mode 100644 index 0000000..c84b104 --- /dev/null +++ b/ansible/roles/vm-install/templates/libvirt-domain.xml.j2 @@ -0,0 +1,70 @@ + + {{ inventory_hostname }} + {{ vm_install.mem * 1024 }} + {{ vm_install.mem * 1024 }} + {{ vm_install.numcpu }} + + hvm +{% if run_installer %} + {{ hostvars[vm_install.host].vm_host.installer.path }}/{{ vmdistro }}-{{ vmdistcodename }}/{{ vm_install.arch | default('amd64') }}/linux + {{ hostvars[vm_install.host].vm_host.installer.path }}/{{ vmdistro }}-{{ vmdistcodename }}/{{ vm_install.arch | default('amd64') }}/initrd.gz + console=ttyS0,115200n8 auto=true interface=auto url=tftp://{{ hostvars[vm_install.host]['ansible_' + hostvars[vm_install.host].vm_host.installer.net_if].ipv4.address }}/vm-{{ inventory_hostname }}-{{ vmdistro }}-{{ vmdistcodename }}.cfg netcfg/choose_interface=enp1s1 netcfg/disable_autoconfig=true netcfg/get_ipaddress={{ vm_network.primary.ip }} netcfg/get_netmask={{ vm_network.primary.mask }} netcfg/get_gateway={{ vm_network.primary.gateway }} netcfg/get_nameservers="{{ vm_network.primary.nameservers | join(' ') }}" netcfg/confirm_static=true netcfg/get_hostname={{ inventory_hostname }} netcfg/get_domain={{ vm_network.primary.domain }} +{% endif %} + + + + + + + + + destroy +{% if run_installer %} + destroy + destroy +{% else %} + restart + restart +{% endif %} + + /usr/bin/kvm + +{% if 'virtio' in vm_install.disks %} +{% for device, lv in vm_install.disks.virtio.items() %} + + + + + +{% endfor %} +{% endif %} + +{% if 'scsi' in vm_install.disks %} + +{% for device, lv in vm_install.disks.scsi.items() %} + + + + + +{% endfor %} +{% endif %} + +{% if vm_install.interfaces %} +{% for if in vm_install.interfaces %} + + + +
+ +{% endfor %} +{% endif %} + + + + + + + + + diff --git a/ansible/roles/vm-install/templates/preseed_debian-stretch.cfg.j2 b/ansible/roles/vm-install/templates/preseed_debian-stretch.cfg.j2 new file mode 100644 index 0000000..5e9dc9f --- /dev/null +++ b/ansible/roles/vm-install/templates/preseed_debian-stretch.cfg.j2 @@ -0,0 +1,105 @@ +######################################################################### +# realraum preseed file for Debian stretch based VMs +######################################################################### + +d-i debian-installer/language string en +d-i debian-installer/country string AT +d-i debian-installer/locale string de_AT.UTF-8 +d-i keyboard-configuration/xkb-keymap select de + + +#d-i netcfg/choose_interface select enp1s1 +#d-i netcfg/disable_autoconfig boolean false +#d-i netcfg/get_ipaddress string {{ vm_network.primary.ip }} +#d-i netcfg/get_netmask string {{ vm_network.primary.mask }} +#d-i netcfg/get_gateway string {{ vm_network.primary.gateway }} +#d-i netcfg/get_nameservers string {{ vm_network.primary.nameservers | join(' ') }} +#d-i netcfg/confirm_static boolean true + +d-i netcfg/get_hostname string {{ inventory_hostname }} +d-i netcfg/get_domain string {{ vm_network.primary.domain }} +d-i netcfg/wireless_wep string + + +d-i mirror/country string manual +d-i mirror/http/hostname string debian.ffgraz.net +d-i mirror/http/directory string /debian +d-i mirror/http/proxy string + + +d-i passwd/make-user boolean false +d-i passwd/root-password password this-very-very-secure-password-will-be-removed-by-latecommand +d-i passwd/root-password-again password this-very-very-secure-password-will-be-removed-by-latecommand + + +d-i clock-setup/utc boolean true +d-i time/zone string Europe/Vienna +d-i clock-setup/ntp boolean false + + +d-i partman-auto/disk string /dev/{{ vm_install.disks.primary }} +d-i partman-auto/method string lvm +d-i partman-lvm/device_remove_lvm boolean true +d-i partman-md/device_remove_md boolean true + +d-i partman-lvm/confirm boolean true +d-i partman-lvm/confirm_nooverwrite boolean true + +d-i partman-auto/expert_recipe string \ + boot-root :: \ + 1000 10000 -1 ext4 \ + $defaultignore{ } $primary{ } $bootable{ } \ + method{ lvm } vg_name{ {{ inventory_hostname }} } \ + . \ + 2048 10000 2560 ext4 \ + $lvmok{ } in_vg{ {{ inventory_hostname }} } \ + method{ format } format{ } \ + use_filesystem{ } filesystem{ ext4 } \ + mountpoint{ / } \ + . \ + 1024 11000 1280 ext4 \ + $lvmok{ } in_vg{ {{ inventory_hostname }} } \ + method{ format } format{ } \ + use_filesystem{ } filesystem{ ext4 } \ + mountpoint{ /var } \ + . \ + 768 10000 768 ext4 \ + $lvmok{ } in_vg{ {{ inventory_hostname }} } \ + method{ format } format{ } \ + use_filesystem{ } filesystem{ ext4 } \ + mountpoint{ /var/log } \ + options/nodev{ nodev } options/noatime{ noatime } \ + options/noexec{ noexec } \ + . \ + 16 20000 -1 ext4 \ + $lvmok{ } in_vg{ {{ inventory_hostname }} } \ + method( keep } lv_name{ dummy } \ + . + +d-i partman-auto-lvm/no_boot boolean true +d-i partman-basicfilesystems/no_swap true +d-i partman-partitioning/confirm_write_new_label boolean true +d-i partman/choose_partition select finish +d-i partman/confirm boolean true +d-i partman/confirm_nooverwrite boolean true + + +d-i base-installer/install-recommends boolean false +d-i apt-setup/security_host string debian.ffgraz.net + +tasksel tasksel/first multiselect +d-i pkgsel/include string openssh-server python +d-i pkgsel/upgrade select safe-upgrade +popularity-contest popularity-contest/participate boolean false + +d-i grub-installer/choose_bootdev string /dev/{{ vm_install.disks.primary }} +d-i grub-installer/only_debian boolean true +d-i grub-installer/with_other_os boolean false + +d-i finish-install/reboot_in_progress note + + +d-i preseed/late_command string \ + lvremove -f {{ inventory_hostname }}/dummy; \ + in-target bash -c "apt-get update -q && apt-get full-upgrade -y -q"; \ + in-target bash -c "passwd -d root; passwd -l root; umask 077; mkdir -p /root/.ssh/; echo -e '{{ sshserver_root_keys }}' > /root/.ssh/authorized_keys" diff --git a/ansible/roles/vm-network/handlers/main.yml b/ansible/roles/vm-network/handlers/main.yml new file mode 100644 index 0000000..f967fa8 --- /dev/null +++ b/ansible/roles/vm-network/handlers/main.yml @@ -0,0 +1,3 @@ +--- +- name: rebuild initramfs + command: update-initramfs -u diff --git a/ansible/roles/vm-network/tasks/main.yml b/ansible/roles/vm-network/tasks/main.yml new file mode 100644 index 0000000..6668a4c --- /dev/null +++ b/ansible/roles/vm-network/tasks/main.yml @@ -0,0 +1,24 @@ +--- +- block: + - name: remove legacy systemd.link units + with_items: + - 50-virtio-kernel-names.link + - 99-default.link + file: + name: "/etc/systemd/network/{{ item }}" + state: absent + + - name: install systemd network link units + template: + src: systemd.link.j2 + dest: "/etc/systemd/network/{{ '%02d' | format(item.idx + 10) }}-{{ item.name }}.link" + with_items: "{{ vm_network.systemd_link.interfaces }}" + notify: rebuild initramfs + + when: vm_network.systemd_link is defined + +- name: install basic interface config + template: + src: interfaces.j2 + dest: /etc/network/interfaces + mode: 0644 diff --git a/ansible/roles/vm-network/templates/interfaces.j2 b/ansible/roles/vm-network/templates/interfaces.j2 new file mode 100644 index 0000000..9092a0b --- /dev/null +++ b/ansible/roles/vm-network/templates/interfaces.j2 @@ -0,0 +1,17 @@ +# This file describes the network interfaces available on your system +# and how to activate them. For more information, see interfaces(5). + +source /etc/network/interfaces.d/* + +# The loopback network interface +auto lo +iface lo inet loopback + +# The primary network interface +auto {{ srv_network.primary.interface }} +iface {{ srv_network.primary.interface }} inet static + address {{ srv_network.primary.ip }} + netmask {{ srv_network.primary.mask }} + gateway {{ srv_network.primary.gateway }} + pre-up echo 0 > /proc/sys/net/ipv6/conf/$IFACE/accept_ra + pre-up echo 0 > /proc/sys/net/ipv6/conf/$IFACE/autoconf diff --git a/ansible/roles/vm-network/templates/systemd.link.j2 b/ansible/roles/vm-network/templates/systemd.link.j2 new file mode 100644 index 0000000..753fd58 --- /dev/null +++ b/ansible/roles/vm-network/templates/systemd.link.j2 @@ -0,0 +1,5 @@ +[Match] +Path=pci-0000:01:{{ "%02d" | format(item.idx) }}.0 + +[Link] +Name={{ item.name }} diff --git a/ansible/vm-install.sh b/ansible/vm-install.sh new file mode 100755 index 0000000..428fdb5 --- /dev/null +++ b/ansible/vm-install.sh @@ -0,0 +1,16 @@ +#!/bin/bash + +if [ -z "$1" ] || [ -z "$2" ] || [ -z "$3" ]; then + echo "$0 " + exit 1 +fi + +echo "installing vm: $1 with $2/$3" +echo "" + +echo "########## clearing old ssh host keys #########" +./remove-known-host.sh "$1" +echo "" + +echo "######## running the install playbook ########" +exec ansible-playbook -e "vmname=$1" -e "vmdistro=$2" -e "vmdistcodename=$3" vm-install.yml diff --git a/ansible/vm-install.yml b/ansible/vm-install.yml new file mode 100644 index 0000000..e82aa33 --- /dev/null +++ b/ansible/vm-install.yml @@ -0,0 +1,21 @@ +--- +- name: Basic Installation + hosts: "{{ vmname }}" + gather_facts: no + pre_tasks: + - name: Gather facts of vm host + setup: + delegate_to: "{{ vm_install.host }}" + delegate_facts: yes + roles: + - role: vm-install + +- import_playbook: "host_playbooks/{{ vmname }}.yml" + +- name: Reboot and wait for VM come back + hosts: "{{ vmname }}" + gather_facts: no + roles: + - role: reboot-and-wait + reboot_delay: 10 + reboot_timeout: 120