From b620bc7810b4ba78e2805b3576147fa9d4b914fe Mon Sep 17 00:00:00 2001 From: Tristan Cacqueray Date: Thu, 8 Nov 2018 04:05:12 +0000 Subject: [PATCH] Remove nodepool-k8s-functional and install-nodepool roles Those roles aren't actually used. Change-Id: I778df2318d43cd3c8b7ed78ba7a1a9808d610fca --- roles/install-nodepool/tasks/main.yaml | 88 ------------------- roles/nodepool-k8s-functional/tasks/main.yaml | 59 ------------- 2 files changed, 147 deletions(-) delete mode 100644 roles/install-nodepool/tasks/main.yaml delete mode 100644 roles/nodepool-k8s-functional/tasks/main.yaml diff --git a/roles/install-nodepool/tasks/main.yaml b/roles/install-nodepool/tasks/main.yaml deleted file mode 100644 index a54a224db..000000000 --- a/roles/install-nodepool/tasks/main.yaml +++ /dev/null @@ -1,88 +0,0 @@ -- name: Install zookeeper - package: - name: zookeeperd - become: yes - -- name: Start zookeeper - service: - name: zookeeper - state: started - become: yes - -- name: Install nodepool - command: pip3 install . - args: - chdir: "{{ zuul.projects['git.openstack.org/openstack-infra/nodepool'].src_dir }}" - become: yes - -- name: Setup logging.conf - copy: - content: | - [loggers] - keys=root,nodepool,requests,openstack - - [handlers] - keys=console,normal - - [formatters] - keys=simple - - [logger_root] - level=WARNING - handlers=console - - [logger_requests] - level=WARNING - handlers=normal - qualname=requests - - [logger_openstack] - level=WARNING - handlers=normal - qualname=openstack - - [logger_gear] - level=DEBUG - handlers=normal - qualname=gear - - [logger_nodepool] - level=DEBUG - handlers=normal - qualname=nodepool - - [handler_console] - level=WARNING - class=StreamHandler - formatter=simple - args=(sys.stdout,) - - [handler_normal] - level=DEBUG - class=FileHandler - formatter=simple - args=('{{ ansible_user_dir }}/work/logs/nodepool/launcher.log',) - - [formatter_simple] - format=%(asctime)s %(levelname)s %(name)s: %(message)s - datefmt= - dest: "{{ ansible_user_dir }}/work/etc/logging.conf" - -- name: Setup nodepool.yaml - copy: - content: | - zookeeper-servers: - - host: localhost - images-dir: "{{ ansible_user_dir }}/work/images/" - build-log-dir: "{{ ansible_user_dir }}/work/logs/nodepool/" - dest: "{{ ansible_user_dir }}/work/etc/nodepool.yaml" - -- name: Setup secure.conf - copy: - content: "" - dest: "{{ ansible_user_dir }}/work/etc/secure.conf" - -- name: Start the service - command: nodepool-launcher -c etc/nodepool.yaml -s etc/secure.conf -l etc/logging.conf -p launcher.pid - args: - chdir: "{{ ansible_user_dir }}/work/" diff --git a/roles/nodepool-k8s-functional/tasks/main.yaml b/roles/nodepool-k8s-functional/tasks/main.yaml deleted file mode 100644 index cfcb1b80b..000000000 --- a/roles/nodepool-k8s-functional/tasks/main.yaml +++ /dev/null @@ -1,59 +0,0 @@ -- name: debug context names - command: kubectl config get-contexts - -- name: Update nodepool.yaml - copy: - content: | - zookeeper-servers: - - host: localhost - images-dir: "{{ ansible_user_dir }}/work/images/" - build-log-dir: "{{ ansible_user_dir }}/work/logs/nodepool/" - labels: - - name: kubernetes-namespace - min-ready: 1 - - name: pod-fedora - min-ready: 1 - providers: - - name: minikube - driver: kubernetes - context: minikube - pools: - - name: main - labels: - - name: kubernetes-namespace - type: namespace - - name: pod-fedora - type: pod - image: docker.io/fedora:28 - dest: "{{ ansible_user_dir }}/work/etc/nodepool.yaml" - -- name: Set nodepool_command facts - set_fact: - nodepool_command: nodepool -c "{{ ansible_user_dir }}/work/etc/nodepool.yaml" - -- name: Wait for nodes - command: "{{ nodepool_command }} list" - register: nodepool_list - until: nodepool_list.stdout - retries: 120 - delay: 2 - -- name: Show nodes - command: "{{ nodepool_command }} list --detail" - -- name: Wait for fedora pod to be running - shell: "{{ nodepool_command }} list | grep 'pod-fedora.*running'" - register: nodepool_list - until: nodepool_list.stdout - retries: 120 - delay: 2 - ignore_errors: yes - -- name: Show nodes - command: "{{ nodepool_command }} list --detail" - -- name: Show namespace - command: kubectl get namespaces - -- name: Show pods - command: kubectl get --all-namespaces=true pods