From f8664e17ef5e6bead61d8471facd2859fd10c180 Mon Sep 17 00:00:00 2001 From: Jan Chaloupka Date: Tue, 19 Sep 2017 15:14:18 +0200 Subject: consolidate etcd_migrate role --- roles/etcd_migrate/tasks/add_ttls.yml | 33 ------------- roles/etcd_migrate/tasks/check.yml | 56 ----------------------- roles/etcd_migrate/tasks/check_cluster_health.yml | 23 ---------- roles/etcd_migrate/tasks/check_cluster_status.yml | 32 ------------- roles/etcd_migrate/tasks/clean_data.yml | 5 -- roles/etcd_migrate/tasks/configure.yml | 13 ------ roles/etcd_migrate/tasks/main.yml | 25 ---------- roles/etcd_migrate/tasks/migrate.yml | 56 ----------------------- 8 files changed, 243 deletions(-) delete mode 100644 roles/etcd_migrate/tasks/add_ttls.yml delete mode 100644 roles/etcd_migrate/tasks/check.yml delete mode 100644 roles/etcd_migrate/tasks/check_cluster_health.yml delete mode 100644 roles/etcd_migrate/tasks/check_cluster_status.yml delete mode 100644 roles/etcd_migrate/tasks/clean_data.yml delete mode 100644 roles/etcd_migrate/tasks/configure.yml delete mode 100644 roles/etcd_migrate/tasks/main.yml delete mode 100644 roles/etcd_migrate/tasks/migrate.yml (limited to 'roles/etcd_migrate/tasks') diff --git a/roles/etcd_migrate/tasks/add_ttls.yml b/roles/etcd_migrate/tasks/add_ttls.yml deleted file mode 100644 index c10465af9..000000000 --- a/roles/etcd_migrate/tasks/add_ttls.yml +++ /dev/null @@ -1,33 +0,0 @@ ---- -# To be executed on first master -- slurp: - src: "{{ openshift.common.config_base }}/master/master-config.yaml" - register: g_master_config_output - -- set_fact: - accessTokenMaxAgeSeconds: "{{ (g_master_config_output.content|b64decode|from_yaml).oauthConfig.tokenConfig.accessTokenMaxAgeSeconds | default(86400) }}" - authroizeTokenMaxAgeSeconds: "{{ (g_master_config_output.content|b64decode|from_yaml).oauthConfig.tokenConfig.authroizeTokenMaxAgeSeconds | default(500) }}" - controllerLeaseTTL: "{{ (g_master_config_output.content|b64decode|from_yaml).controllerLeaseTTL | default(30) }}" -- name: Re-introduce leases (as a replacement for key TTLs) - command: > - oadm migrate etcd-ttl \ - --cert {{ r_etcd_common_master_peer_cert_file }} \ - --key {{ r_etcd_common_master_peer_key_file }} \ - --cacert {{ r_etcd_common_master_peer_ca_file }} \ - --etcd-address 'https://{{ etcd_peer }}:{{ etcd_client_port }}' \ - --ttl-keys-prefix {{ item.keys }} \ - --lease-duration {{ item.ttl }} - environment: - ETCDCTL_API: 3 - PATH: "/usr/local/bin:/var/usrlocal/bin:{{ ansible_env.PATH }}" - with_items: - - keys: "/kubernetes.io/events" - ttl: "1h" - - keys: "/kubernetes.io/masterleases" - ttl: "10s" - - keys: "/openshift.io/oauth/accesstokens" - ttl: "{{ accessTokenMaxAgeSeconds }}s" - - keys: "/openshift.io/oauth/authorizetokens" - ttl: "{{ authroizeTokenMaxAgeSeconds }}s" - - keys: "/openshift.io/leases/controllers" - ttl: "{{ controllerLeaseTTL }}s" diff --git a/roles/etcd_migrate/tasks/check.yml b/roles/etcd_migrate/tasks/check.yml deleted file mode 100644 index 0804d9e1c..000000000 --- a/roles/etcd_migrate/tasks/check.yml +++ /dev/null @@ -1,56 +0,0 @@ ---- - -# Check the cluster is healthy -- include: check_cluster_health.yml - -# Check if the member has v3 data already -# Run the migration only if the data are v2 -- name: Check if there are any v3 data - command: > - etcdctl --cert {{ etcd_peer_cert_file }} --key {{ etcd_peer_key_file }} --cacert {{ etcd_peer_ca_file }} --endpoints 'https://{{ etcd_peer }}:{{ etcd_client_port }}' get "" --from-key --keys-only -w json --limit 1 - environment: - ETCDCTL_API: 3 - register: l_etcdctl_output - -- fail: - msg: "Unable to get a number of v3 keys" - when: l_etcdctl_output.rc != 0 - -- fail: - msg: "The etcd has at least one v3 key" - when: "'count' in (l_etcdctl_output.stdout | from_json) and (l_etcdctl_output.stdout | from_json).count != 0" - - -# TODO(jchaloup): once the until loop can be used over include/block, -# remove the repetive code -# - until loop not supported over include statement (nor block) -# https://github.com/ansible/ansible/issues/17098 -# - with_items not supported over block - -# Check the cluster status for the first time -- include: check_cluster_status.yml - -# Check the cluster status for the second time -- block: - - debug: - msg: "l_etcd_cluster_status_ok: {{ l_etcd_cluster_status_ok }}" - - name: Wait a while before another check - pause: - seconds: 5 - when: not l_etcd_cluster_status_ok | bool - - - include: check_cluster_status.yml - when: not l_etcd_cluster_status_ok | bool - - -# Check the cluster status for the third time -- block: - - debug: - msg: "l_etcd_cluster_status_ok: {{ l_etcd_cluster_status_ok }}" - - name: Wait a while before another check - pause: - seconds: 5 - when: not l_etcd_cluster_status_ok | bool - - - include: check_cluster_status.yml - when: not l_etcd_cluster_status_ok | bool diff --git a/roles/etcd_migrate/tasks/check_cluster_health.yml b/roles/etcd_migrate/tasks/check_cluster_health.yml deleted file mode 100644 index 201d83f99..000000000 --- a/roles/etcd_migrate/tasks/check_cluster_health.yml +++ /dev/null @@ -1,23 +0,0 @@ ---- -- name: Check cluster health - command: > - etcdctl --cert-file {{ etcd_peer_cert_file }} --key-file {{ etcd_peer_key_file }} --ca-file {{ etcd_peer_ca_file }} --endpoint https://{{ etcd_peer }}:{{ etcd_client_port }} cluster-health - register: etcd_cluster_health - changed_when: false - failed_when: false - -- name: Assume a member is not healthy - set_fact: - etcd_member_healthy: false - -- name: Get member item health status - set_fact: - etcd_member_healthy: true - with_items: "{{ etcd_cluster_health.stdout_lines }}" - when: "(etcd_peer in item) and ('is healthy' in item)" - -- name: Check the etcd cluster health - # TODO(jchaloup): should we fail or ask user if he wants to continue? Or just wait until the cluster is healthy? - fail: - msg: "Etcd member {{ etcd_peer }} is not healthy" - when: not etcd_member_healthy diff --git a/roles/etcd_migrate/tasks/check_cluster_status.yml b/roles/etcd_migrate/tasks/check_cluster_status.yml deleted file mode 100644 index b69fb5a52..000000000 --- a/roles/etcd_migrate/tasks/check_cluster_status.yml +++ /dev/null @@ -1,32 +0,0 @@ ---- -# etcd_ip originates from etcd_common role -- name: Check cluster status - command: > - etcdctl --cert {{ etcd_peer_cert_file }} --key {{ etcd_peer_key_file }} --cacert {{ etcd_peer_ca_file }} --endpoints 'https://{{ etcd_peer }}:{{ etcd_client_port }}' -w json endpoint status - environment: - ETCDCTL_API: 3 - register: l_etcd_cluster_status - -- name: Retrieve raftIndex - set_fact: - etcd_member_raft_index: "{{ (l_etcd_cluster_status.stdout | from_json)[0]['Status']['raftIndex'] }}" - -- block: - # http://docs.ansible.com/ansible/playbooks_filters.html#extracting-values-from-containers - - name: Group all raftIndices into a list - set_fact: - etcd_members_raft_indices: "{{ groups['oo_etcd_to_migrate'] | map('extract', hostvars, 'etcd_member_raft_index') | list | unique }}" - - - name: Check the minimum and the maximum of raftIndices is at most 1 - set_fact: - etcd_members_raft_indices_diff: "{{ ((etcd_members_raft_indices | max | int) - (etcd_members_raft_indices | min | int)) | int }}" - - - debug: - msg: "Raft indices difference: {{ etcd_members_raft_indices_diff }}" - - when: inventory_hostname in groups.oo_etcd_to_migrate[0] - -# The cluster raft status is ok if the difference of the max and min raft index is at most 1 -- name: capture the status - set_fact: - l_etcd_cluster_status_ok: "{{ hostvars[groups.oo_etcd_to_migrate[0]]['etcd_members_raft_indices_diff'] | int < 2 }}" diff --git a/roles/etcd_migrate/tasks/clean_data.yml b/roles/etcd_migrate/tasks/clean_data.yml deleted file mode 100644 index 95a0e7c0a..000000000 --- a/roles/etcd_migrate/tasks/clean_data.yml +++ /dev/null @@ -1,5 +0,0 @@ ---- -- name: Remove member data - file: - path: /var/lib/etcd/member - state: absent diff --git a/roles/etcd_migrate/tasks/configure.yml b/roles/etcd_migrate/tasks/configure.yml deleted file mode 100644 index a305d5bf3..000000000 --- a/roles/etcd_migrate/tasks/configure.yml +++ /dev/null @@ -1,13 +0,0 @@ ---- -- name: Configure master to use etcd3 storage backend - yedit: - src: /etc/origin/master/master-config.yaml - key: "{{ item.key }}" - value: "{{ item.value }}" - with_items: - - key: kubernetesMasterConfig.apiServerArguments.storage-backend - value: - - etcd3 - - key: kubernetesMasterConfig.apiServerArguments.storage-media-type - value: - - application/vnd.kubernetes.protobuf diff --git a/roles/etcd_migrate/tasks/main.yml b/roles/etcd_migrate/tasks/main.yml deleted file mode 100644 index e82f6a6b4..000000000 --- a/roles/etcd_migrate/tasks/main.yml +++ /dev/null @@ -1,25 +0,0 @@ ---- -- name: Fail if invalid r_etcd_migrate_action provided - fail: - msg: "etcd_migrate role can only be called with 'check', 'migrate', 'configure', 'add_ttls', or 'clean_data'" - when: r_etcd_migrate_action not in ['check', 'migrate', 'configure', 'add_ttls', 'clean_data'] - -- name: Include main action task file - include: "{{ r_etcd_migrate_action }}.yml" - -# 2. migrate v2 datadir into v3: -# ETCDCTL_API=3 ./etcdctl migrate --data-dir=${data_dir} --no-ttl -# backup the etcd datadir first -# Provide a way for an operator to specify transformer - -# 3. re-configure OpenShift master at /etc/origin/master/master-config.yml -# set storage-backend to “etcd3” -# 4. we could leave the master restart to current logic (there is already the code ready (single vs. HA master)) - -# Run -# etcdctl --cert-file /etc/etcd/peer.crt --key-file /etc/etcd/peer.key --ca-file /etc/etcd/ca.crt --endpoint https://172.16.186.45:2379 cluster-health -# to check the cluster health (from the etcdctl.sh aliases file) - -# Another assumption: -# - in order to migrate all etcd v2 data into v3, we need to shut down the cluster (let's verify that on Wednesday meeting) -# - diff --git a/roles/etcd_migrate/tasks/migrate.yml b/roles/etcd_migrate/tasks/migrate.yml deleted file mode 100644 index 54a9c74ff..000000000 --- a/roles/etcd_migrate/tasks/migrate.yml +++ /dev/null @@ -1,56 +0,0 @@ ---- -# Should this be run in a serial manner? -- set_fact: - l_etcd_service: "{{ 'etcd_container' if openshift.common.is_containerized else 'etcd' }}" - -- name: Migrate etcd data - command: > - etcdctl migrate --data-dir={{ etcd_data_dir }} - environment: - ETCDCTL_API: 3 - register: l_etcdctl_migrate -# TODO(jchaloup): If any of the members fails, we need to restore all members to v2 from the pre-migrate backup -- name: Check the etcd v2 data are correctly migrated - fail: - msg: "Failed to migrate a member" - when: "'finished transforming keys' not in l_etcdctl_migrate.stdout and 'no v2 keys to migrate' not in l_etcdctl_migrate.stdout" -- name: Migration message - debug: - msg: "Etcd migration finished with: {{ l_etcdctl_migrate.stdout }}" -- name: Set ETCD_FORCE_NEW_CLUSTER=true on first etcd host - lineinfile: - line: "ETCD_FORCE_NEW_CLUSTER=true" - dest: /etc/etcd/etcd.conf - backup: true -- name: Start etcd - systemd: - name: "{{ l_etcd_service }}" - state: started -- name: Wait for cluster to become healthy after bringing up first member - command: > - etcdctl --cert-file {{ etcd_peer_cert_file }} --key-file {{ etcd_peer_key_file }} --ca-file {{ etcd_peer_ca_file }} --endpoint https://{{ etcd_peer }}:{{ etcd_client_port }} cluster-health - register: l_etcd_migrate_health - until: l_etcd_migrate_health.rc == 0 - retries: 3 - delay: 30 -- name: Unset ETCD_FORCE_NEW_CLUSTER=true on first etcd host - lineinfile: - line: "ETCD_FORCE_NEW_CLUSTER=true" - dest: /etc/etcd/etcd.conf - state: absent - backup: true -- name: Restart first etcd host - systemd: - name: "{{ l_etcd_service }}" - state: restarted - -- name: Wait for cluster to become healthy after bringing up first member - command: > - etcdctl --cert-file {{ etcd_peer_cert_file }} --key-file {{ etcd_peer_key_file }} --ca-file {{ etcd_peer_ca_file }} --endpoint https://{{ etcd_peer }}:{{ etcd_client_port }} cluster-health - register: l_etcd_migrate_health - until: l_etcd_migrate_health.rc == 0 - retries: 3 - delay: 30 - -- set_fact: - r_etcd_migrate_success: true -- cgit v1.2.3