2025-08-29 16:46:51.892132 | Job console starting 2025-08-29 16:46:51.904872 | Updating git repos 2025-08-29 16:46:51.991362 | Cloning repos into workspace 2025-08-29 16:46:52.193212 | Restoring repo states 2025-08-29 16:46:52.211514 | Merging changes 2025-08-29 16:46:52.211534 | Checking out repos 2025-08-29 16:46:52.464334 | Preparing playbooks 2025-08-29 16:46:53.066978 | Running Ansible setup 2025-08-29 16:46:57.238814 | PRE-RUN START: [trusted : github.com/osism/zuul-config/playbooks/base/pre.yaml@main] 2025-08-29 16:46:58.032086 | 2025-08-29 16:46:58.032263 | PLAY [Base pre] 2025-08-29 16:46:58.050359 | 2025-08-29 16:46:58.050501 | TASK [Setup log path fact] 2025-08-29 16:46:58.081886 | orchestrator | ok 2025-08-29 16:46:58.100633 | 2025-08-29 16:46:58.100774 | TASK [set-zuul-log-path-fact : Set log path for a build] 2025-08-29 16:46:58.149713 | orchestrator | ok 2025-08-29 16:46:58.166225 | 2025-08-29 16:46:58.167198 | TASK [emit-job-header : Print job information] 2025-08-29 16:46:58.219680 | # Job Information 2025-08-29 16:46:58.220063 | Ansible Version: 2.16.14 2025-08-29 16:46:58.220129 | Job: testbed-deploy-stable-in-a-nutshell-ubuntu-24.04 2025-08-29 16:46:58.220190 | Pipeline: post 2025-08-29 16:46:58.220231 | Executor: 521e9411259a 2025-08-29 16:46:58.220268 | Triggered by: https://github.com/osism/testbed/commit/ef698e79bb91e5c1a90863c5f63ed3556fcd4722 2025-08-29 16:46:58.220312 | Event ID: 7e34c700-84e1-11f0-9853-a2f38d9dedb1 2025-08-29 16:46:58.230735 | 2025-08-29 16:46:58.230913 | LOOP [emit-job-header : Print node information] 2025-08-29 16:46:58.351659 | orchestrator | ok: 2025-08-29 16:46:58.351905 | orchestrator | # Node Information 2025-08-29 16:46:58.351939 | orchestrator | Inventory Hostname: orchestrator 2025-08-29 16:46:58.351963 | orchestrator | Hostname: zuul-static-regiocloud-infra-1 2025-08-29 16:46:58.351986 | orchestrator | Username: zuul-testbed04 2025-08-29 16:46:58.352007 | orchestrator | Distro: Debian 12.11 2025-08-29 16:46:58.352033 | orchestrator | Provider: static-testbed 2025-08-29 16:46:58.352054 | orchestrator | Region: 2025-08-29 16:46:58.352075 | orchestrator | Label: testbed-orchestrator 2025-08-29 16:46:58.352096 | orchestrator | Product Name: OpenStack Nova 2025-08-29 16:46:58.352115 | orchestrator | Interface IP: 81.163.193.140 2025-08-29 16:46:58.372133 | 2025-08-29 16:46:58.372293 | TASK [log-inventory : Ensure Zuul Ansible directory exists] 2025-08-29 16:46:58.894542 | orchestrator -> localhost | changed 2025-08-29 16:46:58.910875 | 2025-08-29 16:46:58.911034 | TASK [log-inventory : Copy ansible inventory to logs dir] 2025-08-29 16:47:00.012507 | orchestrator -> localhost | changed 2025-08-29 16:47:00.036212 | 2025-08-29 16:47:00.036364 | TASK [add-build-sshkey : Check to see if ssh key was already created for this build] 2025-08-29 16:47:00.355275 | orchestrator -> localhost | ok 2025-08-29 16:47:00.371783 | 2025-08-29 16:47:00.371974 | TASK [add-build-sshkey : Create a new key in workspace based on build UUID] 2025-08-29 16:47:00.411530 | orchestrator | ok 2025-08-29 16:47:00.432322 | orchestrator | included: /var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey/tasks/create-key-and-replace.yaml 2025-08-29 16:47:00.440744 | 2025-08-29 16:47:00.440850 | TASK [add-build-sshkey : Create Temp SSH key] 2025-08-29 16:47:02.956386 | orchestrator -> localhost | Generating public/private rsa key pair. 2025-08-29 16:47:02.956930 | orchestrator -> localhost | Your identification has been saved in /var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/work/931a11d02915441d8037f098b05d6016_id_rsa 2025-08-29 16:47:02.957046 | orchestrator -> localhost | Your public key has been saved in /var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/work/931a11d02915441d8037f098b05d6016_id_rsa.pub 2025-08-29 16:47:02.957123 | orchestrator -> localhost | The key fingerprint is: 2025-08-29 16:47:02.957199 | orchestrator -> localhost | SHA256:of20j7/Q8235ao0MZp8kTyHL8DxamP4fs/bu3c+3z+A zuul-build-sshkey 2025-08-29 16:47:02.957265 | orchestrator -> localhost | The key's randomart image is: 2025-08-29 16:47:02.957349 | orchestrator -> localhost | +---[RSA 3072]----+ 2025-08-29 16:47:02.957412 | orchestrator -> localhost | | | 2025-08-29 16:47:02.957473 | orchestrator -> localhost | | | 2025-08-29 16:47:02.957531 | orchestrator -> localhost | | . | 2025-08-29 16:47:02.957625 | orchestrator -> localhost | | o .. . . | 2025-08-29 16:47:02.957687 | orchestrator -> localhost | | . S .B o . | 2025-08-29 16:47:02.957754 | orchestrator -> localhost | | ooo% o | 2025-08-29 16:47:02.957812 | orchestrator -> localhost | | .+=oX++.| 2025-08-29 16:47:02.957867 | orchestrator -> localhost | | o+ +OX*| 2025-08-29 16:47:02.957925 | orchestrator -> localhost | | .o==EB^| 2025-08-29 16:47:02.957982 | orchestrator -> localhost | +----[SHA256]-----+ 2025-08-29 16:47:02.958175 | orchestrator -> localhost | ok: Runtime: 0:00:01.978174 2025-08-29 16:47:02.973407 | 2025-08-29 16:47:02.973554 | TASK [add-build-sshkey : Remote setup ssh keys (linux)] 2025-08-29 16:47:03.009730 | orchestrator | ok 2025-08-29 16:47:03.022921 | orchestrator | included: /var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey/tasks/remote-linux.yaml 2025-08-29 16:47:03.032672 | 2025-08-29 16:47:03.032773 | TASK [add-build-sshkey : Remove previously added zuul-build-sshkey] 2025-08-29 16:47:03.057286 | orchestrator | skipping: Conditional result was False 2025-08-29 16:47:03.067386 | 2025-08-29 16:47:03.067513 | TASK [add-build-sshkey : Enable access via build key on all nodes] 2025-08-29 16:47:03.743645 | orchestrator | changed 2025-08-29 16:47:03.753133 | 2025-08-29 16:47:03.753266 | TASK [add-build-sshkey : Make sure user has a .ssh] 2025-08-29 16:47:04.046090 | orchestrator | ok 2025-08-29 16:47:04.056052 | 2025-08-29 16:47:04.056211 | TASK [add-build-sshkey : Install build private key as SSH key on all nodes] 2025-08-29 16:47:04.494411 | orchestrator | ok 2025-08-29 16:47:04.504028 | 2025-08-29 16:47:04.504160 | TASK [add-build-sshkey : Install build public key as SSH key on all nodes] 2025-08-29 16:47:04.924905 | orchestrator | ok 2025-08-29 16:47:04.933978 | 2025-08-29 16:47:04.934109 | TASK [add-build-sshkey : Remote setup ssh keys (windows)] 2025-08-29 16:47:04.969022 | orchestrator | skipping: Conditional result was False 2025-08-29 16:47:04.978695 | 2025-08-29 16:47:04.978854 | TASK [remove-zuul-sshkey : Remove master key from local agent] 2025-08-29 16:47:05.437402 | orchestrator -> localhost | changed 2025-08-29 16:47:05.452352 | 2025-08-29 16:47:05.452485 | TASK [add-build-sshkey : Add back temp key] 2025-08-29 16:47:05.810459 | orchestrator -> localhost | Identity added: /var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/work/931a11d02915441d8037f098b05d6016_id_rsa (zuul-build-sshkey) 2025-08-29 16:47:05.810781 | orchestrator -> localhost | ok: Runtime: 0:00:00.021814 2025-08-29 16:47:05.818691 | 2025-08-29 16:47:05.818801 | TASK [add-build-sshkey : Verify we can still SSH to all nodes] 2025-08-29 16:47:06.283685 | orchestrator | ok 2025-08-29 16:47:06.292662 | 2025-08-29 16:47:06.292796 | TASK [add-build-sshkey : Verify we can still SSH to all nodes (windows)] 2025-08-29 16:47:06.327867 | orchestrator | skipping: Conditional result was False 2025-08-29 16:47:06.388046 | 2025-08-29 16:47:06.388176 | TASK [start-zuul-console : Start zuul_console daemon.] 2025-08-29 16:47:06.790783 | orchestrator | ok 2025-08-29 16:47:06.818383 | 2025-08-29 16:47:06.818616 | TASK [validate-host : Define zuul_info_dir fact] 2025-08-29 16:47:06.868366 | orchestrator | ok 2025-08-29 16:47:06.879704 | 2025-08-29 16:47:06.879836 | TASK [validate-host : Ensure Zuul Ansible directory exists] 2025-08-29 16:47:07.187968 | orchestrator -> localhost | ok 2025-08-29 16:47:07.199057 | 2025-08-29 16:47:07.199190 | TASK [validate-host : Collect information about the host] 2025-08-29 16:47:08.411158 | orchestrator | ok 2025-08-29 16:47:08.427042 | 2025-08-29 16:47:08.427156 | TASK [validate-host : Sanitize hostname] 2025-08-29 16:47:08.504798 | orchestrator | ok 2025-08-29 16:47:08.513732 | 2025-08-29 16:47:08.513876 | TASK [validate-host : Write out all ansible variables/facts known for each host] 2025-08-29 16:47:09.086666 | orchestrator -> localhost | changed 2025-08-29 16:47:09.094588 | 2025-08-29 16:47:09.094728 | TASK [validate-host : Collect information about zuul worker] 2025-08-29 16:47:09.549618 | orchestrator | ok 2025-08-29 16:47:09.561182 | 2025-08-29 16:47:09.561348 | TASK [validate-host : Write out all zuul information for each host] 2025-08-29 16:47:10.134378 | orchestrator -> localhost | changed 2025-08-29 16:47:10.146244 | 2025-08-29 16:47:10.146357 | TASK [prepare-workspace-log : Start zuul_console daemon.] 2025-08-29 16:47:10.435111 | orchestrator | ok 2025-08-29 16:47:10.443758 | 2025-08-29 16:47:10.443878 | TASK [prepare-workspace-log : Synchronize src repos to workspace directory.] 2025-08-29 16:47:27.592215 | orchestrator | changed: 2025-08-29 16:47:27.592558 | orchestrator | .d..t...... src/ 2025-08-29 16:47:27.592639 | orchestrator | .d..t...... src/github.com/ 2025-08-29 16:47:27.592681 | orchestrator | .d..t...... src/github.com/osism/ 2025-08-29 16:47:27.592718 | orchestrator | .d..t...... src/github.com/osism/ansible-collection-commons/ 2025-08-29 16:47:27.592752 | orchestrator | RedHat.yml 2025-08-29 16:47:27.609461 | orchestrator | .L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml 2025-08-29 16:47:27.609479 | orchestrator | RedHat.yml 2025-08-29 16:47:27.609530 | orchestrator | = 2.2.0"... 2025-08-29 16:47:44.132027 | orchestrator | 16:47:44.131 STDOUT terraform: - Finding latest version of hashicorp/null... 2025-08-29 16:47:44.170004 | orchestrator | 16:47:44.169 STDOUT terraform: - Finding terraform-provider-openstack/openstack versions matching ">= 1.53.0"... 2025-08-29 16:47:44.652710 | orchestrator | 16:47:44.652 STDOUT terraform: - Installing hashicorp/local v2.5.3... 2025-08-29 16:47:45.262403 | orchestrator | 16:47:45.262 STDOUT terraform: - Installed hashicorp/local v2.5.3 (signed, key ID 0C0AF313E5FD9F80) 2025-08-29 16:47:45.342290 | orchestrator | 16:47:45.342 STDOUT terraform: - Installing hashicorp/null v3.2.4... 2025-08-29 16:47:45.853541 | orchestrator | 16:47:45.853 STDOUT terraform: - Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80) 2025-08-29 16:47:46.274420 | orchestrator | 16:47:46.274 STDOUT terraform: - Installing terraform-provider-openstack/openstack v3.3.2... 2025-08-29 16:47:47.160147 | orchestrator | 16:47:47.159 STDOUT terraform: - Installed terraform-provider-openstack/openstack v3.3.2 (signed, key ID 4F80527A391BEFD2) 2025-08-29 16:47:47.160239 | orchestrator | 16:47:47.159 STDOUT terraform: Providers are signed by their developers. 2025-08-29 16:47:47.160261 | orchestrator | 16:47:47.159 STDOUT terraform: If you'd like to know more about provider signing, you can read about it here: 2025-08-29 16:47:47.160278 | orchestrator | 16:47:47.159 STDOUT terraform: https://opentofu.org/docs/cli/plugins/signing/ 2025-08-29 16:47:47.160295 | orchestrator | 16:47:47.159 STDOUT terraform: OpenTofu has created a lock file .terraform.lock.hcl to record the provider 2025-08-29 16:47:47.160325 | orchestrator | 16:47:47.159 STDOUT terraform: selections it made above. Include this file in your version control repository 2025-08-29 16:47:47.160359 | orchestrator | 16:47:47.160 STDOUT terraform: so that OpenTofu can guarantee to make the same selections by default when 2025-08-29 16:47:47.160378 | orchestrator | 16:47:47.160 STDOUT terraform: you run "tofu init" in the future. 2025-08-29 16:47:47.161318 | orchestrator | 16:47:47.160 STDOUT terraform: OpenTofu has been successfully initialized! 2025-08-29 16:47:47.161373 | orchestrator | 16:47:47.161 STDOUT terraform: You may now begin working with OpenTofu. Try running "tofu plan" to see 2025-08-29 16:47:47.161390 | orchestrator | 16:47:47.161 STDOUT terraform: any changes that are required for your infrastructure. All OpenTofu commands 2025-08-29 16:47:47.161404 | orchestrator | 16:47:47.161 STDOUT terraform: should now work. 2025-08-29 16:47:47.161417 | orchestrator | 16:47:47.161 STDOUT terraform: If you ever set or change modules or backend configuration for OpenTofu, 2025-08-29 16:47:47.161436 | orchestrator | 16:47:47.161 STDOUT terraform: rerun this command to reinitialize your working directory. If you forget, other 2025-08-29 16:47:47.161525 | orchestrator | 16:47:47.161 STDOUT terraform: commands will detect it and remind you to do so if necessary. 2025-08-29 16:47:47.258482 | orchestrator | 16:47:47.258 WARN  The `TERRAGRUNT_TFPATH` environment variable is deprecated and will be removed in a future version of Terragrunt. Use `TG_TF_PATH=/home/zuul-testbed04/terraform` instead. 2025-08-29 16:47:47.258659 | orchestrator | 16:47:47.258 WARN  The `workspace` command is deprecated and will be removed in a future version of Terragrunt. Use `terragrunt run -- workspace` instead. 2025-08-29 16:47:47.466375 | orchestrator | 16:47:47.466 STDOUT terraform: Created and switched to workspace "ci"! 2025-08-29 16:47:47.466428 | orchestrator | 16:47:47.466 STDOUT terraform: You're now on a new, empty workspace. Workspaces isolate their state, 2025-08-29 16:47:47.466436 | orchestrator | 16:47:47.466 STDOUT terraform: so if you run "tofu plan" OpenTofu will not see any existing state 2025-08-29 16:47:47.466443 | orchestrator | 16:47:47.466 STDOUT terraform: for this configuration. 2025-08-29 16:47:47.610675 | orchestrator | 16:47:47.608 WARN  The `TERRAGRUNT_TFPATH` environment variable is deprecated and will be removed in a future version of Terragrunt. Use `TG_TF_PATH=/home/zuul-testbed04/terraform` instead. 2025-08-29 16:47:47.610737 | orchestrator | 16:47:47.609 WARN  The `fmt` command is deprecated and will be removed in a future version of Terragrunt. Use `terragrunt run -- fmt` instead. 2025-08-29 16:47:47.714312 | orchestrator | 16:47:47.714 STDOUT terraform: ci.auto.tfvars 2025-08-29 16:47:47.716697 | orchestrator | 16:47:47.716 STDOUT terraform: default_custom.tf 2025-08-29 16:47:47.819325 | orchestrator | 16:47:47.819 WARN  The `TERRAGRUNT_TFPATH` environment variable is deprecated and will be removed in a future version of Terragrunt. Use `TG_TF_PATH=/home/zuul-testbed04/terraform` instead. 2025-08-29 16:47:48.693018 | orchestrator | 16:47:48.692 STDOUT terraform: data.openstack_networking_network_v2.public: Reading... 2025-08-29 16:47:49.187505 | orchestrator | 16:47:49.187 STDOUT terraform: data.openstack_networking_network_v2.public: Read complete after 0s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a] 2025-08-29 16:47:49.397365 | orchestrator | 16:47:49.397 STDOUT terraform: OpenTofu used the selected providers to generate the following execution 2025-08-29 16:47:49.397419 | orchestrator | 16:47:49.397 STDOUT terraform: plan. Resource actions are indicated with the following symbols: 2025-08-29 16:47:49.397427 | orchestrator | 16:47:49.397 STDOUT terraform:  + create 2025-08-29 16:47:49.397476 | orchestrator | 16:47:49.397 STDOUT terraform:  <= read (data resources) 2025-08-29 16:47:49.397508 | orchestrator | 16:47:49.397 STDOUT terraform: OpenTofu will perform the following actions: 2025-08-29 16:47:49.397732 | orchestrator | 16:47:49.397 STDOUT terraform:  # data.openstack_images_image_v2.image will be read during apply 2025-08-29 16:47:49.397774 | orchestrator | 16:47:49.397 STDOUT terraform:  # (config refers to values not yet known) 2025-08-29 16:47:49.397906 | orchestrator | 16:47:49.397 STDOUT terraform:  <= data "openstack_images_image_v2" "image" { 2025-08-29 16:47:49.397916 | orchestrator | 16:47:49.397 STDOUT terraform:  + checksum = (known after apply) 2025-08-29 16:47:49.397967 | orchestrator | 16:47:49.397 STDOUT terraform:  + created_at = (known after apply) 2025-08-29 16:47:49.397977 | orchestrator | 16:47:49.397 STDOUT terraform:  + file = (known after apply) 2025-08-29 16:47:49.398046 | orchestrator | 16:47:49.397 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.398058 | orchestrator | 16:47:49.398 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.398120 | orchestrator | 16:47:49.398 STDOUT terraform:  + min_disk_gb = (known after apply) 2025-08-29 16:47:49.398126 | orchestrator | 16:47:49.398 STDOUT terraform:  + min_ram_mb = (known after apply) 2025-08-29 16:47:49.398162 | orchestrator | 16:47:49.398 STDOUT terraform:  + most_recent = true 2025-08-29 16:47:49.398169 | orchestrator | 16:47:49.398 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.398208 | orchestrator | 16:47:49.398 STDOUT terraform:  + protected = (known after apply) 2025-08-29 16:47:49.398253 | orchestrator | 16:47:49.398 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.398262 | orchestrator | 16:47:49.398 STDOUT terraform:  + schema = (known after apply) 2025-08-29 16:47:49.398304 | orchestrator | 16:47:49.398 STDOUT terraform:  + size_bytes = (known after apply) 2025-08-29 16:47:49.398338 | orchestrator | 16:47:49.398 STDOUT terraform:  + tags = (known after apply) 2025-08-29 16:47:49.398348 | orchestrator | 16:47:49.398 STDOUT terraform:  + updated_at = (known after apply) 2025-08-29 16:47:49.398364 | orchestrator | 16:47:49.398 STDOUT terraform:  } 2025-08-29 16:47:49.398577 | orchestrator | 16:47:49.398 STDOUT terraform:  # data.openstack_images_image_v2.image_node will be read during apply 2025-08-29 16:47:49.398625 | orchestrator | 16:47:49.398 STDOUT terraform:  # (config refers to values not yet known) 2025-08-29 16:47:49.398700 | orchestrator | 16:47:49.398 STDOUT terraform:  <= data "openstack_images_image_v2" "image_node" { 2025-08-29 16:47:49.398709 | orchestrator | 16:47:49.398 STDOUT terraform:  + checksum = (known after apply) 2025-08-29 16:47:49.398714 | orchestrator | 16:47:49.398 STDOUT terraform:  + created_at = (known after apply) 2025-08-29 16:47:49.398740 | orchestrator | 16:47:49.398 STDOUT terraform:  + file = (known after apply) 2025-08-29 16:47:49.398815 | orchestrator | 16:47:49.398 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.398820 | orchestrator | 16:47:49.398 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.398826 | orchestrator | 16:47:49.398 STDOUT terraform:  + min_disk_gb = (known after apply) 2025-08-29 16:47:49.398896 | orchestrator | 16:47:49.398 STDOUT terraform:  + min_ram_mb = (known after apply) 2025-08-29 16:47:49.398907 | orchestrator | 16:47:49.398 STDOUT terraform:  + most_recent = true 2025-08-29 16:47:49.398980 | orchestrator | 16:47:49.398 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.399026 | orchestrator | 16:47:49.398 STDOUT terraform:  + protected = (known after apply) 2025-08-29 16:47:49.399058 | orchestrator | 16:47:49.399 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.399067 | orchestrator | 16:47:49.399 STDOUT terraform:  + schema = (known after apply) 2025-08-29 16:47:49.399113 | orchestrator | 16:47:49.399 STDOUT terraform:  + size_bytes = (known after apply) 2025-08-29 16:47:49.399136 | orchestrator | 16:47:49.399 STDOUT terraform:  + tags = (known after apply) 2025-08-29 16:47:49.399176 | orchestrator | 16:47:49.399 STDOUT terraform:  + updated_at = (known after apply) 2025-08-29 16:47:49.399181 | orchestrator | 16:47:49.399 STDOUT terraform:  } 2025-08-29 16:47:49.399337 | orchestrator | 16:47:49.399 STDOUT terraform:  # local_file.MANAGER_ADDRESS will be created 2025-08-29 16:47:49.399387 | orchestrator | 16:47:49.399 STDOUT terraform:  + resource "local_file" "MANAGER_ADDRESS" { 2025-08-29 16:47:49.399434 | orchestrator | 16:47:49.399 STDOUT terraform:  + content = (known after apply) 2025-08-29 16:47:49.399444 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_base64sha256 = (known after apply) 2025-08-29 16:47:49.399481 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_base64sha512 = (known after apply) 2025-08-29 16:47:49.399515 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_md5 = (known after apply) 2025-08-29 16:47:49.399548 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_sha1 = (known after apply) 2025-08-29 16:47:49.399582 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_sha256 = (known after apply) 2025-08-29 16:47:49.399629 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_sha512 = (known after apply) 2025-08-29 16:47:49.399664 | orchestrator | 16:47:49.399 STDOUT terraform:  + directory_permission = "0777" 2025-08-29 16:47:49.399671 | orchestrator | 16:47:49.399 STDOUT terraform:  + file_permission = "0644" 2025-08-29 16:47:49.399707 | orchestrator | 16:47:49.399 STDOUT terraform:  + filename = ".MANAGER_ADDRESS.ci" 2025-08-29 16:47:49.399760 | orchestrator | 16:47:49.399 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.399765 | orchestrator | 16:47:49.399 STDOUT terraform:  } 2025-08-29 16:47:49.399880 | orchestrator | 16:47:49.399 STDOUT terraform:  # local_file.id_rsa_pub will be created 2025-08-29 16:47:49.399887 | orchestrator | 16:47:49.399 STDOUT terraform:  + resource "local_file" "id_rsa_pub" { 2025-08-29 16:47:49.399932 | orchestrator | 16:47:49.399 STDOUT terraform:  + content = (known after apply) 2025-08-29 16:47:49.399980 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_base64sha256 = (known after apply) 2025-08-29 16:47:49.400007 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_base64sha512 = (known after apply) 2025-08-29 16:47:49.400106 | orchestrator | 16:47:49.399 STDOUT terraform:  + content_md5 = (known after apply) 2025-08-29 16:47:49.400168 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_sha1 = (known after apply) 2025-08-29 16:47:49.400177 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_sha256 = (known after apply) 2025-08-29 16:47:49.400215 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_sha512 = (known after apply) 2025-08-29 16:47:49.400238 | orchestrator | 16:47:49.400 STDOUT terraform:  + directory_permission = "0777" 2025-08-29 16:47:49.400263 | orchestrator | 16:47:49.400 STDOUT terraform:  + file_permission = "0644" 2025-08-29 16:47:49.400330 | orchestrator | 16:47:49.400 STDOUT terraform:  + filename = ".id_rsa.ci.pub" 2025-08-29 16:47:49.400335 | orchestrator | 16:47:49.400 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.400341 | orchestrator | 16:47:49.400 STDOUT terraform:  } 2025-08-29 16:47:49.400463 | orchestrator | 16:47:49.400 STDOUT terraform:  # local_file.inventory will be created 2025-08-29 16:47:49.400488 | orchestrator | 16:47:49.400 STDOUT terraform:  + resource "local_file" "inventory" { 2025-08-29 16:47:49.400522 | orchestrator | 16:47:49.400 STDOUT terraform:  + content = (known after apply) 2025-08-29 16:47:49.400576 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_base64sha256 = (known after apply) 2025-08-29 16:47:49.400657 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_base64sha512 = (known after apply) 2025-08-29 16:47:49.400663 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_md5 = (known after apply) 2025-08-29 16:47:49.400693 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_sha1 = (known after apply) 2025-08-29 16:47:49.400771 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_sha256 = (known after apply) 2025-08-29 16:47:49.400776 | orchestrator | 16:47:49.400 STDOUT terraform:  + content_sha512 = (known after apply) 2025-08-29 16:47:49.400802 | orchestrator | 16:47:49.400 STDOUT terraform:  + directory_permission = "0777" 2025-08-29 16:47:49.400882 | orchestrator | 16:47:49.400 STDOUT terraform:  + file_permission = "0644" 2025-08-29 16:47:49.400887 | orchestrator | 16:47:49.400 STDOUT terraform:  + filename = "inventory.ci" 2025-08-29 16:47:49.400893 | orchestrator | 16:47:49.400 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.400897 | orchestrator | 16:47:49.400 STDOUT terraform:  } 2025-08-29 16:47:49.401131 | orchestrator | 16:47:49.400 STDOUT terraform:  # local_sensitive_file.id_rsa will be created 2025-08-29 16:47:49.401137 | orchestrator | 16:47:49.401 STDOUT terraform:  + resource "local_sensitive_file" "id_rsa" { 2025-08-29 16:47:49.401142 | orchestrator | 16:47:49.401 STDOUT terraform:  + content = (sensitive value) 2025-08-29 16:47:49.401146 | orchestrator | 16:47:49.401 STDOUT terraform:  + content_base64sha256 = (known after apply) 2025-08-29 16:47:49.401200 | orchestrator | 16:47:49.401 STDOUT terraform:  + content_base64sha512 = (known after apply) 2025-08-29 16:47:49.401234 | orchestrator | 16:47:49.401 STDOUT terraform:  + content_md5 = (known after apply) 2025-08-29 16:47:49.401287 | orchestrator | 16:47:49.401 STDOUT terraform:  + content_sha1 = (known after apply) 2025-08-29 16:47:49.401311 | orchestrator | 16:47:49.401 STDOUT terraform:  + content_sha256 = (known after apply) 2025-08-29 16:47:49.401367 | orchestrator | 16:47:49.401 STDOUT terraform:  + content_sha512 = (known after apply) 2025-08-29 16:47:49.401374 | orchestrator | 16:47:49.401 STDOUT terraform:  + directory_permission = "0700" 2025-08-29 16:47:49.401438 | orchestrator | 16:47:49.401 STDOUT terraform:  + file_permission = "0600" 2025-08-29 16:47:49.401444 | orchestrator | 16:47:49.401 STDOUT terraform:  + filename = ".id_rsa.ci" 2025-08-29 16:47:49.401449 | orchestrator | 16:47:49.401 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.401468 | orchestrator | 16:47:49.401 STDOUT terraform:  } 2025-08-29 16:47:49.401530 | orchestrator | 16:47:49.401 STDOUT terraform:  # null_resource.node_semaphore will be created 2025-08-29 16:47:49.401564 | orchestrator | 16:47:49.401 STDOUT terraform:  + resource "null_resource" "node_semaphore" { 2025-08-29 16:47:49.401588 | orchestrator | 16:47:49.401 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.401605 | orchestrator | 16:47:49.401 STDOUT terraform:  } 2025-08-29 16:47:49.401771 | orchestrator | 16:47:49.401 STDOUT terraform:  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created 2025-08-29 16:47:49.401832 | orchestrator | 16:47:49.401 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "manager_base_volume" { 2025-08-29 16:47:49.401879 | orchestrator | 16:47:49.401 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.401888 | orchestrator | 16:47:49.401 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.401925 | orchestrator | 16:47:49.401 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.401955 | orchestrator | 16:47:49.401 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.401996 | orchestrator | 16:47:49.401 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.402058 | orchestrator | 16:47:49.401 STDOUT terraform:  + name = "testbed-volume-manager-base" 2025-08-29 16:47:49.402122 | orchestrator | 16:47:49.402 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.402130 | orchestrator | 16:47:49.402 STDOUT terraform:  + size = 80 2025-08-29 16:47:49.402136 | orchestrator | 16:47:49.402 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.402176 | orchestrator | 16:47:49.402 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.402184 | orchestrator | 16:47:49.402 STDOUT terraform:  } 2025-08-29 16:47:49.402404 | orchestrator | 16:47:49.402 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created 2025-08-29 16:47:49.402443 | orchestrator | 16:47:49.402 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_base_volume" { 2025-08-29 16:47:49.402495 | orchestrator | 16:47:49.402 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.402503 | orchestrator | 16:47:49.402 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.402571 | orchestrator | 16:47:49.402 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.402576 | orchestrator | 16:47:49.402 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.402619 | orchestrator | 16:47:49.402 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.402650 | orchestrator | 16:47:49.402 STDOUT terraform:  + name = "testbed-volume-0-node-base" 2025-08-29 16:47:49.402690 | orchestrator | 16:47:49.402 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.402726 | orchestrator | 16:47:49.402 STDOUT terraform:  + size = 80 2025-08-29 16:47:49.402735 | orchestrator | 16:47:49.402 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.402756 | orchestrator | 16:47:49.402 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.402762 | orchestrator | 16:47:49.402 STDOUT terraform:  } 2025-08-29 16:47:49.402890 | orchestrator | 16:47:49.402 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created 2025-08-29 16:47:49.402938 | orchestrator | 16:47:49.402 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_base_volume" { 2025-08-29 16:47:49.402985 | orchestrator | 16:47:49.402 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.402999 | orchestrator | 16:47:49.402 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.403033 | orchestrator | 16:47:49.402 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.403072 | orchestrator | 16:47:49.403 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.403106 | orchestrator | 16:47:49.403 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.403161 | orchestrator | 16:47:49.403 STDOUT terraform:  + name = "testbed-volume-1-node-base" 2025-08-29 16:47:49.403206 | orchestrator | 16:47:49.403 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.403214 | orchestrator | 16:47:49.403 STDOUT terraform:  + size = 80 2025-08-29 16:47:49.403245 | orchestrator | 16:47:49.403 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.403252 | orchestrator | 16:47:49.403 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.403267 | orchestrator | 16:47:49.403 STDOUT terraform:  } 2025-08-29 16:47:49.403455 | orchestrator | 16:47:49.403 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created 2025-08-29 16:47:49.403511 | orchestrator | 16:47:49.403 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_base_volume" { 2025-08-29 16:47:49.403547 | orchestrator | 16:47:49.403 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.403602 | orchestrator | 16:47:49.403 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.403618 | orchestrator | 16:47:49.403 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.403654 | orchestrator | 16:47:49.403 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.403712 | orchestrator | 16:47:49.403 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.403741 | orchestrator | 16:47:49.403 STDOUT terraform:  + name = "testbed-volume-2-node-base" 2025-08-29 16:47:49.403784 | orchestrator | 16:47:49.403 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.403810 | orchestrator | 16:47:49.403 STDOUT terraform:  + size = 80 2025-08-29 16:47:49.403832 | orchestrator | 16:47:49.403 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.403880 | orchestrator | 16:47:49.403 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.403885 | orchestrator | 16:47:49.403 STDOUT terraform:  } 2025-08-29 16:47:49.404017 | orchestrator | 16:47:49.403 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created 2025-08-29 16:47:49.404053 | orchestrator | 16:47:49.403 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_base_volume" { 2025-08-29 16:47:49.404112 | orchestrator | 16:47:49.404 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.404117 | orchestrator | 16:47:49.404 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.404149 | orchestrator | 16:47:49.404 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.404190 | orchestrator | 16:47:49.404 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.404223 | orchestrator | 16:47:49.404 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.404270 | orchestrator | 16:47:49.404 STDOUT terraform:  + name = "testbed-volume-3-node-base" 2025-08-29 16:47:49.404307 | orchestrator | 16:47:49.404 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.404358 | orchestrator | 16:47:49.404 STDOUT terraform:  + size = 80 2025-08-29 16:47:49.404363 | orchestrator | 16:47:49.404 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.404368 | orchestrator | 16:47:49.404 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.404388 | orchestrator | 16:47:49.404 STDOUT terraform:  } 2025-08-29 16:47:49.404618 | orchestrator | 16:47:49.404 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created 2025-08-29 16:47:49.404656 | orchestrator | 16:47:49.404 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_base_volume" { 2025-08-29 16:47:49.404686 | orchestrator | 16:47:49.404 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.404709 | orchestrator | 16:47:49.404 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.404754 | orchestrator | 16:47:49.404 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.404791 | orchestrator | 16:47:49.404 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.404827 | orchestrator | 16:47:49.404 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.404872 | orchestrator | 16:47:49.404 STDOUT terraform:  + name = "testbed-volume-4-node-base" 2025-08-29 16:47:49.404934 | orchestrator | 16:47:49.404 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.404942 | orchestrator | 16:47:49.404 STDOUT terraform:  + size = 80 2025-08-29 16:47:49.404948 | orchestrator | 16:47:49.404 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.404985 | orchestrator | 16:47:49.404 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.404993 | orchestrator | 16:47:49.404 STDOUT terraform:  } 2025-08-29 16:47:49.405111 | orchestrator | 16:47:49.405 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created 2025-08-29 16:47:49.405174 | orchestrator | 16:47:49.405 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_base_volume" { 2025-08-29 16:47:49.405184 | orchestrator | 16:47:49.405 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.405234 | orchestrator | 16:47:49.405 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.405256 | orchestrator | 16:47:49.405 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.405297 | orchestrator | 16:47:49.405 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.405331 | orchestrator | 16:47:49.405 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.405376 | orchestrator | 16:47:49.405 STDOUT terraform:  + name = "testbed-volume-5-node-base" 2025-08-29 16:47:49.405416 | orchestrator | 16:47:49.405 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.405438 | orchestrator | 16:47:49.405 STDOUT terraform:  + size = 80 2025-08-29 16:47:49.405471 | orchestrator | 16:47:49.405 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.405488 | orchestrator | 16:47:49.405 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.405562 | orchestrator | 16:47:49.405 STDOUT terraform:  } 2025-08-29 16:47:49.405704 | orchestrator | 16:47:49.405 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[0] will be created 2025-08-29 16:47:49.405756 | orchestrator | 16:47:49.405 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.405815 | orchestrator | 16:47:49.405 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.405822 | orchestrator | 16:47:49.405 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.405855 | orchestrator | 16:47:49.405 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.405906 | orchestrator | 16:47:49.405 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.405947 | orchestrator | 16:47:49.405 STDOUT terraform:  + name = "testbed-volume-0-node-3" 2025-08-29 16:47:49.405975 | orchestrator | 16:47:49.405 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.406007 | orchestrator | 16:47:49.405 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.406029 | orchestrator | 16:47:49.405 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.406053 | orchestrator | 16:47:49.406 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.406060 | orchestrator | 16:47:49.406 STDOUT terraform:  } 2025-08-29 16:47:49.411467 | orchestrator | 16:47:49.411 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[1] will be created 2025-08-29 16:47:49.411521 | orchestrator | 16:47:49.411 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.411552 | orchestrator | 16:47:49.411 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.412191 | orchestrator | 16:47:49.411 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.412198 | orchestrator | 16:47:49.411 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.412202 | orchestrator | 16:47:49.411 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.412206 | orchestrator | 16:47:49.411 STDOUT terraform:  + name = "testbed-volume-1-node-4" 2025-08-29 16:47:49.412210 | orchestrator | 16:47:49.411 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.412214 | orchestrator | 16:47:49.411 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.412218 | orchestrator | 16:47:49.411 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.412222 | orchestrator | 16:47:49.411 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.412226 | orchestrator | 16:47:49.411 STDOUT terraform:  } 2025-08-29 16:47:49.412230 | orchestrator | 16:47:49.411 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[2] will be created 2025-08-29 16:47:49.412234 | orchestrator | 16:47:49.411 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.412237 | orchestrator | 16:47:49.411 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.412247 | orchestrator | 16:47:49.411 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.412251 | orchestrator | 16:47:49.411 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.412255 | orchestrator | 16:47:49.411 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.412259 | orchestrator | 16:47:49.412 STDOUT terraform:  + name = "testbed-volume-2-node-5" 2025-08-29 16:47:49.412262 | orchestrator | 16:47:49.412 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.412266 | orchestrator | 16:47:49.412 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.412276 | orchestrator | 16:47:49.412 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.412280 | orchestrator | 16:47:49.412 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.412284 | orchestrator | 16:47:49.412 STDOUT terraform:  } 2025-08-29 16:47:49.412289 | orchestrator | 16:47:49.412 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[3] will be created 2025-08-29 16:47:49.412746 | orchestrator | 16:47:49.412 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.412752 | orchestrator | 16:47:49.412 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.412756 | orchestrator | 16:47:49.412 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.412760 | orchestrator | 16:47:49.412 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.412764 | orchestrator | 16:47:49.412 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.412767 | orchestrator | 16:47:49.412 STDOUT terraform:  + name = "testbed-volume-3-node-3" 2025-08-29 16:47:49.412771 | orchestrator | 16:47:49.412 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.412775 | orchestrator | 16:47:49.412 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.412779 | orchestrator | 16:47:49.412 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.412783 | orchestrator | 16:47:49.412 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.412786 | orchestrator | 16:47:49.412 STDOUT terraform:  } 2025-08-29 16:47:49.412790 | orchestrator | 16:47:49.412 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[4] will be created 2025-08-29 16:47:49.412794 | orchestrator | 16:47:49.412 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.412798 | orchestrator | 16:47:49.412 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.412803 | orchestrator | 16:47:49.412 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.412807 | orchestrator | 16:47:49.412 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.412813 | orchestrator | 16:47:49.412 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.412854 | orchestrator | 16:47:49.412 STDOUT terraform:  + name = "testbed-volume-4-node-4" 2025-08-29 16:47:49.412930 | orchestrator | 16:47:49.412 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.412939 | orchestrator | 16:47:49.412 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.412945 | orchestrator | 16:47:49.412 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.412974 | orchestrator | 16:47:49.412 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.412981 | orchestrator | 16:47:49.412 STDOUT terraform:  } 2025-08-29 16:47:49.413034 | orchestrator | 16:47:49.412 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[5] will be created 2025-08-29 16:47:49.413091 | orchestrator | 16:47:49.413 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.413840 | orchestrator | 16:47:49.413 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.413845 | orchestrator | 16:47:49.413 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.413849 | orchestrator | 16:47:49.413 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.413853 | orchestrator | 16:47:49.413 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.413857 | orchestrator | 16:47:49.413 STDOUT terraform:  + name = "testbed-volume-5-node-5" 2025-08-29 16:47:49.413861 | orchestrator | 16:47:49.413 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.413867 | orchestrator | 16:47:49.413 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.413871 | orchestrator | 16:47:49.413 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.413875 | orchestrator | 16:47:49.413 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.413879 | orchestrator | 16:47:49.413 STDOUT terraform:  } 2025-08-29 16:47:49.413883 | orchestrator | 16:47:49.413 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[6] will be created 2025-08-29 16:47:49.413887 | orchestrator | 16:47:49.413 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.413890 | orchestrator | 16:47:49.413 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.413894 | orchestrator | 16:47:49.413 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.413898 | orchestrator | 16:47:49.413 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.413902 | orchestrator | 16:47:49.413 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.413906 | orchestrator | 16:47:49.413 STDOUT terraform:  + name = "testbed-volume-6-node-3" 2025-08-29 16:47:49.413909 | orchestrator | 16:47:49.413 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.413913 | orchestrator | 16:47:49.413 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.413917 | orchestrator | 16:47:49.413 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.413921 | orchestrator | 16:47:49.413 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.413925 | orchestrator | 16:47:49.413 STDOUT terraform:  } 2025-08-29 16:47:49.413928 | orchestrator | 16:47:49.413 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[7] will be created 2025-08-29 16:47:49.413934 | orchestrator | 16:47:49.413 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.413942 | orchestrator | 16:47:49.413 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.413945 | orchestrator | 16:47:49.413 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.413951 | orchestrator | 16:47:49.413 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.413991 | orchestrator | 16:47:49.413 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.414028 | orchestrator | 16:47:49.413 STDOUT terraform:  + name = "testbed-volume-7-node-4" 2025-08-29 16:47:49.414071 | orchestrator | 16:47:49.414 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.414097 | orchestrator | 16:47:49.414 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.414127 | orchestrator | 16:47:49.414 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.414147 | orchestrator | 16:47:49.414 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.414153 | orchestrator | 16:47:49.414 STDOUT terraform:  } 2025-08-29 16:47:49.414208 | orchestrator | 16:47:49.414 STDOUT terraform:  # openstack_blockstorage_volume_v3.node_volume[8] will be created 2025-08-29 16:47:49.414253 | orchestrator | 16:47:49.414 STDOUT terraform:  + resource "openstack_blockstorage_volume_v3" "node_volume" { 2025-08-29 16:47:49.414290 | orchestrator | 16:47:49.414 STDOUT terraform:  + attachment = (known after apply) 2025-08-29 16:47:49.414325 | orchestrator | 16:47:49.414 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.414384 | orchestrator | 16:47:49.414 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.414390 | orchestrator | 16:47:49.414 STDOUT terraform:  + metadata = (known after apply) 2025-08-29 16:47:49.414440 | orchestrator | 16:47:49.414 STDOUT terraform:  + name = "testbed-volume-8-node-5" 2025-08-29 16:47:49.414509 | orchestrator | 16:47:49.414 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.414544 | orchestrator | 16:47:49.414 STDOUT terraform:  + size = 20 2025-08-29 16:47:49.414551 | orchestrator | 16:47:49.414 STDOUT terraform:  + volume_retype_policy = "never" 2025-08-29 16:47:49.414602 | orchestrator | 16:47:49.414 STDOUT terraform:  + volume_type = "ssd" 2025-08-29 16:47:49.414610 | orchestrator | 16:47:49.414 STDOUT terraform:  } 2025-08-29 16:47:49.414646 | orchestrator | 16:47:49.414 STDOUT terraform:  # openstack_compute_instance_v2.manager_server will be created 2025-08-29 16:47:49.414698 | orchestrator | 16:47:49.414 STDOUT terraform:  + resource "openstack_compute_instance_v2" "manager_server" { 2025-08-29 16:47:49.414731 | orchestrator | 16:47:49.414 STDOUT terraform:  + access_ip_v4 = (known after apply) 2025-08-29 16:47:49.414755 | orchestrator | 16:47:49.414 STDOUT terraform:  + access_ip_v6 = (known after apply) 2025-08-29 16:47:49.414790 | orchestrator | 16:47:49.414 STDOUT terraform:  + all_metadata = (known after apply) 2025-08-29 16:47:49.414821 | orchestrator | 16:47:49.414 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.414860 | orchestrator | 16:47:49.414 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.414872 | orchestrator | 16:47:49.414 STDOUT terraform:  + config_drive = true 2025-08-29 16:47:49.414898 | orchestrator | 16:47:49.414 STDOUT terraform:  + created = (known after apply) 2025-08-29 16:47:49.414933 | orchestrator | 16:47:49.414 STDOUT terraform:  + flavor_id = (known after apply) 2025-08-29 16:47:49.414975 | orchestrator | 16:47:49.414 STDOUT terraform:  + flavor_name = "OSISM-4V-16" 2025-08-29 16:47:49.414985 | orchestrator | 16:47:49.414 STDOUT terraform:  + force_delete = false 2025-08-29 16:47:49.415026 | orchestrator | 16:47:49.414 STDOUT terraform:  + hypervisor_hostname = (known after apply) 2025-08-29 16:47:49.415080 | orchestrator | 16:47:49.415 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.415087 | orchestrator | 16:47:49.415 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.415121 | orchestrator | 16:47:49.415 STDOUT terraform:  + image_name = (known after apply) 2025-08-29 16:47:49.415157 | orchestrator | 16:47:49.415 STDOUT terraform:  + key_pair = "testbed" 2025-08-29 16:47:49.415195 | orchestrator | 16:47:49.415 STDOUT terraform:  + name = "testbed-manager" 2025-08-29 16:47:49.415201 | orchestrator | 16:47:49.415 STDOUT terraform:  + power_state = "active" 2025-08-29 16:47:49.415330 | orchestrator | 16:47:49.415 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.415376 | orchestrator | 16:47:49.415 STDOUT terraform:  + security_groups = (known after apply) 2025-08-29 16:47:49.415381 | orchestrator | 16:47:49.415 STDOUT terraform:  + stop_before_destroy = false 2025-08-29 16:47:49.415391 | orchestrator | 16:47:49.415 STDOUT terraform:  + updated = (known after apply) 2025-08-29 16:47:49.415395 | orchestrator | 16:47:49.415 STDOUT terraform:  + user_data = (sensitive value) 2025-08-29 16:47:49.415399 | orchestrator | 16:47:49.415 STDOUT terraform:  + block_device { 2025-08-29 16:47:49.415403 | orchestrator | 16:47:49.415 STDOUT terraform:  + boot_index = 0 2025-08-29 16:47:49.415412 | orchestrator | 16:47:49.415 STDOUT terraform:  + delete_on_termination = false 2025-08-29 16:47:49.415570 | orchestrator | 16:47:49.415 STDOUT terraform:  + destination_type = "volume" 2025-08-29 16:47:49.415584 | orchestrator | 16:47:49.415 STDOUT terraform:  + multiattach = false 2025-08-29 16:47:49.415621 | orchestrator | 16:47:49.415 STDOUT terraform:  + source_type = "volume" 2025-08-29 16:47:49.415632 | orchestrator | 16:47:49.415 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.415647 | orchestrator | 16:47:49.415 STDOUT terraform:  } 2025-08-29 16:47:49.415657 | orchestrator | 16:47:49.415 STDOUT terraform:  + network { 2025-08-29 16:47:49.415667 | orchestrator | 16:47:49.415 STDOUT terraform:  + access_network = false 2025-08-29 16:47:49.415680 | orchestrator | 16:47:49.415 STDOUT terraform:  + fixed_ip_v4 = (known after apply) 2025-08-29 16:47:49.416685 | orchestrator | 16:47:49.415 STDOUT terraform:  + fixed_ip_v6 = (known after apply) 2025-08-29 16:47:49.416704 | orchestrator | 16:47:49.415 STDOUT terraform:  + mac = (known after apply) 2025-08-29 16:47:49.416732 | orchestrator | 16:47:49.415 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.416742 | orchestrator | 16:47:49.415 STDOUT terraform:  + port = (known after apply) 2025-08-29 16:47:49.416752 | orchestrator | 16:47:49.415 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.416762 | orchestrator | 16:47:49.415 STDOUT terraform:  } 2025-08-29 16:47:49.416773 | orchestrator | 16:47:49.415 STDOUT terraform:  } 2025-08-29 16:47:49.416783 | orchestrator | 16:47:49.415 STDOUT terraform:  # openstack_compute_instance_v2.node_server[0] will be created 2025-08-29 16:47:49.416793 | orchestrator | 16:47:49.415 STDOUT terraform:  + resource "openstack_compute_instance_v2" "node_server" { 2025-08-29 16:47:49.416802 | orchestrator | 16:47:49.415 STDOUT terraform:  + access_ip_v4 = (known after apply) 2025-08-29 16:47:49.416823 | orchestrator | 16:47:49.415 STDOUT terraform:  + access_ip_v6 = (known after apply) 2025-08-29 16:47:49.416833 | orchestrator | 16:47:49.416 STDOUT terraform:  + all_metadata = (known after apply) 2025-08-29 16:47:49.416842 | orchestrator | 16:47:49.416 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.416852 | orchestrator | 16:47:49.416 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.416861 | orchestrator | 16:47:49.416 STDOUT terraform:  + config_drive = true 2025-08-29 16:47:49.416871 | orchestrator | 16:47:49.416 STDOUT terraform:  + created = (known after apply) 2025-08-29 16:47:49.416880 | orchestrator | 16:47:49.416 STDOUT terraform:  + flavor_id = (known after apply) 2025-08-29 16:47:49.416889 | orchestrator | 16:47:49.416 STDOUT terraform:  + flavor_name = "OSISM-8V-32" 2025-08-29 16:47:49.416899 | orchestrator | 16:47:49.416 STDOUT terraform:  + force_delete = false 2025-08-29 16:47:49.416909 | orchestrator | 16:47:49.416 STDOUT terraform:  + hypervisor_hostname = (known after apply) 2025-08-29 16:47:49.416919 | orchestrator | 16:47:49.416 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.416928 | orchestrator | 16:47:49.416 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.416938 | orchestrator | 16:47:49.416 STDOUT terraform:  + image_name = (known after apply) 2025-08-29 16:47:49.416947 | orchestrator | 16:47:49.416 STDOUT terraform:  + key_pair = "testbed" 2025-08-29 16:47:49.416957 | orchestrator | 16:47:49.416 STDOUT terraform:  + name = "testbed-node-0" 2025-08-29 16:47:49.416967 | orchestrator | 16:47:49.416 STDOUT terraform:  + power_state = "active" 2025-08-29 16:47:49.416976 | orchestrator | 16:47:49.416 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.416986 | orchestrator | 16:47:49.416 STDOUT terraform:  + security_groups = (known after apply) 2025-08-29 16:47:49.416995 | orchestrator | 16:47:49.416 STDOUT terraform:  + stop_before_destroy = false 2025-08-29 16:47:49.417004 | orchestrator | 16:47:49.416 STDOUT terraform:  + updated = (known after apply) 2025-08-29 16:47:49.417014 | orchestrator | 16:47:49.416 STDOUT terraform:  + user_data = "ae09e46b224a6ca206a9ed4f8f8a4f8520827854" 2025-08-29 16:47:49.417025 | orchestrator | 16:47:49.416 STDOUT terraform:  + block_device { 2025-08-29 16:47:49.417047 | orchestrator | 16:47:49.416 STDOUT terraform:  + boot_index = 0 2025-08-29 16:47:49.417057 | orchestrator | 16:47:49.416 STDOUT terraform:  + delete_on_termination = false 2025-08-29 16:47:49.417067 | orchestrator | 16:47:49.416 STDOUT terraform:  + destination_type = "volume" 2025-08-29 16:47:49.417076 | orchestrator | 16:47:49.416 STDOUT terraform:  + multiattach = false 2025-08-29 16:47:49.417086 | orchestrator | 16:47:49.416 STDOUT terraform:  + source_type = "volume" 2025-08-29 16:47:49.417096 | orchestrator | 16:47:49.416 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.417105 | orchestrator | 16:47:49.416 STDOUT terraform:  } 2025-08-29 16:47:49.417115 | orchestrator | 16:47:49.416 STDOUT terraform:  + network { 2025-08-29 16:47:49.417125 | orchestrator | 16:47:49.416 STDOUT terraform:  + access_network = false 2025-08-29 16:47:49.417135 | orchestrator | 16:47:49.416 STDOUT terraform:  + fixed_ip_v4 = (known after apply) 2025-08-29 16:47:49.417144 | orchestrator | 16:47:49.416 STDOUT terraform:  + fixed_ip_v6 = (known after apply) 2025-08-29 16:47:49.417154 | orchestrator | 16:47:49.416 STDOUT terraform:  + mac = (known after apply) 2025-08-29 16:47:49.417163 | orchestrator | 16:47:49.416 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.417173 | orchestrator | 16:47:49.416 STDOUT terraform:  + port = (known after apply) 2025-08-29 16:47:49.417186 | orchestrator | 16:47:49.417 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.417196 | orchestrator | 16:47:49.417 STDOUT terraform:  } 2025-08-29 16:47:49.417206 | orchestrator | 16:47:49.417 STDOUT terraform:  } 2025-08-29 16:47:49.417219 | orchestrator | 16:47:49.417 STDOUT terraform:  # openstack_compute_instance_v2.node_server[1] will be created 2025-08-29 16:47:49.417803 | orchestrator | 16:47:49.417 STDOUT terraform:  + resource "openstack_compute_instance_v2" "node_server" { 2025-08-29 16:47:49.417819 | orchestrator | 16:47:49.417 STDOUT terraform:  + access_ip_v4 = (known after apply) 2025-08-29 16:47:49.417830 | orchestrator | 16:47:49.417 STDOUT terraform:  + access_ip_v6 = (known after apply) 2025-08-29 16:47:49.417840 | orchestrator | 16:47:49.417 STDOUT terraform:  + all_metadata = (known after apply) 2025-08-29 16:47:49.417850 | orchestrator | 16:47:49.417 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.417860 | orchestrator | 16:47:49.417 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.417869 | orchestrator | 16:47:49.417 STDOUT terraform:  + config_drive = true 2025-08-29 16:47:49.417879 | orchestrator | 16:47:49.417 STDOUT terraform:  + created = (known after apply) 2025-08-29 16:47:49.417889 | orchestrator | 16:47:49.417 STDOUT terraform:  + flavor_id = (known after apply) 2025-08-29 16:47:49.417898 | orchestrator | 16:47:49.417 STDOUT terraform:  + flavor_name = "OSISM-8V-32" 2025-08-29 16:47:49.417908 | orchestrator | 16:47:49.417 STDOUT terraform:  + force_delete = false 2025-08-29 16:47:49.417924 | orchestrator | 16:47:49.417 STDOUT terraform:  + hypervisor_hostname = (known after apply) 2025-08-29 16:47:49.417942 | orchestrator | 16:47:49.417 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.417952 | orchestrator | 16:47:49.417 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.417961 | orchestrator | 16:47:49.417 STDOUT terraform:  + image_name = (known after apply) 2025-08-29 16:47:49.417971 | orchestrator | 16:47:49.417 STDOUT terraform:  + key_pair = "testbed" 2025-08-29 16:47:49.417985 | orchestrator | 16:47:49.417 STDOUT terraform:  + name = "testbed-node-1" 2025-08-29 16:47:49.417994 | orchestrator | 16:47:49.417 STDOUT terraform:  + power_state = "active" 2025-08-29 16:47:49.418004 | orchestrator | 16:47:49.417 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.418068 | orchestrator | 16:47:49.417 STDOUT terraform:  + security_groups = (known after apply) 2025-08-29 16:47:49.418081 | orchestrator | 16:47:49.417 STDOUT terraform:  + stop_before_destroy = false 2025-08-29 16:47:49.418091 | orchestrator | 16:47:49.417 STDOUT terraform:  + updated = (known after apply) 2025-08-29 16:47:49.418105 | orchestrator | 16:47:49.417 STDOUT terraform:  + user_data = "ae09e46b224a6ca206a9ed4f8f8a4f8520827854" 2025-08-29 16:47:49.418115 | orchestrator | 16:47:49.418 STDOUT terraform:  + block_device { 2025-08-29 16:47:49.418125 | orchestrator | 16:47:49.418 STDOUT terraform:  + boot_index = 0 2025-08-29 16:47:49.418141 | orchestrator | 16:47:49.418 STDOUT terraform:  + delete_on_termination = false 2025-08-29 16:47:49.418154 | orchestrator | 16:47:49.418 STDOUT terraform:  + destination_type = "volume" 2025-08-29 16:47:49.418164 | orchestrator | 16:47:49.418 STDOUT terraform:  + multiattach = false 2025-08-29 16:47:49.418177 | orchestrator | 16:47:49.418 STDOUT terraform:  + source_type = "volume" 2025-08-29 16:47:49.418312 | orchestrator | 16:47:49.418 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.418325 | orchestrator | 16:47:49.418 STDOUT terraform:  } 2025-08-29 16:47:49.418336 | orchestrator | 16:47:49.418 STDOUT terraform:  + network { 2025-08-29 16:47:49.418345 | orchestrator | 16:47:49.418 STDOUT terraform:  + access_network = false 2025-08-29 16:47:49.418355 | orchestrator | 16:47:49.418 STDOUT terraform:  + fixed_ip_v4 = (known after apply) 2025-08-29 16:47:49.418369 | orchestrator | 16:47:49.418 STDOUT terraform:  + fixed_ip_v6 = (known after apply) 2025-08-29 16:47:49.418379 | orchestrator | 16:47:49.418 STDOUT terraform:  + mac = (known after apply) 2025-08-29 16:47:49.418388 | orchestrator | 16:47:49.418 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.418401 | orchestrator | 16:47:49.418 STDOUT terraform:  + port = (known after apply) 2025-08-29 16:47:49.418471 | orchestrator | 16:47:49.418 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.418484 | orchestrator | 16:47:49.418 STDOUT terraform:  } 2025-08-29 16:47:49.418493 | orchestrator | 16:47:49.418 STDOUT terraform:  } 2025-08-29 16:47:49.418507 | orchestrator | 16:47:49.418 STDOUT terraform:  # openstack_compute_instance_v2.node_server[2] will be created 2025-08-29 16:47:49.418761 | orchestrator | 16:47:49.418 STDOUT terraform:  + resource "openstack_compute_instance_v2" "node_server" { 2025-08-29 16:47:49.418789 | orchestrator | 16:47:49.418 STDOUT terraform:  + access_ip_v4 = (known after apply) 2025-08-29 16:47:49.418799 | orchestrator | 16:47:49.418 STDOUT terraform:  + access_ip_v6 = (known after apply) 2025-08-29 16:47:49.418809 | orchestrator | 16:47:49.418 STDOUT terraform:  + all_metadata = (known after apply) 2025-08-29 16:47:49.418819 | orchestrator | 16:47:49.418 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.418828 | orchestrator | 16:47:49.418 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.418836 | orchestrator | 16:47:49.418 STDOUT terraform:  + config_drive = true 2025-08-29 16:47:49.418847 | orchestrator | 16:47:49.418 STDOUT terraform:  + created = (known after apply) 2025-08-29 16:47:49.418855 | orchestrator | 16:47:49.418 STDOUT terraform:  + flavor_id = (known after apply) 2025-08-29 16:47:49.418866 | orchestrator | 16:47:49.418 STDOUT terraform:  + flavor_name = "OSISM-8V-32" 2025-08-29 16:47:49.418899 | orchestrator | 16:47:49.418 STDOUT terraform:  + force_delete = false 2025-08-29 16:47:49.418933 | orchestrator | 16:47:49.418 STDOUT terraform:  + hypervisor_hostname = (known after apply) 2025-08-29 16:47:49.418976 | orchestrator | 16:47:49.418 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.419037 | orchestrator | 16:47:49.418 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.419048 | orchestrator | 16:47:49.418 STDOUT terraform:  + image_name = (known after apply) 2025-08-29 16:47:49.419058 | orchestrator | 16:47:49.419 STDOUT terraform:  + key_pair = "testbed" 2025-08-29 16:47:49.419097 | orchestrator | 16:47:49.419 STDOUT terraform:  + name = "testbed-node-2" 2025-08-29 16:47:49.419109 | orchestrator | 16:47:49.419 STDOUT terraform:  + power_state = "active" 2025-08-29 16:47:49.419183 | orchestrator | 16:47:49.419 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.419196 | orchestrator | 16:47:49.419 STDOUT terraform:  + security_groups = (known after apply) 2025-08-29 16:47:49.419225 | orchestrator | 16:47:49.419 STDOUT terraform:  + stop_before_destroy = false 2025-08-29 16:47:49.419295 | orchestrator | 16:47:49.419 STDOUT terraform:  + updated = (known after apply) 2025-08-29 16:47:49.419308 | orchestrator | 16:47:49.419 STDOUT terraform:  + user_data = "ae09e46b224a6ca206a9ed4f8f8a4f8520827854" 2025-08-29 16:47:49.419319 | orchestrator | 16:47:49.419 STDOUT terraform:  + block_device { 2025-08-29 16:47:49.419352 | orchestrator | 16:47:49.419 STDOUT terraform:  + boot_index = 0 2025-08-29 16:47:49.419386 | orchestrator | 16:47:49.419 STDOUT terraform:  + delete_on_termination = false 2025-08-29 16:47:49.419422 | orchestrator | 16:47:49.419 STDOUT terraform:  + destination_type = "volume" 2025-08-29 16:47:49.419434 | orchestrator | 16:47:49.419 STDOUT terraform:  + multiattach = false 2025-08-29 16:47:49.419499 | orchestrator | 16:47:49.419 STDOUT terraform:  + source_type = "volume" 2025-08-29 16:47:49.419515 | orchestrator | 16:47:49.419 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.419529 | orchestrator | 16:47:49.419 STDOUT terraform:  } 2025-08-29 16:47:49.419540 | orchestrator | 16:47:49.419 STDOUT terraform:  + network { 2025-08-29 16:47:49.419548 | orchestrator | 16:47:49.419 STDOUT terraform:  + access_network = false 2025-08-29 16:47:49.419658 | orchestrator | 16:47:49.419 STDOUT terraform:  + fixed_ip_v4 = (known after apply) 2025-08-29 16:47:49.419674 | orchestrator | 16:47:49.419 STDOUT terraform:  + fixed_ip_v6 = (known after apply) 2025-08-29 16:47:49.419683 | orchestrator | 16:47:49.419 STDOUT terraform:  + mac = (known after apply) 2025-08-29 16:47:49.419694 | orchestrator | 16:47:49.419 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.419702 | orchestrator | 16:47:49.419 STDOUT terraform:  + port = (known after apply) 2025-08-29 16:47:49.419757 | orchestrator | 16:47:49.419 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.419768 | orchestrator | 16:47:49.419 STDOUT terraform:  } 2025-08-29 16:47:49.419778 | orchestrator | 16:47:49.419 STDOUT terraform:  } 2025-08-29 16:47:49.419789 | orchestrator | 16:47:49.419 STDOUT terraform:  # openstack_compute_instance_v2.node_server[3] will be created 2025-08-29 16:47:49.419848 | orchestrator | 16:47:49.419 STDOUT terraform:  + resource "openstack_compute_instance_v2" "node_server" { 2025-08-29 16:47:49.419861 | orchestrator | 16:47:49.419 STDOUT terraform:  + access_ip_v4 = (known after apply) 2025-08-29 16:47:49.419972 | orchestrator | 16:47:49.419 STDOUT terraform:  + access_ip_v6 = (known after apply) 2025-08-29 16:47:49.419986 | orchestrator | 16:47:49.419 STDOUT terraform:  + all_metadata = (known after apply) 2025-08-29 16:47:49.420048 | orchestrator | 16:47:49.419 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.420058 | orchestrator | 16:47:49.420 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.420069 | orchestrator | 16:47:49.420 STDOUT terraform:  + config_drive = true 2025-08-29 16:47:49.420080 | orchestrator | 16:47:49.420 STDOUT terraform:  + created = (known after apply) 2025-08-29 16:47:49.420135 | orchestrator | 16:47:49.420 STDOUT terraform:  + flavor_id = (known after apply) 2025-08-29 16:47:49.420149 | orchestrator | 16:47:49.420 STDOUT terraform:  + flavor_name = "OSISM-8V-32" 2025-08-29 16:47:49.420181 | orchestrator | 16:47:49.420 STDOUT terraform:  + force_delete = false 2025-08-29 16:47:49.420210 | orchestrator | 16:47:49.420 STDOUT terraform:  + hypervisor_hostname = (known after apply) 2025-08-29 16:47:49.420265 | orchestrator | 16:47:49.420 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.420278 | orchestrator | 16:47:49.420 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.420332 | orchestrator | 16:47:49.420 STDOUT terraform:  + image_name = (known after apply) 2025-08-29 16:47:49.420345 | orchestrator | 16:47:49.420 STDOUT terraform:  + key_pair = "testbed" 2025-08-29 16:47:49.420406 | orchestrator | 16:47:49.420 STDOUT terraform:  + name = "testbed-node-3" 2025-08-29 16:47:49.420417 | orchestrator | 16:47:49.420 STDOUT terraform:  + power_state = "active" 2025-08-29 16:47:49.420439 | orchestrator | 16:47:49.420 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.420449 | orchestrator | 16:47:49.420 STDOUT terraform:  + security_groups = (known after apply) 2025-08-29 16:47:49.420478 | orchestrator | 16:47:49.420 STDOUT terraform:  + stop_before_destroy = false 2025-08-29 16:47:49.420517 | orchestrator | 16:47:49.420 STDOUT terraform:  + updated = (known after apply) 2025-08-29 16:47:49.420576 | orchestrator | 16:47:49.420 STDOUT terraform:  + user_data = "ae09e46b224a6ca206a9ed4f8f8a4f8520827854" 2025-08-29 16:47:49.420587 | orchestrator | 16:47:49.420 STDOUT terraform:  + block_device { 2025-08-29 16:47:49.420655 | orchestrator | 16:47:49.420 STDOUT terraform:  + boot_index = 0 2025-08-29 16:47:49.420670 | orchestrator | 16:47:49.420 STDOUT terraform:  + delete_on_termination = false 2025-08-29 16:47:49.420681 | orchestrator | 16:47:49.420 STDOUT terraform:  + destination_type = "volume" 2025-08-29 16:47:49.420692 | orchestrator | 16:47:49.420 STDOUT terraform:  + multiattach = false 2025-08-29 16:47:49.420770 | orchestrator | 16:47:49.420 STDOUT terraform:  + source_type = "volume" 2025-08-29 16:47:49.420780 | orchestrator | 16:47:49.420 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.420791 | orchestrator | 16:47:49.420 STDOUT terraform:  } 2025-08-29 16:47:49.420800 | orchestrator | 16:47:49.420 STDOUT terraform:  + network { 2025-08-29 16:47:49.420810 | orchestrator | 16:47:49.420 STDOUT terraform:  + access_network = false 2025-08-29 16:47:49.420840 | orchestrator | 16:47:49.420 STDOUT terraform:  + fixed_ip_v4 = (known after apply) 2025-08-29 16:47:49.420901 | orchestrator | 16:47:49.420 STDOUT terraform:  + fixed_ip_v6 = (known after apply) 2025-08-29 16:47:49.420914 | orchestrator | 16:47:49.420 STDOUT terraform:  + mac = (known after apply) 2025-08-29 16:47:49.420924 | orchestrator | 16:47:49.420 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.421028 | orchestrator | 16:47:49.420 STDOUT terraform:  + port = (known after apply) 2025-08-29 16:47:49.421042 | orchestrator | 16:47:49.420 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.421050 | orchestrator | 16:47:49.421 STDOUT terraform:  } 2025-08-29 16:47:49.421060 | orchestrator | 16:47:49.421 STDOUT terraform:  } 2025-08-29 16:47:49.421098 | orchestrator | 16:47:49.421 STDOUT terraform:  # openstack_compute_instance_v2.node_server[4] will be created 2025-08-29 16:47:49.421134 | orchestrator | 16:47:49.421 STDOUT terraform:  + resource "openstack_compute_instance_v2" "node_server" { 2025-08-29 16:47:49.421187 | orchestrator | 16:47:49.421 STDOUT terraform:  + access_ip_v4 = (known after apply) 2025-08-29 16:47:49.421198 | orchestrator | 16:47:49.421 STDOUT terraform:  + access_ip_v6 = (known after apply) 2025-08-29 16:47:49.421256 | orchestrator | 16:47:49.421 STDOUT terraform:  + all_metadata = (known after apply) 2025-08-29 16:47:49.421267 | orchestrator | 16:47:49.421 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.421298 | orchestrator | 16:47:49.421 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.421315 | orchestrator | 16:47:49.421 STDOUT terraform:  + config_drive = true 2025-08-29 16:47:49.421364 | orchestrator | 16:47:49.421 STDOUT terraform:  + created = (known after apply) 2025-08-29 16:47:49.421375 | orchestrator | 16:47:49.421 STDOUT terraform:  + flavor_id = (known after apply) 2025-08-29 16:47:49.421424 | orchestrator | 16:47:49.421 STDOUT terraform:  + flavor_name = "OSISM-8V-32" 2025-08-29 16:47:49.421442 | orchestrator | 16:47:49.421 STDOUT terraform:  + force_delete = false 2025-08-29 16:47:49.421468 | orchestrator | 16:47:49.421 STDOUT terraform:  + hypervisor_hostname = (known after apply) 2025-08-29 16:47:49.421521 | orchestrator | 16:47:49.421 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.421533 | orchestrator | 16:47:49.421 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.421650 | orchestrator | 16:47:49.421 STDOUT terraform:  + image_name = (known after apply) 2025-08-29 16:47:49.421661 | orchestrator | 16:47:49.421 STDOUT terraform:  + key_pair = "testbed" 2025-08-29 16:47:49.421667 | orchestrator | 16:47:49.421 STDOUT terraform:  + name = "testbed-node-4" 2025-08-29 16:47:49.421677 | orchestrator | 16:47:49.421 STDOUT terraform:  + power_state = "active" 2025-08-29 16:47:49.421685 | orchestrator | 16:47:49.421 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.421730 | orchestrator | 16:47:49.421 STDOUT terraform:  + security_groups = (known after apply) 2025-08-29 16:47:49.421746 | orchestrator | 16:47:49.421 STDOUT terraform:  + stop_before_destroy = false 2025-08-29 16:47:49.421778 | orchestrator | 16:47:49.421 STDOUT terraform:  + updated = (known after apply) 2025-08-29 16:47:49.421830 | orchestrator | 16:47:49.421 STDOUT terraform:  + user_data = "ae09e46b224a6ca206a9ed4f8f8a4f8520827854" 2025-08-29 16:47:49.421846 | orchestrator | 16:47:49.421 STDOUT terraform:  + block_device { 2025-08-29 16:47:49.421855 | orchestrator | 16:47:49.421 STDOUT terraform:  + boot_index = 0 2025-08-29 16:47:49.421898 | orchestrator | 16:47:49.421 STDOUT terraform:  + delete_on_termination = false 2025-08-29 16:47:49.421914 | orchestrator | 16:47:49.421 STDOUT terraform:  + destination_type = "volume" 2025-08-29 16:47:49.421945 | orchestrator | 16:47:49.421 STDOUT terraform:  + multiattach = false 2025-08-29 16:47:49.421975 | orchestrator | 16:47:49.421 STDOUT terraform:  + source_type = "volume" 2025-08-29 16:47:49.422052 | orchestrator | 16:47:49.421 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.422068 | orchestrator | 16:47:49.422 STDOUT terraform:  } 2025-08-29 16:47:49.422077 | orchestrator | 16:47:49.422 STDOUT terraform:  + network { 2025-08-29 16:47:49.422086 | orchestrator | 16:47:49.422 STDOUT terraform:  + access_network = false 2025-08-29 16:47:49.422148 | orchestrator | 16:47:49.422 STDOUT terraform:  + fixed_ip_v4 = (known after apply) 2025-08-29 16:47:49.422161 | orchestrator | 16:47:49.422 STDOUT terraform:  + fixed_ip_v6 = (known after apply) 2025-08-29 16:47:49.422223 | orchestrator | 16:47:49.422 STDOUT terraform:  + mac = (known after apply) 2025-08-29 16:47:49.422242 | orchestrator | 16:47:49.422 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.422251 | orchestrator | 16:47:49.422 STDOUT terraform:  + port = (known after apply) 2025-08-29 16:47:49.422278 | orchestrator | 16:47:49.422 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.422288 | orchestrator | 16:47:49.422 STDOUT terraform:  } 2025-08-29 16:47:49.422297 | orchestrator | 16:47:49.422 STDOUT terraform:  } 2025-08-29 16:47:49.422339 | orchestrator | 16:47:49.422 STDOUT terraform:  # openstack_compute_instance_v2.node_server[5] will be created 2025-08-29 16:47:49.422393 | orchestrator | 16:47:49.422 STDOUT terraform:  + resource "openstack_compute_instance_v2" "node_server" { 2025-08-29 16:47:49.422419 | orchestrator | 16:47:49.422 STDOUT terraform:  + access_ip_v4 = (known after apply) 2025-08-29 16:47:49.422468 | orchestrator | 16:47:49.422 STDOUT terraform:  + access_ip_v6 = (known after apply) 2025-08-29 16:47:49.422484 | orchestrator | 16:47:49.422 STDOUT terraform:  + all_metadata = (known after apply) 2025-08-29 16:47:49.422521 | orchestrator | 16:47:49.422 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.422608 | orchestrator | 16:47:49.422 STDOUT terraform:  + availability_zone = "nova" 2025-08-29 16:47:49.422618 | orchestrator | 16:47:49.422 STDOUT terraform:  + config_drive = true 2025-08-29 16:47:49.422628 | orchestrator | 16:47:49.422 STDOUT terraform:  + created = (known after apply) 2025-08-29 16:47:49.422637 | orchestrator | 16:47:49.422 STDOUT terraform:  + flavor_id = (known after apply) 2025-08-29 16:47:49.422670 | orchestrator | 16:47:49.422 STDOUT terraform:  + flavor_name = "OSISM-8V-32" 2025-08-29 16:47:49.422680 | orchestrator | 16:47:49.422 STDOUT terraform:  + force_delete = false 2025-08-29 16:47:49.422742 | orchestrator | 16:47:49.422 STDOUT terraform:  + hypervisor_hostname = (known after apply) 2025-08-29 16:47:49.422758 | orchestrator | 16:47:49.422 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.422789 | orchestrator | 16:47:49.422 STDOUT terraform:  + image_id = (known after apply) 2025-08-29 16:47:49.422843 | orchestrator | 16:47:49.422 STDOUT terraform:  + image_name = (known after apply) 2025-08-29 16:47:49.422854 | orchestrator | 16:47:49.422 STDOUT terraform:  + key_pair = "testbed" 2025-08-29 16:47:49.422913 | orchestrator | 16:47:49.422 STDOUT terraform:  + name = "testbed-node-5" 2025-08-29 16:47:49.422921 | orchestrator | 16:47:49.422 STDOUT terraform:  + power_state = "active" 2025-08-29 16:47:49.422931 | orchestrator | 16:47:49.422 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.422975 | orchestrator | 16:47:49.422 STDOUT terraform:  + security_groups = (known after apply) 2025-08-29 16:47:49.422990 | orchestrator | 16:47:49.422 STDOUT terraform:  + stop_before_destroy = false 2025-08-29 16:47:49.423021 | orchestrator | 16:47:49.422 STDOUT terraform:  + updated = (known after apply) 2025-08-29 16:47:49.423094 | orchestrator | 16:47:49.423 STDOUT terraform:  + user_data = "ae09e46b224a6ca206a9ed4f8f8a4f8520827854" 2025-08-29 16:47:49.423103 | orchestrator | 16:47:49.423 STDOUT terraform:  + block_device { 2025-08-29 16:47:49.423118 | orchestrator | 16:47:49.423 STDOUT terraform:  + boot_index = 0 2025-08-29 16:47:49.423185 | orchestrator | 16:47:49.423 STDOUT terraform:  + delete_on_termination = false 2025-08-29 16:47:49.423197 | orchestrator | 16:47:49.423 STDOUT terraform:  + destination_type = "volume" 2025-08-29 16:47:49.423230 | orchestrator | 16:47:49.423 STDOUT terraform:  + multiattach = false 2025-08-29 16:47:49.423284 | orchestrator | 16:47:49.423 STDOUT terraform:  + source_type = "volume" 2025-08-29 16:47:49.423295 | orchestrator | 16:47:49.423 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.423302 | orchestrator | 16:47:49.423 STDOUT terraform:  } 2025-08-29 16:47:49.423311 | orchestrator | 16:47:49.423 STDOUT terraform:  + network { 2025-08-29 16:47:49.423341 | orchestrator | 16:47:49.423 STDOUT terraform:  + access_network = false 2025-08-29 16:47:49.423393 | orchestrator | 16:47:49.423 STDOUT terraform:  + fixed_ip_v4 = (known after apply) 2025-08-29 16:47:49.423402 | orchestrator | 16:47:49.423 STDOUT terraform:  + fixed_ip_v6 = (known after apply) 2025-08-29 16:47:49.423493 | orchestrator | 16:47:49.423 STDOUT terraform:  + mac = (known after apply) 2025-08-29 16:47:49.423502 | orchestrator | 16:47:49.423 STDOUT terraform:  + name = (known after apply) 2025-08-29 16:47:49.423513 | orchestrator | 16:47:49.423 STDOUT terraform:  + port = (known after apply) 2025-08-29 16:47:49.423523 | orchestrator | 16:47:49.423 STDOUT terraform:  + uuid = (known after apply) 2025-08-29 16:47:49.423530 | orchestrator | 16:47:49.423 STDOUT terraform:  } 2025-08-29 16:47:49.423536 | orchestrator | 16:47:49.423 STDOUT terraform:  } 2025-08-29 16:47:49.423582 | orchestrator | 16:47:49.423 STDOUT terraform:  # openstack_compute_keypair_v2.key will be created 2025-08-29 16:47:49.423630 | orchestrator | 16:47:49.423 STDOUT terraform:  + resource "openstack_compute_keypair_v2" "key" { 2025-08-29 16:47:49.423641 | orchestrator | 16:47:49.423 STDOUT terraform:  + fingerprint = (known after apply) 2025-08-29 16:47:49.423697 | orchestrator | 16:47:49.423 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.423707 | orchestrator | 16:47:49.423 STDOUT terraform:  + name = "testbed" 2025-08-29 16:47:49.423716 | orchestrator | 16:47:49.423 STDOUT terraform:  + private_key = (sensitive value) 2025-08-29 16:47:49.423767 | orchestrator | 16:47:49.423 STDOUT terraform:  + public_key = (known after apply) 2025-08-29 16:47:49.423781 | orchestrator | 16:47:49.423 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.423794 | orchestrator | 16:47:49.423 STDOUT terraform:  + user_id = (known after apply) 2025-08-29 16:47:49.423801 | orchestrator | 16:47:49.423 STDOUT terraform:  } 2025-08-29 16:47:49.423909 | orchestrator | 16:47:49.423 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created 2025-08-29 16:47:49.423923 | orchestrator | 16:47:49.423 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.423932 | orchestrator | 16:47:49.423 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.423943 | orchestrator | 16:47:49.423 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.423997 | orchestrator | 16:47:49.423 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.424011 | orchestrator | 16:47:49.423 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.424034 | orchestrator | 16:47:49.423 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.424047 | orchestrator | 16:47:49.424 STDOUT terraform:  } 2025-08-29 16:47:49.424121 | orchestrator | 16:47:49.424 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created 2025-08-29 16:47:49.424136 | orchestrator | 16:47:49.424 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.424201 | orchestrator | 16:47:49.424 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.424249 | orchestrator | 16:47:49.424 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.424288 | orchestrator | 16:47:49.424 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.424302 | orchestrator | 16:47:49.424 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.424354 | orchestrator | 16:47:49.424 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.424366 | orchestrator | 16:47:49.424 STDOUT terraform:  } 2025-08-29 16:47:49.424375 | orchestrator | 16:47:49.424 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created 2025-08-29 16:47:49.424449 | orchestrator | 16:47:49.424 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.424461 | orchestrator | 16:47:49.424 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.424503 | orchestrator | 16:47:49.424 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.424517 | orchestrator | 16:47:49.424 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.424567 | orchestrator | 16:47:49.424 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.424580 | orchestrator | 16:47:49.424 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.424604 | orchestrator | 16:47:49.424 STDOUT terraform:  } 2025-08-29 16:47:49.424679 | orchestrator | 16:47:49.424 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created 2025-08-29 16:47:49.424691 | orchestrator | 16:47:49.424 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.424733 | orchestrator | 16:47:49.424 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.424748 | orchestrator | 16:47:49.424 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.424802 | orchestrator | 16:47:49.424 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.424810 | orchestrator | 16:47:49.424 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.424818 | orchestrator | 16:47:49.424 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.424827 | orchestrator | 16:47:49.424 STDOUT terraform:  } 2025-08-29 16:47:49.424894 | orchestrator | 16:47:49.424 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created 2025-08-29 16:47:49.424943 | orchestrator | 16:47:49.424 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.424954 | orchestrator | 16:47:49.424 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.425025 | orchestrator | 16:47:49.424 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.425034 | orchestrator | 16:47:49.424 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.425042 | orchestrator | 16:47:49.425 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.425087 | orchestrator | 16:47:49.425 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.425095 | orchestrator | 16:47:49.425 STDOUT terraform:  } 2025-08-29 16:47:49.425133 | orchestrator | 16:47:49.425 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created 2025-08-29 16:47:49.425224 | orchestrator | 16:47:49.425 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.425233 | orchestrator | 16:47:49.425 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.425239 | orchestrator | 16:47:49.425 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.425248 | orchestrator | 16:47:49.425 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.425354 | orchestrator | 16:47:49.425 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.425364 | orchestrator | 16:47:49.425 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.425370 | orchestrator | 16:47:49.425 STDOUT terraform:  } 2025-08-29 16:47:49.425405 | orchestrator | 16:47:49.425 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created 2025-08-29 16:47:49.425453 | orchestrator | 16:47:49.425 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.425506 | orchestrator | 16:47:49.425 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.425515 | orchestrator | 16:47:49.425 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.425523 | orchestrator | 16:47:49.425 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.425625 | orchestrator | 16:47:49.425 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.425637 | orchestrator | 16:47:49.425 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.425643 | orchestrator | 16:47:49.425 STDOUT terraform:  } 2025-08-29 16:47:49.425653 | orchestrator | 16:47:49.425 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created 2025-08-29 16:47:49.425719 | orchestrator | 16:47:49.425 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.425735 | orchestrator | 16:47:49.425 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.425779 | orchestrator | 16:47:49.425 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.425790 | orchestrator | 16:47:49.425 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.425831 | orchestrator | 16:47:49.425 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.425845 | orchestrator | 16:47:49.425 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.425852 | orchestrator | 16:47:49.425 STDOUT terraform:  } 2025-08-29 16:47:49.425929 | orchestrator | 16:47:49.425 STDOUT terraform:  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created 2025-08-29 16:47:49.426143 | orchestrator | 16:47:49.425 STDOUT terraform:  + resource "openstack_compute_volume_attach_v2" "node_volume_attachment" { 2025-08-29 16:47:49.426300 | orchestrator | 16:47:49.426 STDOUT terraform:  + device = (known after apply) 2025-08-29 16:47:49.426361 | orchestrator | 16:47:49.426 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.426420 | orchestrator | 16:47:49.426 STDOUT terraform:  + instance_id = (known after apply) 2025-08-29 16:47:49.426478 | orchestrator | 16:47:49.426 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.426547 | orchestrator | 16:47:49.426 STDOUT terraform:  + volume_id = (known after apply) 2025-08-29 16:47:49.426575 | orchestrator | 16:47:49.426 STDOUT terraform:  } 2025-08-29 16:47:49.426724 | orchestrator | 16:47:49.426 STDOUT terraform:  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created 2025-08-29 16:47:49.426861 | orchestrator | 16:47:49.426 STDOUT terraform:  + resource "openstack_networking_floatingip_associate_v2" "manager_floating_ip_association" { 2025-08-29 16:47:49.426921 | orchestrator | 16:47:49.426 STDOUT terraform:  + fixed_ip = (known after apply) 2025-08-29 16:47:49.426978 | orchestrator | 16:47:49.426 STDOUT terraform:  + floating_ip = (known after apply) 2025-08-29 16:47:49.427055 | orchestrator | 16:47:49.426 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.427125 | orchestrator | 16:47:49.427 STDOUT terraform:  + port_id = (known after apply) 2025-08-29 16:47:49.427185 | orchestrator | 16:47:49.427 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.427210 | orchestrator | 16:47:49.427 STDOUT terraform:  } 2025-08-29 16:47:49.427314 | orchestrator | 16:47:49.427 STDOUT terraform:  # openstack_networking_floatingip_v2.manager_floating_ip will be created 2025-08-29 16:47:49.427416 | orchestrator | 16:47:49.427 STDOUT terraform:  + resource "openstack_networking_floatingip_v2" "manager_floating_ip" { 2025-08-29 16:47:49.427468 | orchestrator | 16:47:49.427 STDOUT terraform:  + address = (known after apply) 2025-08-29 16:47:49.427520 | orchestrator | 16:47:49.427 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.427568 | orchestrator | 16:47:49.427 STDOUT terraform:  + dns_domain = (known after apply) 2025-08-29 16:47:49.427666 | orchestrator | 16:47:49.427 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.427733 | orchestrator | 16:47:49.427 STDOUT terraform:  + fixed_ip = (known after apply) 2025-08-29 16:47:49.427779 | orchestrator | 16:47:49.427 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.427815 | orchestrator | 16:47:49.427 STDOUT terraform:  + pool = "public" 2025-08-29 16:47:49.427857 | orchestrator | 16:47:49.427 STDOUT terraform:  + port_id = (known after apply) 2025-08-29 16:47:49.427909 | orchestrator | 16:47:49.427 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.427953 | orchestrator | 16:47:49.427 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.427997 | orchestrator | 16:47:49.427 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.428019 | orchestrator | 16:47:49.427 STDOUT terraform:  } 2025-08-29 16:47:49.428125 | orchestrator | 16:47:49.428 STDOUT terraform:  # openstack_networking_network_v2.net_management will be created 2025-08-29 16:47:49.428208 | orchestrator | 16:47:49.428 STDOUT terraform:  + resource "openstack_networking_network_v2" "net_management" { 2025-08-29 16:47:49.428286 | orchestrator | 16:47:49.428 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.428372 | orchestrator | 16:47:49.428 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.428414 | orchestrator | 16:47:49.428 STDOUT terraform:  + availability_zone_hints = [ 2025-08-29 16:47:49.428439 | orchestrator | 16:47:49.428 STDOUT terraform:  + "nova", 2025-08-29 16:47:49.428461 | orchestrator | 16:47:49.428 STDOUT terraform:  ] 2025-08-29 16:47:49.428527 | orchestrator | 16:47:49.428 STDOUT terraform:  + dns_domain = (known after apply) 2025-08-29 16:47:49.428625 | orchestrator | 16:47:49.428 STDOUT terraform:  + external = (known after apply) 2025-08-29 16:47:49.428691 | orchestrator | 16:47:49.428 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.428765 | orchestrator | 16:47:49.428 STDOUT terraform:  + mtu = (known after apply) 2025-08-29 16:47:49.428838 | orchestrator | 16:47:49.428 STDOUT terraform:  + name = "net-testbed-management" 2025-08-29 16:47:49.428910 | orchestrator | 16:47:49.428 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.428990 | orchestrator | 16:47:49.428 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.429058 | orchestrator | 16:47:49.428 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.429128 | orchestrator | 16:47:49.429 STDOUT terraform:  + shared = (known after apply) 2025-08-29 16:47:49.429200 | orchestrator | 16:47:49.429 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.429266 | orchestrator | 16:47:49.429 STDOUT terraform:  + transparent_vlan = (known after apply) 2025-08-29 16:47:49.429308 | orchestrator | 16:47:49.429 STDOUT terraform:  + segments (known after apply) 2025-08-29 16:47:49.429330 | orchestrator | 16:47:49.429 STDOUT terraform:  } 2025-08-29 16:47:49.429415 | orchestrator | 16:47:49.429 STDOUT terraform:  # openstack_networking_port_v2.manager_port_management will be created 2025-08-29 16:47:49.429510 | orchestrator | 16:47:49.429 STDOUT terraform:  + resource "openstack_networking_port_v2" "manager_port_management" { 2025-08-29 16:47:49.429575 | orchestrator | 16:47:49.429 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.429674 | orchestrator | 16:47:49.429 STDOUT terraform:  + all_fixed_ips = (known after apply) 2025-08-29 16:47:49.429746 | orchestrator | 16:47:49.429 STDOUT terraform:  + all_security_group_ids = (known after apply) 2025-08-29 16:47:49.429813 | orchestrator | 16:47:49.429 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.429879 | orchestrator | 16:47:49.429 STDOUT terraform:  + device_id = (known after apply) 2025-08-29 16:47:49.429955 | orchestrator | 16:47:49.429 STDOUT terraform:  + device_owner = (known after apply) 2025-08-29 16:47:49.434066 | orchestrator | 16:47:49.429 STDOUT terraform:  + dns_assignment = (known after apply) 2025-08-29 16:47:49.434085 | orchestrator | 16:47:49.430 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.434090 | orchestrator | 16:47:49.430 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.434095 | orchestrator | 16:47:49.430 STDOUT terraform:  + mac_address = (known after apply) 2025-08-29 16:47:49.434100 | orchestrator | 16:47:49.430 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.434104 | orchestrator | 16:47:49.430 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.434109 | orchestrator | 16:47:49.430 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.434114 | orchestrator | 16:47:49.430 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.434118 | orchestrator | 16:47:49.430 STDOUT terraform:  + security_group_ids = (known after apply) 2025-08-29 16:47:49.434123 | orchestrator | 16:47:49.430 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.434128 | orchestrator | 16:47:49.430 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434134 | orchestrator | 16:47:49.430 STDOUT terraform:  + ip_address = "192.168.112.0/20" 2025-08-29 16:47:49.434139 | orchestrator | 16:47:49.430 STDOUT terraform:  } 2025-08-29 16:47:49.434143 | orchestrator | 16:47:49.430 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434148 | orchestrator | 16:47:49.430 STDOUT terraform:  + ip_address = "192.168.16.8/20" 2025-08-29 16:47:49.434153 | orchestrator | 16:47:49.430 STDOUT terraform:  } 2025-08-29 16:47:49.434157 | orchestrator | 16:47:49.430 STDOUT terraform:  + binding (known after apply) 2025-08-29 16:47:49.434162 | orchestrator | 16:47:49.430 STDOUT terraform:  + fixed_ip { 2025-08-29 16:47:49.434167 | orchestrator | 16:47:49.430 STDOUT terraform:  + ip_address = "192.168.16.5" 2025-08-29 16:47:49.434172 | orchestrator | 16:47:49.430 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.434176 | orchestrator | 16:47:49.431 STDOUT terraform:  } 2025-08-29 16:47:49.434181 | orchestrator | 16:47:49.431 STDOUT terraform:  } 2025-08-29 16:47:49.434186 | orchestrator | 16:47:49.431 STDOUT terraform:  # openstack_networking_port_v2.node_port_management[0] will be created 2025-08-29 16:47:49.434191 | orchestrator | 16:47:49.431 STDOUT terraform:  + resource "openstack_networking_port_v2" "node_port_management" { 2025-08-29 16:47:49.434195 | orchestrator | 16:47:49.431 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.434200 | orchestrator | 16:47:49.431 STDOUT terraform:  + all_fixed_ips = (known after apply) 2025-08-29 16:47:49.434204 | orchestrator | 16:47:49.431 STDOUT terraform:  + all_security_group_ids = (known after apply) 2025-08-29 16:47:49.434209 | orchestrator | 16:47:49.431 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.434214 | orchestrator | 16:47:49.431 STDOUT terraform:  + device_id = (known after apply) 2025-08-29 16:47:49.434224 | orchestrator | 16:47:49.431 STDOUT terraform:  + device_owner = (known after apply) 2025-08-29 16:47:49.434228 | orchestrator | 16:47:49.431 STDOUT terraform:  + dns_assignment = (known after apply) 2025-08-29 16:47:49.434233 | orchestrator | 16:47:49.431 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.434237 | orchestrator | 16:47:49.431 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.434246 | orchestrator | 16:47:49.431 STDOUT terraform:  + mac_address = (known after apply) 2025-08-29 16:47:49.434251 | orchestrator | 16:47:49.431 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.434255 | orchestrator | 16:47:49.431 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.434260 | orchestrator | 16:47:49.431 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.434271 | orchestrator | 16:47:49.431 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.434276 | orchestrator | 16:47:49.431 STDOUT terraform:  + security_group_ids = (known after apply) 2025-08-29 16:47:49.434280 | orchestrator | 16:47:49.432 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.434285 | orchestrator | 16:47:49.432 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434289 | orchestrator | 16:47:49.432 STDOUT terraform:  + ip_address = "192.168.112.0/20" 2025-08-29 16:47:49.434294 | orchestrator | 16:47:49.432 STDOUT terraform:  } 2025-08-29 16:47:49.434299 | orchestrator | 16:47:49.432 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434303 | orchestrator | 16:47:49.432 STDOUT terraform:  + ip_address = "192.168.16.254/20" 2025-08-29 16:47:49.434308 | orchestrator | 16:47:49.432 STDOUT terraform:  } 2025-08-29 16:47:49.434312 | orchestrator | 16:47:49.432 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434317 | orchestrator | 16:47:49.432 STDOUT terraform:  + ip_address = "192.168.16.8/20" 2025-08-29 16:47:49.434322 | orchestrator | 16:47:49.432 STDOUT terraform:  } 2025-08-29 16:47:49.434326 | orchestrator | 16:47:49.432 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434331 | orchestrator | 16:47:49.432 STDOUT terraform:  + ip_address = "192.168.16.9/20" 2025-08-29 16:47:49.434336 | orchestrator | 16:47:49.432 STDOUT terraform:  } 2025-08-29 16:47:49.434340 | orchestrator | 16:47:49.432 STDOUT terraform:  + binding (known after apply) 2025-08-29 16:47:49.434345 | orchestrator | 16:47:49.432 STDOUT terraform:  + fixed_ip { 2025-08-29 16:47:49.434349 | orchestrator | 16:47:49.432 STDOUT terraform:  + ip_address = "192.168.16.10" 2025-08-29 16:47:49.434354 | orchestrator | 16:47:49.432 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.434358 | orchestrator | 16:47:49.432 STDOUT terraform:  } 2025-08-29 16:47:49.434363 | orchestrator | 16:47:49.432 STDOUT terraform:  } 2025-08-29 16:47:49.434368 | orchestrator | 16:47:49.432 STDOUT terraform:  # openstack_networking_port_v2.node_port_management[1] will be created 2025-08-29 16:47:49.434372 | orchestrator | 16:47:49.432 STDOUT terraform:  + resource "openstack_networking_port_v2" "node_port_management" { 2025-08-29 16:47:49.434380 | orchestrator | 16:47:49.432 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.434385 | orchestrator | 16:47:49.432 STDOUT terraform:  + all_fixed_ips = (known after apply) 2025-08-29 16:47:49.434389 | orchestrator | 16:47:49.432 STDOUT terraform:  + all_security_group_ids = (known after apply) 2025-08-29 16:47:49.434396 | orchestrator | 16:47:49.432 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.434401 | orchestrator | 16:47:49.432 STDOUT terraform:  + device_id = (known after apply) 2025-08-29 16:47:49.434406 | orchestrator | 16:47:49.433 STDOUT terraform:  + device_owner = (known after apply) 2025-08-29 16:47:49.434410 | orchestrator | 16:47:49.433 STDOUT terraform:  + dns_assignment = (known after apply) 2025-08-29 16:47:49.434415 | orchestrator | 16:47:49.433 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.434419 | orchestrator | 16:47:49.433 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.434424 | orchestrator | 16:47:49.433 STDOUT terraform:  + mac_address = (known after apply) 2025-08-29 16:47:49.434428 | orchestrator | 16:47:49.433 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.434433 | orchestrator | 16:47:49.433 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.434437 | orchestrator | 16:47:49.433 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.434442 | orchestrator | 16:47:49.433 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.434446 | orchestrator | 16:47:49.433 STDOUT terraform:  + security_group_ids = (known after apply) 2025-08-29 16:47:49.434454 | orchestrator | 16:47:49.433 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.434459 | orchestrator | 16:47:49.433 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434463 | orchestrator | 16:47:49.433 STDOUT terraform:  + ip_address = "192.168.112.0/20" 2025-08-29 16:47:49.434468 | orchestrator | 16:47:49.433 STDOUT terraform:  } 2025-08-29 16:47:49.434472 | orchestrator | 16:47:49.433 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434477 | orchestrator | 16:47:49.433 STDOUT terraform:  + ip_address = "192.168.16.254/20" 2025-08-29 16:47:49.434481 | orchestrator | 16:47:49.433 STDOUT terraform:  } 2025-08-29 16:47:49.434486 | orchestrator | 16:47:49.433 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434490 | orchestrator | 16:47:49.433 STDOUT terraform:  + ip_address = "192.168.16.8/20" 2025-08-29 16:47:49.434495 | orchestrator | 16:47:49.433 STDOUT terraform:  } 2025-08-29 16:47:49.434499 | orchestrator | 16:47:49.433 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.434504 | orchestrator | 16:47:49.433 STDOUT terraform:  + ip_address = "192.168.16.9/20" 2025-08-29 16:47:49.434509 | orchestrator | 16:47:49.433 STDOUT terraform:  } 2025-08-29 16:47:49.434513 | orchestrator | 16:47:49.433 STDOUT terraform:  + binding (known after apply) 2025-08-29 16:47:49.435347 | orchestrator | 16:47:49.433 STDOUT terraform:  + fixed_ip { 2025-08-29 16:47:49.435399 | orchestrator | 16:47:49.434 STDOUT terraform:  + ip_address = "192.168.16.11" 2025-08-29 16:47:49.435405 | orchestrator | 16:47:49.434 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.435410 | orchestrator | 16:47:49.434 STDOUT terraform:  } 2025-08-29 16:47:49.435414 | orchestrator | 16:47:49.434 STDOUT terraform:  } 2025-08-29 16:47:49.435418 | orchestrator | 16:47:49.434 STDOUT terraform:  # openstack_networking_port_v2.node_port_management[2] will be created 2025-08-29 16:47:49.435423 | orchestrator | 16:47:49.434 STDOUT terraform:  + resource "openstack_networking_port_v2" "node_port_management" { 2025-08-29 16:47:49.435427 | orchestrator | 16:47:49.434 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.435431 | orchestrator | 16:47:49.434 STDOUT terraform:  + all_fixed_ips = (known after apply) 2025-08-29 16:47:49.435435 | orchestrator | 16:47:49.434 STDOUT terraform:  + all_security_group_ids = (known after apply) 2025-08-29 16:47:49.435439 | orchestrator | 16:47:49.434 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.435443 | orchestrator | 16:47:49.435 STDOUT terraform:  + device_id = (known after apply) 2025-08-29 16:47:49.435457 | orchestrator | 16:47:49.435 STDOUT terraform:  + device_owner = (known after apply) 2025-08-29 16:47:49.435461 | orchestrator | 16:47:49.435 STDOUT terraform:  + dns_assignment = (known after apply) 2025-08-29 16:47:49.435465 | orchestrator | 16:47:49.435 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.435469 | orchestrator | 16:47:49.435 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.435473 | orchestrator | 16:47:49.435 STDOUT terraform:  + mac_address = (known after apply) 2025-08-29 16:47:49.435477 | orchestrator | 16:47:49.435 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.435486 | orchestrator | 16:47:49.435 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.435490 | orchestrator | 16:47:49.435 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.435493 | orchestrator | 16:47:49.435 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.435497 | orchestrator | 16:47:49.435 STDOUT terraform:  + security_group_ids = (known after apply) 2025-08-29 16:47:49.435503 | orchestrator | 16:47:49.435 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.435534 | orchestrator | 16:47:49.435 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.435573 | orchestrator | 16:47:49.435 STDOUT terraform:  + ip_address = "192.168.112.0/20" 2025-08-29 16:47:49.435786 | orchestrator | 16:47:49.435 STDOUT terraform:  } 2025-08-29 16:47:49.436006 | orchestrator | 16:47:49.435 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.436046 | orchestrator | 16:47:49.435 STDOUT terraform:  + ip_address = "192.168.16.254/20" 2025-08-29 16:47:49.436074 | orchestrator | 16:47:49.435 STDOUT terraform:  } 2025-08-29 16:47:49.436100 | orchestrator | 16:47:49.435 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.436173 | orchestrator | 16:47:49.435 STDOUT terraform:  + ip_address = "192.168.16.8/20" 2025-08-29 16:47:49.436246 | orchestrator | 16:47:49.435 STDOUT terraform:  } 2025-08-29 16:47:49.436293 | orchestrator | 16:47:49.435 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.436527 | orchestrator | 16:47:49.435 STDOUT terraform:  + ip_address = "192.168.16.9/20" 2025-08-29 16:47:49.436560 | orchestrator | 16:47:49.435 STDOUT terraform:  } 2025-08-29 16:47:49.437153 | orchestrator | 16:47:49.435 STDOUT terraform:  + binding (known after apply) 2025-08-29 16:47:49.437209 | orchestrator | 16:47:49.435 STDOUT terraform:  + fixed_ip { 2025-08-29 16:47:49.437403 | orchestrator | 16:47:49.435 STDOUT terraform:  + ip_address = "192.168.16.12" 2025-08-29 16:47:49.437519 | orchestrator | 16:47:49.435 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.437611 | orchestrator | 16:47:49.435 STDOUT terraform:  } 2025-08-29 16:47:49.437616 | orchestrator | 16:47:49.435 STDOUT terraform:  } 2025-08-29 16:47:49.437690 | orchestrator | 16:47:49.435 STDOUT terraform:  # openstack_networking_port_v2.node_port_management[3] will be created 2025-08-29 16:47:49.437726 | orchestrator | 16:47:49.435 STDOUT terraform:  + resource "openstack_networking_port_v2" "node_port_management" { 2025-08-29 16:47:49.437799 | orchestrator | 16:47:49.435 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.437835 | orchestrator | 16:47:49.436 STDOUT terraform:  + all_fixed_ips = (known after apply) 2025-08-29 16:47:49.437874 | orchestrator | 16:47:49.436 STDOUT terraform:  + all_security_group_ids = (known after apply) 2025-08-29 16:47:49.437951 | orchestrator | 16:47:49.436 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.437998 | orchestrator | 16:47:49.436 STDOUT terraform:  + device_id = (known after apply) 2025-08-29 16:47:49.438067 | orchestrator | 16:47:49.436 STDOUT terraform:  + device_owner = (known after apply) 2025-08-29 16:47:49.442076 | orchestrator | 16:47:49.436 STDOUT terraform:  + dns_assignment = (known after apply) 2025-08-29 16:47:49.442121 | orchestrator | 16:47:49.436 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.442133 | orchestrator | 16:47:49.436 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442137 | orchestrator | 16:47:49.436 STDOUT terraform:  + mac_address = (known after apply) 2025-08-29 16:47:49.442141 | orchestrator | 16:47:49.436 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.442145 | orchestrator | 16:47:49.436 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.442148 | orchestrator | 16:47:49.436 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.442152 | orchestrator | 16:47:49.436 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442156 | orchestrator | 16:47:49.436 STDOUT terraform:  + security_group_ids = (known after apply) 2025-08-29 16:47:49.442160 | orchestrator | 16:47:49.436 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442164 | orchestrator | 16:47:49.436 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442168 | orchestrator | 16:47:49.436 STDOUT terraform:  + ip_address = "192.168.112.0/20" 2025-08-29 16:47:49.442181 | orchestrator | 16:47:49.436 STDOUT terraform:  } 2025-08-29 16:47:49.442185 | orchestrator | 16:47:49.436 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442189 | orchestrator | 16:47:49.436 STDOUT terraform:  + ip_address = "192.168.16.254/20" 2025-08-29 16:47:49.442193 | orchestrator | 16:47:49.436 STDOUT terraform:  } 2025-08-29 16:47:49.442196 | orchestrator | 16:47:49.436 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442200 | orchestrator | 16:47:49.436 STDOUT terraform:  + ip_address = "192.168.16.8/20" 2025-08-29 16:47:49.442204 | orchestrator | 16:47:49.436 STDOUT terraform:  } 2025-08-29 16:47:49.442208 | orchestrator | 16:47:49.436 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442211 | orchestrator | 16:47:49.436 STDOUT terraform:  + ip_address = "192.168.16.9/20" 2025-08-29 16:47:49.442215 | orchestrator | 16:47:49.436 STDOUT terraform:  } 2025-08-29 16:47:49.442219 | orchestrator | 16:47:49.436 STDOUT terraform:  + binding (known after apply) 2025-08-29 16:47:49.442223 | orchestrator | 16:47:49.436 STDOUT terraform:  + fixed_ip { 2025-08-29 16:47:49.442227 | orchestrator | 16:47:49.437 STDOUT terraform:  + ip_address = "192.168.16.13" 2025-08-29 16:47:49.442231 | orchestrator | 16:47:49.437 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.442235 | orchestrator | 16:47:49.437 STDOUT terraform:  } 2025-08-29 16:47:49.442250 | orchestrator | 16:47:49.437 STDOUT terraform:  } 2025-08-29 16:47:49.442254 | orchestrator | 16:47:49.437 STDOUT terraform:  # openstack_networking_port_v2.node_port_management[4] will be created 2025-08-29 16:47:49.442258 | orchestrator | 16:47:49.437 STDOUT terraform:  + resource "openstack_networking_port_v2" "node_port_management" { 2025-08-29 16:47:49.442262 | orchestrator | 16:47:49.437 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.442266 | orchestrator | 16:47:49.437 STDOUT terraform:  + all_fixed_ips = (known after apply) 2025-08-29 16:47:49.442270 | orchestrator | 16:47:49.437 STDOUT terraform:  + all_security_group_ids = (known after apply) 2025-08-29 16:47:49.442273 | orchestrator | 16:47:49.437 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.442277 | orchestrator | 16:47:49.437 STDOUT terraform:  + device_id = (known after apply) 2025-08-29 16:47:49.442281 | orchestrator | 16:47:49.437 STDOUT terraform:  + device_owner = (known after apply) 2025-08-29 16:47:49.442285 | orchestrator | 16:47:49.437 STDOUT terraform:  + dns_assignment = (known after apply) 2025-08-29 16:47:49.442288 | orchestrator | 16:47:49.437 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.442292 | orchestrator | 16:47:49.437 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442296 | orchestrator | 16:47:49.437 STDOUT terraform:  + mac_address = (known after apply) 2025-08-29 16:47:49.442299 | orchestrator | 16:47:49.437 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.442306 | orchestrator | 16:47:49.437 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.442313 | orchestrator | 16:47:49.437 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.442317 | orchestrator | 16:47:49.437 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442320 | orchestrator | 16:47:49.437 STDOUT terraform:  + security_group_ids = (known after apply) 2025-08-29 16:47:49.442324 | orchestrator | 16:47:49.437 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442328 | orchestrator | 16:47:49.437 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442332 | orchestrator | 16:47:49.437 STDOUT terraform:  + ip_address = "192.168.112.0/20" 2025-08-29 16:47:49.442335 | orchestrator | 16:47:49.437 STDOUT terraform:  } 2025-08-29 16:47:49.442339 | orchestrator | 16:47:49.437 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442343 | orchestrator | 16:47:49.437 STDOUT terraform:  + ip_address = "192.168.16.254/20" 2025-08-29 16:47:49.442347 | orchestrator | 16:47:49.437 STDOUT terraform:  } 2025-08-29 16:47:49.442350 | orchestrator | 16:47:49.437 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442354 | orchestrator | 16:47:49.437 STDOUT terraform:  + ip_address = "192.168.16.8/20" 2025-08-29 16:47:49.442358 | orchestrator | 16:47:49.437 STDOUT terraform:  } 2025-08-29 16:47:49.442362 | orchestrator | 16:47:49.437 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442365 | orchestrator | 16:47:49.437 STDOUT terraform:  + ip_address = "192.168.16.9/20" 2025-08-29 16:47:49.442369 | orchestrator | 16:47:49.437 STDOUT terraform:  } 2025-08-29 16:47:49.442373 | orchestrator | 16:47:49.437 STDOUT terraform:  + binding (known after apply) 2025-08-29 16:47:49.442377 | orchestrator | 16:47:49.437 STDOUT terraform:  + fixed_ip { 2025-08-29 16:47:49.442380 | orchestrator | 16:47:49.437 STDOUT terraform:  + ip_address = "192.168.16.14" 2025-08-29 16:47:49.442384 | orchestrator | 16:47:49.437 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.442388 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442392 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442396 | orchestrator | 16:47:49.438 STDOUT terraform:  # openstack_networking_port_v2.node_port_management[5] will be created 2025-08-29 16:47:49.442402 | orchestrator | 16:47:49.438 STDOUT terraform:  + resource "openstack_networking_port_v2" "node_port_management" { 2025-08-29 16:47:49.442406 | orchestrator | 16:47:49.438 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.442410 | orchestrator | 16:47:49.438 STDOUT terraform:  + all_fixed_ips = (known after apply) 2025-08-29 16:47:49.442414 | orchestrator | 16:47:49.438 STDOUT terraform:  + all_security_group_ids = (known after apply) 2025-08-29 16:47:49.442418 | orchestrator | 16:47:49.438 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.442421 | orchestrator | 16:47:49.438 STDOUT terraform:  + device_id = (known after apply) 2025-08-29 16:47:49.442425 | orchestrator | 16:47:49.438 STDOUT terraform:  + device_owner = (known after apply) 2025-08-29 16:47:49.442429 | orchestrator | 16:47:49.438 STDOUT terraform:  + dns_assignment = (known after apply) 2025-08-29 16:47:49.442439 | orchestrator | 16:47:49.438 STDOUT terraform:  + dns_name = (known after apply) 2025-08-29 16:47:49.442443 | orchestrator | 16:47:49.438 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442447 | orchestrator | 16:47:49.438 STDOUT terraform:  + mac_address = (known after apply) 2025-08-29 16:47:49.442450 | orchestrator | 16:47:49.438 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.442454 | orchestrator | 16:47:49.438 STDOUT terraform:  + port_security_enabled = (known after apply) 2025-08-29 16:47:49.442458 | orchestrator | 16:47:49.438 STDOUT terraform:  + qos_policy_id = (known after apply) 2025-08-29 16:47:49.442462 | orchestrator | 16:47:49.438 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442466 | orchestrator | 16:47:49.438 STDOUT terraform:  + security_group_ids = (known after apply) 2025-08-29 16:47:49.442469 | orchestrator | 16:47:49.438 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442473 | orchestrator | 16:47:49.438 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442477 | orchestrator | 16:47:49.438 STDOUT terraform:  + ip_address = "192.168.112.0/20" 2025-08-29 16:47:49.442481 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442484 | orchestrator | 16:47:49.438 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442488 | orchestrator | 16:47:49.438 STDOUT terraform:  + ip_address = "192.168.16.254/20" 2025-08-29 16:47:49.442492 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442496 | orchestrator | 16:47:49.438 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442500 | orchestrator | 16:47:49.438 STDOUT terraform:  + ip_address = "192.168.16.8/20" 2025-08-29 16:47:49.442503 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442507 | orchestrator | 16:47:49.438 STDOUT terraform:  + allowed_address_pairs { 2025-08-29 16:47:49.442511 | orchestrator | 16:47:49.438 STDOUT terraform:  + ip_address = "192.168.16.9/20" 2025-08-29 16:47:49.442515 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442518 | orchestrator | 16:47:49.438 STDOUT terraform:  + binding (known after apply) 2025-08-29 16:47:49.442522 | orchestrator | 16:47:49.438 STDOUT terraform:  + fixed_ip { 2025-08-29 16:47:49.442526 | orchestrator | 16:47:49.438 STDOUT terraform:  + ip_address = "192.168.16.15" 2025-08-29 16:47:49.442530 | orchestrator | 16:47:49.438 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.442534 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442537 | orchestrator | 16:47:49.438 STDOUT terraform:  } 2025-08-29 16:47:49.442541 | orchestrator | 16:47:49.438 STDOUT terraform:  # openstack_networking_router_interface_v2.router_interface will be created 2025-08-29 16:47:49.442545 | orchestrator | 16:47:49.439 STDOUT terraform:  + resource "openstack_networking_router_interface_v2" "router_interface" { 2025-08-29 16:47:49.442549 | orchestrator | 16:47:49.439 STDOUT terraform:  + force_destroy = false 2025-08-29 16:47:49.442552 | orchestrator | 16:47:49.439 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442563 | orchestrator | 16:47:49.439 STDOUT terraform:  + port_id = (known after apply) 2025-08-29 16:47:49.442568 | orchestrator | 16:47:49.439 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442571 | orchestrator | 16:47:49.439 STDOUT terraform:  + router_id = (known after apply) 2025-08-29 16:47:49.442575 | orchestrator | 16:47:49.439 STDOUT terraform:  + subnet_id = (known after apply) 2025-08-29 16:47:49.442579 | orchestrator | 16:47:49.439 STDOUT terraform:  } 2025-08-29 16:47:49.442583 | orchestrator | 16:47:49.439 STDOUT terraform:  # openstack_networking_router_v2.router will be created 2025-08-29 16:47:49.442587 | orchestrator | 16:47:49.439 STDOUT terraform:  + resource "openstack_networking_router_v2" "router" { 2025-08-29 16:47:49.442606 | orchestrator | 16:47:49.439 STDOUT terraform:  + admin_state_up = (known after apply) 2025-08-29 16:47:49.442610 | orchestrator | 16:47:49.439 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.442614 | orchestrator | 16:47:49.439 STDOUT terraform:  + availability_zone_hints = [ 2025-08-29 16:47:49.442620 | orchestrator | 16:47:49.439 STDOUT terraform:  + "nova", 2025-08-29 16:47:49.442624 | orchestrator | 16:47:49.439 STDOUT terraform:  ] 2025-08-29 16:47:49.442628 | orchestrator | 16:47:49.439 STDOUT terraform:  + distributed = (known after apply) 2025-08-29 16:47:49.442632 | orchestrator | 16:47:49.439 STDOUT terraform:  + enable_snat = (known after apply) 2025-08-29 16:47:49.442635 | orchestrator | 16:47:49.439 STDOUT terraform:  + external_network_id = "e6be7364-bfd8-4de7-8120-8f41c69a139a" 2025-08-29 16:47:49.442645 | orchestrator | 16:47:49.439 STDOUT terraform:  + external_qos_policy_id = (known after apply) 2025-08-29 16:47:49.442649 | orchestrator | 16:47:49.439 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442652 | orchestrator | 16:47:49.439 STDOUT terraform:  + name = "testbed" 2025-08-29 16:47:49.442656 | orchestrator | 16:47:49.439 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442660 | orchestrator | 16:47:49.439 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442664 | orchestrator | 16:47:49.439 STDOUT terraform:  + external_fixed_ip (known after apply) 2025-08-29 16:47:49.442668 | orchestrator | 16:47:49.439 STDOUT terraform:  } 2025-08-29 16:47:49.442672 | orchestrator | 16:47:49.439 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created 2025-08-29 16:47:49.442677 | orchestrator | 16:47:49.439 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_management_rule1" { 2025-08-29 16:47:49.442681 | orchestrator | 16:47:49.439 STDOUT terraform:  + description = "ssh" 2025-08-29 16:47:49.442685 | orchestrator | 16:47:49.439 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.442689 | orchestrator | 16:47:49.439 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.442693 | orchestrator | 16:47:49.439 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442697 | orchestrator | 16:47:49.439 STDOUT terraform:  + port_range_max = 22 2025-08-29 16:47:49.442700 | orchestrator | 16:47:49.439 STDOUT terraform:  + port_range_min = 22 2025-08-29 16:47:49.442707 | orchestrator | 16:47:49.439 STDOUT terraform:  + protocol = "tcp" 2025-08-29 16:47:49.442711 | orchestrator | 16:47:49.439 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442715 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.442719 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.442723 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_ip_prefix = "0.0.0.0/0" 2025-08-29 16:47:49.442726 | orchestrator | 16:47:49.440 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.442730 | orchestrator | 16:47:49.440 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442737 | orchestrator | 16:47:49.440 STDOUT terraform:  } 2025-08-29 16:47:49.442741 | orchestrator | 16:47:49.440 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created 2025-08-29 16:47:49.442745 | orchestrator | 16:47:49.440 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_management_rule2" { 2025-08-29 16:47:49.442748 | orchestrator | 16:47:49.440 STDOUT terraform:  + description = "wireguard" 2025-08-29 16:47:49.442752 | orchestrator | 16:47:49.440 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.442756 | orchestrator | 16:47:49.440 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.442760 | orchestrator | 16:47:49.440 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442763 | orchestrator | 16:47:49.440 STDOUT terraform:  + port_range_max = 51820 2025-08-29 16:47:49.442767 | orchestrator | 16:47:49.440 STDOUT terraform:  + port_range_min = 51820 2025-08-29 16:47:49.442771 | orchestrator | 16:47:49.440 STDOUT terraform:  + protocol = "udp" 2025-08-29 16:47:49.442775 | orchestrator | 16:47:49.440 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442778 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.442782 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.442786 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_ip_prefix = "0.0.0.0/0" 2025-08-29 16:47:49.442792 | orchestrator | 16:47:49.440 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.442796 | orchestrator | 16:47:49.440 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442800 | orchestrator | 16:47:49.440 STDOUT terraform:  } 2025-08-29 16:47:49.442803 | orchestrator | 16:47:49.440 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created 2025-08-29 16:47:49.442807 | orchestrator | 16:47:49.440 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_management_rule3" { 2025-08-29 16:47:49.442811 | orchestrator | 16:47:49.440 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.442815 | orchestrator | 16:47:49.440 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.442822 | orchestrator | 16:47:49.440 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442826 | orchestrator | 16:47:49.440 STDOUT terraform:  + protocol = "tcp" 2025-08-29 16:47:49.442830 | orchestrator | 16:47:49.440 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442833 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.442837 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.442841 | orchestrator | 16:47:49.440 STDOUT terraform:  + remote_ip_prefix = "192.168.16.0/20" 2025-08-29 16:47:49.442845 | orchestrator | 16:47:49.441 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.442848 | orchestrator | 16:47:49.441 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442852 | orchestrator | 16:47:49.441 STDOUT terraform:  } 2025-08-29 16:47:49.442856 | orchestrator | 16:47:49.441 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created 2025-08-29 16:47:49.442860 | orchestrator | 16:47:49.441 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_management_rule4" { 2025-08-29 16:47:49.442864 | orchestrator | 16:47:49.441 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.442867 | orchestrator | 16:47:49.441 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.442871 | orchestrator | 16:47:49.441 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442877 | orchestrator | 16:47:49.441 STDOUT terraform:  + protocol = "udp" 2025-08-29 16:47:49.442881 | orchestrator | 16:47:49.441 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442885 | orchestrator | 16:47:49.441 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.442889 | orchestrator | 16:47:49.441 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.442893 | orchestrator | 16:47:49.441 STDOUT terraform:  + remote_ip_prefix = "192.168.16.0/20" 2025-08-29 16:47:49.442896 | orchestrator | 16:47:49.441 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.442900 | orchestrator | 16:47:49.441 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442904 | orchestrator | 16:47:49.441 STDOUT terraform:  } 2025-08-29 16:47:49.442908 | orchestrator | 16:47:49.441 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created 2025-08-29 16:47:49.442911 | orchestrator | 16:47:49.441 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_management_rule5" { 2025-08-29 16:47:49.442915 | orchestrator | 16:47:49.441 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.442919 | orchestrator | 16:47:49.441 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.442923 | orchestrator | 16:47:49.441 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442927 | orchestrator | 16:47:49.441 STDOUT terraform:  + protocol = "icmp" 2025-08-29 16:47:49.442935 | orchestrator | 16:47:49.441 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442939 | orchestrator | 16:47:49.441 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.442943 | orchestrator | 16:47:49.441 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.442947 | orchestrator | 16:47:49.441 STDOUT terraform:  + remote_ip_prefix = "0.0.0.0/0" 2025-08-29 16:47:49.442950 | orchestrator | 16:47:49.441 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.442954 | orchestrator | 16:47:49.441 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.442958 | orchestrator | 16:47:49.441 STDOUT terraform:  } 2025-08-29 16:47:49.442962 | orchestrator | 16:47:49.441 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created 2025-08-29 16:47:49.442966 | orchestrator | 16:47:49.442 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_node_rule1" { 2025-08-29 16:47:49.442969 | orchestrator | 16:47:49.442 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.442973 | orchestrator | 16:47:49.442 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.442977 | orchestrator | 16:47:49.442 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.442981 | orchestrator | 16:47:49.442 STDOUT terraform:  + protocol = "tcp" 2025-08-29 16:47:49.442984 | orchestrator | 16:47:49.442 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.442988 | orchestrator | 16:47:49.442 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.442992 | orchestrator | 16:47:49.442 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.442996 | orchestrator | 16:47:49.442 STDOUT terraform:  + remote_ip_prefix = "0.0.0.0/0" 2025-08-29 16:47:49.442999 | orchestrator | 16:47:49.442 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.443003 | orchestrator | 16:47:49.442 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.443007 | orchestrator | 16:47:49.442 STDOUT terraform:  } 2025-08-29 16:47:49.443011 | orchestrator | 16:47:49.442 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created 2025-08-29 16:47:49.443017 | orchestrator | 16:47:49.442 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_node_rule2" { 2025-08-29 16:47:49.443021 | orchestrator | 16:47:49.442 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.443025 | orchestrator | 16:47:49.442 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.443028 | orchestrator | 16:47:49.442 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.443032 | orchestrator | 16:47:49.442 STDOUT terraform:  + protocol = "udp" 2025-08-29 16:47:49.443036 | orchestrator | 16:47:49.442 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.443040 | orchestrator | 16:47:49.442 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.443044 | orchestrator | 16:47:49.442 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.443050 | orchestrator | 16:47:49.442 STDOUT terraform:  + remote_ip_prefix = "0.0.0.0/0" 2025-08-29 16:47:49.443054 | orchestrator | 16:47:49.442 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.443058 | orchestrator | 16:47:49.442 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.443062 | orchestrator | 16:47:49.442 STDOUT terraform:  } 2025-08-29 16:47:49.443066 | orchestrator | 16:47:49.442 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created 2025-08-29 16:47:49.443069 | orchestrator | 16:47:49.442 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_node_rule3" { 2025-08-29 16:47:49.443073 | orchestrator | 16:47:49.442 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.443077 | orchestrator | 16:47:49.442 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.443081 | orchestrator | 16:47:49.442 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.443085 | orchestrator | 16:47:49.442 STDOUT terraform:  + protocol = "icmp" 2025-08-29 16:47:49.443090 | orchestrator | 16:47:49.442 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.443094 | orchestrator | 16:47:49.443 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.443099 | orchestrator | 16:47:49.443 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.443140 | orchestrator | 16:47:49.443 STDOUT terraform:  + remote_ip_prefix = "0.0.0.0/0" 2025-08-29 16:47:49.443178 | orchestrator | 16:47:49.443 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.443213 | orchestrator | 16:47:49.443 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.443219 | orchestrator | 16:47:49.443 STDOUT terraform:  } 2025-08-29 16:47:49.443280 | orchestrator | 16:47:49.443 STDOUT terraform:  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created 2025-08-29 16:47:49.443329 | orchestrator | 16:47:49.443 STDOUT terraform:  + resource "openstack_networking_secgroup_rule_v2" "security_group_rule_vrrp" { 2025-08-29 16:47:49.443354 | orchestrator | 16:47:49.443 STDOUT terraform:  + description = "vrrp" 2025-08-29 16:47:49.443384 | orchestrator | 16:47:49.443 STDOUT terraform:  + direction = "ingress" 2025-08-29 16:47:49.443409 | orchestrator | 16:47:49.443 STDOUT terraform:  + ethertype = "IPv4" 2025-08-29 16:47:49.443469 | orchestrator | 16:47:49.443 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.443476 | orchestrator | 16:47:49.443 STDOUT terraform:  + protocol = "112" 2025-08-29 16:47:49.443510 | orchestrator | 16:47:49.443 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.443546 | orchestrator | 16:47:49.443 STDOUT terraform:  + remote_address_group_id = (known after apply) 2025-08-29 16:47:49.443585 | orchestrator | 16:47:49.443 STDOUT terraform:  + remote_group_id = (known after apply) 2025-08-29 16:47:49.443639 | orchestrator | 16:47:49.443 STDOUT terraform:  + remote_ip_prefix = "0.0.0.0/0" 2025-08-29 16:47:49.443677 | orchestrator | 16:47:49.443 STDOUT terraform:  + security_group_id = (known after apply) 2025-08-29 16:47:49.443714 | orchestrator | 16:47:49.443 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.443720 | orchestrator | 16:47:49.443 STDOUT terraform:  } 2025-08-29 16:47:49.443743 | orchestrator | 16:47:49.443 STDOUT terraform:  # openstack_networking_secgroup 2025-08-29 16:47:49.443824 | orchestrator | 16:47:49.443 STDOUT terraform: _v2.security_group_management will be created 2025-08-29 16:47:49.443870 | orchestrator | 16:47:49.443 STDOUT terraform:  + resource "openstack_networking_secgroup_v2" "security_group_management" { 2025-08-29 16:47:49.443899 | orchestrator | 16:47:49.443 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.443935 | orchestrator | 16:47:49.443 STDOUT terraform:  + description = "management security group" 2025-08-29 16:47:49.443989 | orchestrator | 16:47:49.443 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.443996 | orchestrator | 16:47:49.443 STDOUT terraform:  + name = "testbed-management" 2025-08-29 16:47:49.444022 | orchestrator | 16:47:49.443 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.444052 | orchestrator | 16:47:49.444 STDOUT terraform:  + stateful = (known after apply) 2025-08-29 16:47:49.444080 | orchestrator | 16:47:49.444 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.444087 | orchestrator | 16:47:49.444 STDOUT terraform:  } 2025-08-29 16:47:49.444140 | orchestrator | 16:47:49.444 STDOUT terraform:  # openstack_networking_secgroup_v2.security_group_node will be created 2025-08-29 16:47:49.444202 | orchestrator | 16:47:49.444 STDOUT terraform:  + resource "openstack_networking_secgroup_v2" "security_group_node" { 2025-08-29 16:47:49.444231 | orchestrator | 16:47:49.444 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.444260 | orchestrator | 16:47:49.444 STDOUT terraform:  + description = "node security group" 2025-08-29 16:47:49.444289 | orchestrator | 16:47:49.444 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.444316 | orchestrator | 16:47:49.444 STDOUT terraform:  + name = "testbed-node" 2025-08-29 16:47:49.444346 | orchestrator | 16:47:49.444 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.444374 | orchestrator | 16:47:49.444 STDOUT terraform:  + stateful = (known after apply) 2025-08-29 16:47:49.444402 | orchestrator | 16:47:49.444 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.444409 | orchestrator | 16:47:49.444 STDOUT terraform:  } 2025-08-29 16:47:49.444458 | orchestrator | 16:47:49.444 STDOUT terraform:  # openstack_networking_subnet_v2.subnet_management will be created 2025-08-29 16:47:49.444505 | orchestrator | 16:47:49.444 STDOUT terraform:  + resource "openstack_networking_subnet_v2" "subnet_management" { 2025-08-29 16:47:49.444537 | orchestrator | 16:47:49.444 STDOUT terraform:  + all_tags = (known after apply) 2025-08-29 16:47:49.444568 | orchestrator | 16:47:49.444 STDOUT terraform:  + cidr = "192.168.16.0/20" 2025-08-29 16:47:49.444587 | orchestrator | 16:47:49.444 STDOUT terraform:  + dns_nameservers = [ 2025-08-29 16:47:49.444604 | orchestrator | 16:47:49.444 STDOUT terraform:  + "8.8.8.8", 2025-08-29 16:47:49.444614 | orchestrator | 16:47:49.444 STDOUT terraform:  + "9.9.9.9", 2025-08-29 16:47:49.444632 | orchestrator | 16:47:49.444 STDOUT terraform:  ] 2025-08-29 16:47:49.444648 | orchestrator | 16:47:49.444 STDOUT terraform:  + enable_dhcp = true 2025-08-29 16:47:49.444679 | orchestrator | 16:47:49.444 STDOUT terraform:  + gateway_ip = (known after apply) 2025-08-29 16:47:49.444710 | orchestrator | 16:47:49.444 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.444727 | orchestrator | 16:47:49.444 STDOUT terraform:  + ip_version = 4 2025-08-29 16:47:49.444759 | orchestrator | 16:47:49.444 STDOUT terraform:  + ipv6_address_mode = (known after apply) 2025-08-29 16:47:49.444789 | orchestrator | 16:47:49.444 STDOUT terraform:  + ipv6_ra_mode = (known after apply) 2025-08-29 16:47:49.444830 | orchestrator | 16:47:49.444 STDOUT terraform:  + name = "subnet-testbed-management" 2025-08-29 16:47:49.444861 | orchestrator | 16:47:49.444 STDOUT terraform:  + network_id = (known after apply) 2025-08-29 16:47:49.444878 | orchestrator | 16:47:49.444 STDOUT terraform:  + no_gateway = false 2025-08-29 16:47:49.444912 | orchestrator | 16:47:49.444 STDOUT terraform:  + region = (known after apply) 2025-08-29 16:47:49.444943 | orchestrator | 16:47:49.444 STDOUT terraform:  + service_types = (known after apply) 2025-08-29 16:47:49.444974 | orchestrator | 16:47:49.444 STDOUT terraform:  + tenant_id = (known after apply) 2025-08-29 16:47:49.444990 | orchestrator | 16:47:49.444 STDOUT terraform:  + allocation_pool { 2025-08-29 16:47:49.445013 | orchestrator | 16:47:49.444 STDOUT terraform:  + end = "192.168.31.250" 2025-08-29 16:47:49.445038 | orchestrator | 16:47:49.445 STDOUT terraform:  + start = "192.168.31.200" 2025-08-29 16:47:49.445044 | orchestrator | 16:47:49.445 STDOUT terraform:  } 2025-08-29 16:47:49.445049 | orchestrator | 16:47:49.445 STDOUT terraform:  } 2025-08-29 16:47:49.445081 | orchestrator | 16:47:49.445 STDOUT terraform:  # terraform_data.image will be created 2025-08-29 16:47:49.445104 | orchestrator | 16:47:49.445 STDOUT terraform:  + resource "terraform_data" "image" { 2025-08-29 16:47:49.445129 | orchestrator | 16:47:49.445 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.445146 | orchestrator | 16:47:49.445 STDOUT terraform:  + input = "Ubuntu 24.04" 2025-08-29 16:47:49.445172 | orchestrator | 16:47:49.445 STDOUT terraform:  + output = (known after apply) 2025-08-29 16:47:49.445178 | orchestrator | 16:47:49.445 STDOUT terraform:  } 2025-08-29 16:47:49.445215 | orchestrator | 16:47:49.445 STDOUT terraform:  # terraform_data.image_node will be created 2025-08-29 16:47:49.445239 | orchestrator | 16:47:49.445 STDOUT terraform:  + resource "terraform_data" "image_node" { 2025-08-29 16:47:49.445266 | orchestrator | 16:47:49.445 STDOUT terraform:  + id = (known after apply) 2025-08-29 16:47:49.445283 | orchestrator | 16:47:49.445 STDOUT terraform:  + input = "Ubuntu 24.04" 2025-08-29 16:47:49.445307 | orchestrator | 16:47:49.445 STDOUT terraform:  + output = (known after apply) 2025-08-29 16:47:49.445313 | orchestrator | 16:47:49.445 STDOUT terraform:  } 2025-08-29 16:47:49.445346 | orchestrator | 16:47:49.445 STDOUT terraform: Plan: 64 to add, 0 to change, 0 to destroy. 2025-08-29 16:47:49.445356 | orchestrator | 16:47:49.445 STDOUT terraform: Changes to Outputs: 2025-08-29 16:47:49.445383 | orchestrator | 16:47:49.445 STDOUT terraform:  + manager_address = (sensitive value) 2025-08-29 16:47:49.445407 | orchestrator | 16:47:49.445 STDOUT terraform:  + private_key = (sensitive value) 2025-08-29 16:47:49.722832 | orchestrator | 16:47:49.722 STDOUT terraform: terraform_data.image: Creating... 2025-08-29 16:47:49.722892 | orchestrator | 16:47:49.722 STDOUT terraform: terraform_data.image_node: Creating... 2025-08-29 16:47:49.722900 | orchestrator | 16:47:49.722 STDOUT terraform: terraform_data.image: Creation complete after 0s [id=57fffb01-1ef5-f5e4-7647-4e3c104ecadf] 2025-08-29 16:47:49.723047 | orchestrator | 16:47:49.722 STDOUT terraform: terraform_data.image_node: Creation complete after 0s [id=37362a65-0acc-bd1b-6db3-8cce75ac46dd] 2025-08-29 16:47:49.739608 | orchestrator | 16:47:49.739 STDOUT terraform: data.openstack_images_image_v2.image: Reading... 2025-08-29 16:47:49.742733 | orchestrator | 16:47:49.740 STDOUT terraform: data.openstack_images_image_v2.image_node: Reading... 2025-08-29 16:47:49.747064 | orchestrator | 16:47:49.746 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[1]: Creating... 2025-08-29 16:47:49.747414 | orchestrator | 16:47:49.746 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[5]: Creating... 2025-08-29 16:47:49.747542 | orchestrator | 16:47:49.747 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[2]: Creating... 2025-08-29 16:47:49.748128 | orchestrator | 16:47:49.747 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[6]: Creating... 2025-08-29 16:47:49.751998 | orchestrator | 16:47:49.751 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[8]: Creating... 2025-08-29 16:47:49.752456 | orchestrator | 16:47:49.752 STDOUT terraform: openstack_networking_network_v2.net_management: Creating... 2025-08-29 16:47:49.752698 | orchestrator | 16:47:49.752 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[0]: Creating... 2025-08-29 16:47:49.766040 | orchestrator | 16:47:49.765 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[7]: Creating... 2025-08-29 16:47:50.206038 | orchestrator | 16:47:50.205 STDOUT terraform: data.openstack_images_image_v2.image: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4] 2025-08-29 16:47:50.206129 | orchestrator | 16:47:50.205 STDOUT terraform: data.openstack_images_image_v2.image_node: Read complete after 0s [id=846820b2-039e-4b42-adad-daf72e0f8ea4] 2025-08-29 16:47:50.216226 | orchestrator | 16:47:50.216 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[3]: Creating... 2025-08-29 16:47:50.221010 | orchestrator | 16:47:50.220 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[4]: Creating... 2025-08-29 16:47:50.774077 | orchestrator | 16:47:50.773 STDOUT terraform: openstack_networking_network_v2.net_management: Creation complete after 1s [id=3030863b-afb3-4c81-b5a0-dd7d948f61f3] 2025-08-29 16:47:50.779682 | orchestrator | 16:47:50.779 STDOUT terraform: openstack_compute_keypair_v2.key: Creating... 2025-08-29 16:47:50.897997 | orchestrator | 16:47:50.897 STDOUT terraform: openstack_compute_keypair_v2.key: Creation complete after 0s [id=testbed] 2025-08-29 16:47:50.907123 | orchestrator | 16:47:50.906 STDOUT terraform: openstack_blockstorage_volume_v3.manager_base_volume[0]: Creating... 2025-08-29 16:47:53.393410 | orchestrator | 16:47:53.392 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 3s [id=588899d3-5e4c-43b2-b3cb-36cecbb176f6] 2025-08-29 16:47:53.395305 | orchestrator | 16:47:53.394 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 3s [id=90c73877-8ee8-469e-bb0b-db8fbc3e9322] 2025-08-29 16:47:53.401535 | orchestrator | 16:47:53.401 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[4]: Creating... 2025-08-29 16:47:53.405988 | orchestrator | 16:47:53.405 STDOUT terraform: openstack_networking_subnet_v2.subnet_management: Creating... 2025-08-29 16:47:53.425772 | orchestrator | 16:47:53.425 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 3s [id=8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f] 2025-08-29 16:47:53.435407 | orchestrator | 16:47:53.435 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[5]: Creating... 2025-08-29 16:47:53.441101 | orchestrator | 16:47:53.440 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 3s [id=8e58fc08-3f1a-44a4-9385-6e01a94c76b6] 2025-08-29 16:47:53.448519 | orchestrator | 16:47:53.448 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[2]: Creating... 2025-08-29 16:47:53.463651 | orchestrator | 16:47:53.463 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 3s [id=c3b7c3a0-9d38-47db-a77f-489f49c79290] 2025-08-29 16:47:53.471080 | orchestrator | 16:47:53.470 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[3]: Creating... 2025-08-29 16:47:53.471570 | orchestrator | 16:47:53.471 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 3s [id=8a1a83e3-79d3-46d9-abf0-648ca1b5a474] 2025-08-29 16:47:53.480158 | orchestrator | 16:47:53.480 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[0]: Creating... 2025-08-29 16:47:53.502864 | orchestrator | 16:47:53.502 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 4s [id=8606d01f-f80c-4f5c-bc07-f0f70fbd69b3] 2025-08-29 16:47:53.513410 | orchestrator | 16:47:53.513 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 4s [id=8cbeb260-c625-4217-b6ca-cc546e4ae7e5] 2025-08-29 16:47:53.514740 | orchestrator | 16:47:53.514 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[1]: Creating... 2025-08-29 16:47:53.530847 | orchestrator | 16:47:53.530 STDOUT terraform: local_file.id_rsa_pub: Creating... 2025-08-29 16:47:53.535786 | orchestrator | 16:47:53.535 STDOUT terraform: local_file.id_rsa_pub: Creation complete after 0s [id=95d1cb3cc57dc032fb44c516fc87f25702f2e739] 2025-08-29 16:47:53.547856 | orchestrator | 16:47:53.547 STDOUT terraform: local_sensitive_file.id_rsa: Creating... 2025-08-29 16:47:53.552095 | orchestrator | 16:47:53.551 STDOUT terraform: local_sensitive_file.id_rsa: Creation complete after 0s [id=0e2346498e2400631fffce2d04d094c09bf9c9c2] 2025-08-29 16:47:53.556494 | orchestrator | 16:47:53.556 STDOUT terraform: openstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=5c881ad7-dcaa-4170-926f-a62d483103a6] 2025-08-29 16:47:54.230321 | orchestrator | 16:47:54.230 STDOUT terraform: openstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 3s [id=8dc8bb9a-5647-425f-98b6-36e4c720b19f] 2025-08-29 16:47:54.435120 | orchestrator | 16:47:54.434 STDOUT terraform: openstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=b0f34010-344a-4ed9-bba2-9872237f2a93] 2025-08-29 16:47:54.447751 | orchestrator | 16:47:54.447 STDOUT terraform: openstack_networking_router_v2.router: Creating... 2025-08-29 16:47:56.793109 | orchestrator | 16:47:56.792 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 4s [id=8099cd87-683e-4d8a-b9af-8a40d3773320] 2025-08-29 16:47:56.847624 | orchestrator | 16:47:56.847 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 4s [id=4f525629-0253-4144-8364-3c913388a34f] 2025-08-29 16:47:56.854228 | orchestrator | 16:47:56.853 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 4s [id=15412d66-e28a-4d80-903c-2982b517c5a1] 2025-08-29 16:47:56.864455 | orchestrator | 16:47:56.864 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 4s [id=ddf3c727-22d6-4195-a504-158fe82cbe5d] 2025-08-29 16:47:56.895058 | orchestrator | 16:47:56.894 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 4s [id=681b03f9-5a09-4bb1-a1f3-7c10614aa46d] 2025-08-29 16:47:56.909175 | orchestrator | 16:47:56.908 STDOUT terraform: openstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=c024de2c-1d0f-4cd2-a007-d9b7894b983c] 2025-08-29 16:47:57.149994 | orchestrator | 16:47:57.149 STDOUT terraform: openstack_networking_router_v2.router: Creation complete after 3s [id=9d6993c9-2bc0-4f79-b4ab-7431acddc365] 2025-08-29 16:47:57.154798 | orchestrator | 16:47:57.154 STDOUT terraform: openstack_networking_secgroup_v2.security_group_node: Creating... 2025-08-29 16:47:57.155836 | orchestrator | 16:47:57.155 STDOUT terraform: openstack_networking_secgroup_v2.security_group_management: Creating... 2025-08-29 16:47:57.156173 | orchestrator | 16:47:57.156 STDOUT terraform: openstack_networking_router_interface_v2.router_interface: Creating... 2025-08-29 16:47:57.336288 | orchestrator | 16:47:57.335 STDOUT terraform: openstack_networking_secgroup_v2.security_group_management: Creation complete after 0s [id=8b1fe255-e278-4731-a7c4-595168f2abdf] 2025-08-29 16:47:57.345855 | orchestrator | 16:47:57.345 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating... 2025-08-29 16:47:57.348811 | orchestrator | 16:47:57.348 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating... 2025-08-29 16:47:57.350164 | orchestrator | 16:47:57.349 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating... 2025-08-29 16:47:57.350570 | orchestrator | 16:47:57.350 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating... 2025-08-29 16:47:57.354872 | orchestrator | 16:47:57.354 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating... 2025-08-29 16:47:57.357802 | orchestrator | 16:47:57.357 STDOUT terraform: openstack_networking_port_v2.manager_port_management: Creating... 2025-08-29 16:47:57.394677 | orchestrator | 16:47:57.394 STDOUT terraform: openstack_networking_secgroup_v2.security_group_node: Creation complete after 0s [id=2e9e3fc0-3d88-4f3e-8344-a7d1e4568997] 2025-08-29 16:47:57.401625 | orchestrator | 16:47:57.401 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating... 2025-08-29 16:47:57.401865 | orchestrator | 16:47:57.401 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating... 2025-08-29 16:47:57.415801 | orchestrator | 16:47:57.415 STDOUT terraform: openstack_networking_port_v2.node_port_management[3]: Creating... 2025-08-29 16:47:57.543465 | orchestrator | 16:47:57.543 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 1s [id=876051c5-822e-432f-ba73-8f86e8d9cacf] 2025-08-29 16:47:57.555703 | orchestrator | 16:47:57.555 STDOUT terraform: openstack_networking_port_v2.node_port_management[0]: Creating... 2025-08-29 16:47:57.556000 | orchestrator | 16:47:57.555 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 1s [id=6985779c-6a71-448f-94c7-457fc72869b0] 2025-08-29 16:47:57.569710 | orchestrator | 16:47:57.569 STDOUT terraform: openstack_networking_port_v2.node_port_management[5]: Creating... 2025-08-29 16:47:57.801810 | orchestrator | 16:47:57.801 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 1s [id=ed2ee91a-0be1-41d3-94d2-2c52a1115f40] 2025-08-29 16:47:57.814442 | orchestrator | 16:47:57.814 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating... 2025-08-29 16:47:57.885949 | orchestrator | 16:47:57.885 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 1s [id=28787a42-e56e-4fcf-a80f-d56696502d4f] 2025-08-29 16:47:57.901787 | orchestrator | 16:47:57.901 STDOUT terraform: openstack_networking_port_v2.node_port_management[2]: Creating... 2025-08-29 16:47:58.093067 | orchestrator | 16:47:58.092 STDOUT terraform: openstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=439b83ae-b76c-43e8-bba2-fd94170abfb8] 2025-08-29 16:47:58.104301 | orchestrator | 16:47:58.103 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 1s [id=4dbe32e2-32d6-40de-a99b-4b1d57ae0762] 2025-08-29 16:47:58.110042 | orchestrator | 16:47:58.109 STDOUT terraform: openstack_networking_port_v2.node_port_management[1]: Creating... 2025-08-29 16:47:58.110092 | orchestrator | 16:47:58.109 STDOUT terraform: openstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=4dac935d-a84d-4be3-830c-60b89606d097] 2025-08-29 16:47:58.116898 | orchestrator | 16:47:58.116 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating... 2025-08-29 16:47:58.120797 | orchestrator | 16:47:58.120 STDOUT terraform: openstack_networking_port_v2.node_port_management[4]: Creating... 2025-08-29 16:47:58.158578 | orchestrator | 16:47:58.158 STDOUT terraform: openstack_networking_port_v2.node_port_management[5]: Creation complete after 0s [id=ff163896-8f2c-47d0-88f7-f5289aa5741f] 2025-08-29 16:47:58.225132 | orchestrator | 16:47:58.224 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 0s [id=c16c45db-653e-4a06-a4d6-87dc63dafc3d] 2025-08-29 16:47:58.253403 | orchestrator | 16:47:58.253 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 1s [id=0b2d3183-1f00-4aab-a0aa-fae498432011] 2025-08-29 16:47:58.335703 | orchestrator | 16:47:58.335 STDOUT terraform: openstack_networking_port_v2.node_port_management[0]: Creation complete after 0s [id=c3682e91-0bc9-4afc-a372-129810224734] 2025-08-29 16:47:58.664131 | orchestrator | 16:47:58.515 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 1s [id=e45b3f0f-b516-4c9a-b96f-1e97d882279d] 2025-08-29 16:47:58.664218 | orchestrator | 16:47:58.516 STDOUT terraform: openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 2s [id=2b5a01bf-259d-425f-971b-c4f1567c21a5] 2025-08-29 16:47:58.664242 | orchestrator | 16:47:58.600 STDOUT terraform: openstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=ca38cb25-bf61-4013-8f49-8127f925ca15] 2025-08-29 16:47:58.664291 | orchestrator | 16:47:58.627 STDOUT terraform: openstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=c6afcf8e-1df2-4630-afe1-b0a00d8cd915] 2025-08-29 16:47:58.702413 | orchestrator | 16:47:58.702 STDOUT terraform: openstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=7de1b8d7-5a09-45f0-967e-92f385f39e56] 2025-08-29 16:47:59.737038 | orchestrator | 16:47:59.736 STDOUT terraform: openstack_networking_router_interface_v2.router_interface: Creation complete after 3s [id=7a9a07df-adc3-4766-90d9-0c8258492e83] 2025-08-29 16:47:59.761929 | orchestrator | 16:47:59.761 STDOUT terraform: openstack_compute_instance_v2.node_server[5]: Creating... 2025-08-29 16:47:59.766004 | orchestrator | 16:47:59.765 STDOUT terraform: openstack_networking_floatingip_v2.manager_floating_ip: Creating... 2025-08-29 16:47:59.770156 | orchestrator | 16:47:59.770 STDOUT terraform: openstack_compute_instance_v2.node_server[1]: Creating... 2025-08-29 16:47:59.772996 | orchestrator | 16:47:59.772 STDOUT terraform: openstack_compute_instance_v2.node_server[0]: Creating... 2025-08-29 16:47:59.781342 | orchestrator | 16:47:59.781 STDOUT terraform: openstack_compute_instance_v2.node_server[3]: Creating... 2025-08-29 16:47:59.786796 | orchestrator | 16:47:59.786 STDOUT terraform: openstack_compute_instance_v2.node_server[2]: Creating... 2025-08-29 16:47:59.790951 | orchestrator | 16:47:59.790 STDOUT terraform: openstack_compute_instance_v2.node_server[4]: Creating... 2025-08-29 16:48:01.192559 | orchestrator | 16:48:01.192 STDOUT terraform: openstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 1s [id=fadbef0d-0584-4092-bc86-6b56b9f400f8] 2025-08-29 16:48:01.206475 | orchestrator | 16:48:01.206 STDOUT terraform: local_file.inventory: Creating... 2025-08-29 16:48:01.209708 | orchestrator | 16:48:01.209 STDOUT terraform: openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating... 2025-08-29 16:48:01.215505 | orchestrator | 16:48:01.215 STDOUT terraform: local_file.MANAGER_ADDRESS: Creating... 2025-08-29 16:48:01.215886 | orchestrator | 16:48:01.215 STDOUT terraform: local_file.inventory: Creation complete after 0s [id=b443f3cf3e05a0c50d62d35c6e77cdc52d38772f] 2025-08-29 16:48:01.219137 | orchestrator | 16:48:01.218 STDOUT terraform: local_file.MANAGER_ADDRESS: Creation complete after 0s [id=b3fe5f348e5b3c4cfaa377c9d90e00f34e90f4e6] 2025-08-29 16:48:01.963097 | orchestrator | 16:48:01.962 STDOUT terraform: openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 1s [id=fadbef0d-0584-4092-bc86-6b56b9f400f8] 2025-08-29 16:48:09.766101 | orchestrator | 16:48:09.765 STDOUT terraform: openstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed] 2025-08-29 16:48:09.777128 | orchestrator | 16:48:09.776 STDOUT terraform: openstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed] 2025-08-29 16:48:09.777206 | orchestrator | 16:48:09.777 STDOUT terraform: openstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed] 2025-08-29 16:48:09.782425 | orchestrator | 16:48:09.782 STDOUT terraform: openstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed] 2025-08-29 16:48:09.788562 | orchestrator | 16:48:09.788 STDOUT terraform: openstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed] 2025-08-29 16:48:09.791787 | orchestrator | 16:48:09.791 STDOUT terraform: openstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed] 2025-08-29 16:48:19.766499 | orchestrator | 16:48:19.766 STDOUT terraform: openstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed] 2025-08-29 16:48:19.777724 | orchestrator | 16:48:19.777 STDOUT terraform: openstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed] 2025-08-29 16:48:19.777792 | orchestrator | 16:48:19.777 STDOUT terraform: openstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed] 2025-08-29 16:48:19.783719 | orchestrator | 16:48:19.783 STDOUT terraform: openstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed] 2025-08-29 16:48:19.789835 | orchestrator | 16:48:19.789 STDOUT terraform: openstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed] 2025-08-29 16:48:19.791815 | orchestrator | 16:48:19.791 STDOUT terraform: openstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed] 2025-08-29 16:48:20.208531 | orchestrator | 16:48:20.208 STDOUT terraform: openstack_compute_instance_v2.node_server[2]: Creation complete after 20s [id=1d35ec6d-7ec1-4d51-8c2a-2e78101693b4] 2025-08-29 16:48:20.224333 | orchestrator | 16:48:20.224 STDOUT terraform: openstack_compute_instance_v2.node_server[4]: Creation complete after 20s [id=18f72a44-59ec-4709-afd6-ef91a757d27d] 2025-08-29 16:48:20.294658 | orchestrator | 16:48:20.294 STDOUT terraform: openstack_compute_instance_v2.node_server[0]: Creation complete after 20s [id=af838305-f6eb-4da4-8214-e19a8131b389] 2025-08-29 16:48:20.348153 | orchestrator | 16:48:20.347 STDOUT terraform: openstack_compute_instance_v2.node_server[3]: Creation complete after 20s [id=e08be3b7-3e1c-4b61-87c0-bbad0dd0ac7a] 2025-08-29 16:48:20.457232 | orchestrator | 16:48:20.456 STDOUT terraform: openstack_compute_instance_v2.node_server[5]: Creation complete after 20s [id=dff09d75-794c-46e3-b40d-0219c30495df] 2025-08-29 16:48:29.777939 | orchestrator | 16:48:29.777 STDOUT terraform: openstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed] 2025-08-29 16:48:30.897375 | orchestrator | 16:48:30.897 STDOUT terraform: openstack_compute_instance_v2.node_server[1]: Creation complete after 31s [id=5698e987-00ce-42c5-a947-313e2b4f6f9f] 2025-08-29 16:48:30.930734 | orchestrator | 16:48:30.930 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating... 2025-08-29 16:48:30.935015 | orchestrator | 16:48:30.934 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating... 2025-08-29 16:48:30.944682 | orchestrator | 16:48:30.944 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating... 2025-08-29 16:48:30.944724 | orchestrator | 16:48:30.944 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating... 2025-08-29 16:48:30.944784 | orchestrator | 16:48:30.944 STDOUT terraform: null_resource.node_semaphore: Creating... 2025-08-29 16:48:30.946441 | orchestrator | 16:48:30.946 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating... 2025-08-29 16:48:30.948325 | orchestrator | 16:48:30.947 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating... 2025-08-29 16:48:30.962063 | orchestrator | 16:48:30.961 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating... 2025-08-29 16:48:30.972360 | orchestrator | 16:48:30.972 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating... 2025-08-29 16:48:30.976592 | orchestrator | 16:48:30.976 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating... 2025-08-29 16:48:30.980326 | orchestrator | 16:48:30.980 STDOUT terraform: null_resource.node_semaphore: Creation complete after 0s [id=5932439781506972546] 2025-08-29 16:48:30.999936 | orchestrator | 16:48:30.999 STDOUT terraform: openstack_compute_instance_v2.manager_server: Creating... 2025-08-29 16:48:34.346584 | orchestrator | 16:48:34.346 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 3s [id=e08be3b7-3e1c-4b61-87c0-bbad0dd0ac7a/8cbeb260-c625-4217-b6ca-cc546e4ae7e5] 2025-08-29 16:48:34.347234 | orchestrator | 16:48:34.346 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 3s [id=18f72a44-59ec-4709-afd6-ef91a757d27d/8a1a83e3-79d3-46d9-abf0-648ca1b5a474] 2025-08-29 16:48:34.361861 | orchestrator | 16:48:34.361 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 3s [id=e08be3b7-3e1c-4b61-87c0-bbad0dd0ac7a/8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f] 2025-08-29 16:48:34.379639 | orchestrator | 16:48:34.379 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 3s [id=dff09d75-794c-46e3-b40d-0219c30495df/c3b7c3a0-9d38-47db-a77f-489f49c79290] 2025-08-29 16:48:34.395783 | orchestrator | 16:48:34.395 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 3s [id=18f72a44-59ec-4709-afd6-ef91a757d27d/8e58fc08-3f1a-44a4-9385-6e01a94c76b6] 2025-08-29 16:48:34.426372 | orchestrator | 16:48:34.425 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 3s [id=dff09d75-794c-46e3-b40d-0219c30495df/5c881ad7-dcaa-4170-926f-a62d483103a6] 2025-08-29 16:48:40.479897 | orchestrator | 16:48:40.479 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 9s [id=18f72a44-59ec-4709-afd6-ef91a757d27d/8606d01f-f80c-4f5c-bc07-f0f70fbd69b3] 2025-08-29 16:48:40.498676 | orchestrator | 16:48:40.498 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 9s [id=dff09d75-794c-46e3-b40d-0219c30495df/588899d3-5e4c-43b2-b3cb-36cecbb176f6] 2025-08-29 16:48:40.526771 | orchestrator | 16:48:40.526 STDOUT terraform: openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 10s [id=e08be3b7-3e1c-4b61-87c0-bbad0dd0ac7a/90c73877-8ee8-469e-bb0b-db8fbc3e9322] 2025-08-29 16:48:41.001027 | orchestrator | 16:48:41.000 STDOUT terraform: openstack_compute_instance_v2.manager_server: Still creating... [10s elapsed] 2025-08-29 16:48:51.002197 | orchestrator | 16:48:51.001 STDOUT terraform: openstack_compute_instance_v2.manager_server: Still creating... [20s elapsed] 2025-08-29 16:48:51.299490 | orchestrator | 16:48:51.299 STDOUT terraform: openstack_compute_instance_v2.manager_server: Creation complete after 20s [id=0c2a4e85-63d5-46a9-85fa-b56ed8d5b0dc] 2025-08-29 16:48:51.335458 | orchestrator | 16:48:51.335 STDOUT terraform: Apply complete! Resources: 64 added, 0 changed, 0 destroyed. 2025-08-29 16:48:51.335541 | orchestrator | 16:48:51.335 STDOUT terraform: Outputs: 2025-08-29 16:48:51.335556 | orchestrator | 16:48:51.335 STDOUT terraform: manager_address = 2025-08-29 16:48:51.335578 | orchestrator | 16:48:51.335 STDOUT terraform: private_key = 2025-08-29 16:48:51.705933 | orchestrator | ok: Runtime: 0:01:07.605524 2025-08-29 16:48:51.739981 | 2025-08-29 16:48:51.740100 | TASK [Fetch manager address] 2025-08-29 16:48:52.194538 | orchestrator | ok 2025-08-29 16:48:52.205289 | 2025-08-29 16:48:52.205421 | TASK [Set manager_host address] 2025-08-29 16:48:52.291134 | orchestrator | ok 2025-08-29 16:48:52.304460 | 2025-08-29 16:48:52.304709 | LOOP [Update ansible collections] 2025-08-29 16:48:53.206160 | orchestrator | [WARNING]: Collection osism.services does not support Ansible version 2.15.2 2025-08-29 16:48:53.206547 | orchestrator | [WARNING]: Collection osism.commons does not support Ansible version 2.15.2 2025-08-29 16:48:53.206656 | orchestrator | Starting galaxy collection install process 2025-08-29 16:48:53.206702 | orchestrator | Process install dependency map 2025-08-29 16:48:53.206738 | orchestrator | Starting collection install process 2025-08-29 16:48:53.206772 | orchestrator | Installing 'osism.commons:999.0.0' to '/home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons' 2025-08-29 16:48:53.206812 | orchestrator | Created collection for osism.commons:999.0.0 at /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons 2025-08-29 16:48:53.206878 | orchestrator | osism.commons:999.0.0 was installed successfully 2025-08-29 16:48:53.206978 | orchestrator | ok: Item: commons Runtime: 0:00:00.555395 2025-08-29 16:48:54.015449 | orchestrator | [WARNING]: Collection osism.services does not support Ansible version 2.15.2 2025-08-29 16:48:54.015720 | orchestrator | [WARNING]: Collection osism.commons does not support Ansible version 2.15.2 2025-08-29 16:48:54.015793 | orchestrator | Starting galaxy collection install process 2025-08-29 16:48:54.015846 | orchestrator | Process install dependency map 2025-08-29 16:48:54.015919 | orchestrator | Starting collection install process 2025-08-29 16:48:54.015969 | orchestrator | Installing 'osism.services:999.0.0' to '/home/zuul-testbed04/.ansible/collections/ansible_collections/osism/services' 2025-08-29 16:48:54.016018 | orchestrator | Created collection for osism.services:999.0.0 at /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/services 2025-08-29 16:48:54.016062 | orchestrator | osism.services:999.0.0 was installed successfully 2025-08-29 16:48:54.016131 | orchestrator | ok: Item: services Runtime: 0:00:00.543695 2025-08-29 16:48:54.038083 | 2025-08-29 16:48:54.038258 | TASK [Wait up to 300 seconds for port 22 to become open and contain "OpenSSH"] 2025-08-29 16:49:04.612516 | orchestrator | ok 2025-08-29 16:49:04.623514 | 2025-08-29 16:49:04.623688 | TASK [Wait a little longer for the manager so that everything is ready] 2025-08-29 16:50:04.675104 | orchestrator | ok 2025-08-29 16:50:04.686445 | 2025-08-29 16:50:04.686578 | TASK [Fetch manager ssh hostkey] 2025-08-29 16:50:06.276622 | orchestrator | Output suppressed because no_log was given 2025-08-29 16:50:06.291535 | 2025-08-29 16:50:06.291760 | TASK [Get ssh keypair from terraform environment] 2025-08-29 16:50:06.827590 | orchestrator | ok: Runtime: 0:00:00.010673 2025-08-29 16:50:06.838805 | 2025-08-29 16:50:06.838962 | TASK [Point out that the following task takes some time and does not give any output] 2025-08-29 16:50:06.889968 | orchestrator | ok: The task 'Run manager part 0' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minutes for this task to complete. 2025-08-29 16:50:06.901054 | 2025-08-29 16:50:06.901194 | TASK [Run manager part 0] 2025-08-29 16:50:08.469077 | orchestrator | [WARNING]: Collection osism.commons does not support Ansible version 2.15.2 2025-08-29 16:50:08.512956 | orchestrator | 2025-08-29 16:50:08.513043 | orchestrator | PLAY [Wait for cloud-init to finish] ******************************************* 2025-08-29 16:50:08.513061 | orchestrator | 2025-08-29 16:50:08.513091 | orchestrator | TASK [Check /var/lib/cloud/instance/boot-finished] ***************************** 2025-08-29 16:50:10.332533 | orchestrator | ok: [testbed-manager] 2025-08-29 16:50:10.332586 | orchestrator | 2025-08-29 16:50:10.332606 | orchestrator | PLAY [Run manager part 0] ****************************************************** 2025-08-29 16:50:10.332616 | orchestrator | 2025-08-29 16:50:10.332674 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 16:50:12.160442 | orchestrator | ok: [testbed-manager] 2025-08-29 16:50:12.160582 | orchestrator | 2025-08-29 16:50:12.160598 | orchestrator | TASK [Get home directory of ansible user] ************************************** 2025-08-29 16:50:12.863388 | orchestrator | ok: [testbed-manager] 2025-08-29 16:50:12.863453 | orchestrator | 2025-08-29 16:50:12.863467 | orchestrator | TASK [Set repo_path fact] ****************************************************** 2025-08-29 16:50:12.917063 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:50:12.917120 | orchestrator | 2025-08-29 16:50:12.917133 | orchestrator | TASK [Update package cache] **************************************************** 2025-08-29 16:50:12.948507 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:50:12.948554 | orchestrator | 2025-08-29 16:50:12.948561 | orchestrator | TASK [Install required packages] *********************************************** 2025-08-29 16:50:12.972246 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:50:12.972286 | orchestrator | 2025-08-29 16:50:12.972291 | orchestrator | TASK [Remove some python packages] ********************************************* 2025-08-29 16:50:12.997269 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:50:12.997327 | orchestrator | 2025-08-29 16:50:12.997336 | orchestrator | TASK [Set venv_command fact (RedHat)] ****************************************** 2025-08-29 16:50:13.031156 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:50:13.031212 | orchestrator | 2025-08-29 16:50:13.031221 | orchestrator | TASK [Fail if Ubuntu version is lower than 22.04] ****************************** 2025-08-29 16:50:13.060218 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:50:13.060268 | orchestrator | 2025-08-29 16:50:13.060276 | orchestrator | TASK [Fail if Debian version is lower than 12] ********************************* 2025-08-29 16:50:13.091428 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:50:13.091473 | orchestrator | 2025-08-29 16:50:13.091481 | orchestrator | TASK [Set APT options on manager] ********************************************** 2025-08-29 16:50:13.823127 | orchestrator | changed: [testbed-manager] 2025-08-29 16:50:13.824203 | orchestrator | 2025-08-29 16:50:13.824221 | orchestrator | TASK [Update APT cache and run dist-upgrade] *********************************** 2025-08-29 16:52:46.085981 | orchestrator | changed: [testbed-manager] 2025-08-29 16:52:46.086104 | orchestrator | 2025-08-29 16:52:46.086124 | orchestrator | TASK [Install HWE kernel package on Ubuntu] ************************************ 2025-08-29 16:53:59.268465 | orchestrator | changed: [testbed-manager] 2025-08-29 16:53:59.268560 | orchestrator | 2025-08-29 16:53:59.268575 | orchestrator | TASK [Install required packages] *********************************************** 2025-08-29 16:54:21.162706 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:21.162798 | orchestrator | 2025-08-29 16:54:21.162816 | orchestrator | TASK [Remove some python packages] ********************************************* 2025-08-29 16:54:29.482169 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:29.482211 | orchestrator | 2025-08-29 16:54:29.482219 | orchestrator | TASK [Set venv_command fact (Debian)] ****************************************** 2025-08-29 16:54:29.527670 | orchestrator | ok: [testbed-manager] 2025-08-29 16:54:29.527768 | orchestrator | 2025-08-29 16:54:29.527783 | orchestrator | TASK [Get current user] ******************************************************** 2025-08-29 16:54:30.315970 | orchestrator | ok: [testbed-manager] 2025-08-29 16:54:30.316059 | orchestrator | 2025-08-29 16:54:30.316077 | orchestrator | TASK [Create venv directory] *************************************************** 2025-08-29 16:54:30.999844 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:30.999888 | orchestrator | 2025-08-29 16:54:30.999897 | orchestrator | TASK [Install netaddr in venv] ************************************************* 2025-08-29 16:54:36.848119 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:36.848200 | orchestrator | 2025-08-29 16:54:36.848237 | orchestrator | TASK [Install ansible-core in venv] ******************************************** 2025-08-29 16:54:42.218223 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:42.218294 | orchestrator | 2025-08-29 16:54:42.218309 | orchestrator | TASK [Install requests >= 2.32.2] ********************************************** 2025-08-29 16:54:44.775817 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:44.775912 | orchestrator | 2025-08-29 16:54:44.775938 | orchestrator | TASK [Install docker >= 7.1.0] ************************************************* 2025-08-29 16:54:46.520522 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:46.520601 | orchestrator | 2025-08-29 16:54:46.520616 | orchestrator | TASK [Create directories in /opt/src] ****************************************** 2025-08-29 16:54:47.547056 | orchestrator | changed: [testbed-manager] => (item=osism/ansible-collection-commons) 2025-08-29 16:54:47.547103 | orchestrator | changed: [testbed-manager] => (item=osism/ansible-collection-services) 2025-08-29 16:54:47.547110 | orchestrator | 2025-08-29 16:54:47.547116 | orchestrator | TASK [Sync sources in /opt/src] ************************************************ 2025-08-29 16:54:47.585508 | orchestrator | [DEPRECATION WARNING]: The connection's stdin object is deprecated. Call 2025-08-29 16:54:47.585576 | orchestrator | display.prompt_until(msg) instead. This feature will be removed in version 2025-08-29 16:54:47.585591 | orchestrator | 2.19. Deprecation warnings can be disabled by setting 2025-08-29 16:54:47.585603 | orchestrator | deprecation_warnings=False in ansible.cfg. 2025-08-29 16:54:50.764084 | orchestrator | changed: [testbed-manager] => (item=osism/ansible-collection-commons) 2025-08-29 16:54:50.764146 | orchestrator | changed: [testbed-manager] => (item=osism/ansible-collection-services) 2025-08-29 16:54:50.764152 | orchestrator | 2025-08-29 16:54:50.764157 | orchestrator | TASK [Create /usr/share/ansible directory] ************************************* 2025-08-29 16:54:51.302859 | orchestrator | changed: [testbed-manager] 2025-08-29 16:54:51.302925 | orchestrator | 2025-08-29 16:54:51.302936 | orchestrator | TASK [Install collections from Ansible galaxy] ********************************* 2025-08-29 16:56:11.535002 | orchestrator | changed: [testbed-manager] => (item=ansible.netcommon) 2025-08-29 16:56:11.535087 | orchestrator | changed: [testbed-manager] => (item=ansible.posix) 2025-08-29 16:56:11.535104 | orchestrator | changed: [testbed-manager] => (item=community.docker>=3.10.2) 2025-08-29 16:56:11.535115 | orchestrator | 2025-08-29 16:56:11.535126 | orchestrator | TASK [Install local collections] *********************************************** 2025-08-29 16:56:13.761943 | orchestrator | changed: [testbed-manager] => (item=ansible-collection-commons) 2025-08-29 16:56:13.761977 | orchestrator | changed: [testbed-manager] => (item=ansible-collection-services) 2025-08-29 16:56:13.761982 | orchestrator | 2025-08-29 16:56:13.761986 | orchestrator | PLAY [Create operator user] **************************************************** 2025-08-29 16:56:13.761991 | orchestrator | 2025-08-29 16:56:13.761995 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 16:56:15.186731 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:15.186847 | orchestrator | 2025-08-29 16:56:15.186866 | orchestrator | TASK [osism.commons.operator : Gather variables for each operating system] ***** 2025-08-29 16:56:15.235720 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:15.235796 | orchestrator | 2025-08-29 16:56:15.235809 | orchestrator | TASK [osism.commons.operator : Set operator_groups variable to default value] *** 2025-08-29 16:56:15.299461 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:15.299515 | orchestrator | 2025-08-29 16:56:15.299522 | orchestrator | TASK [osism.commons.operator : Create operator group] ************************** 2025-08-29 16:56:16.072929 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:16.073662 | orchestrator | 2025-08-29 16:56:16.073697 | orchestrator | TASK [osism.commons.operator : Create user] ************************************ 2025-08-29 16:56:16.762743 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:16.762850 | orchestrator | 2025-08-29 16:56:16.762867 | orchestrator | TASK [osism.commons.operator : Add user to additional groups] ****************** 2025-08-29 16:56:18.070446 | orchestrator | changed: [testbed-manager] => (item=adm) 2025-08-29 16:56:18.070505 | orchestrator | changed: [testbed-manager] => (item=sudo) 2025-08-29 16:56:18.070517 | orchestrator | 2025-08-29 16:56:18.070544 | orchestrator | TASK [osism.commons.operator : Copy user sudoers file] ************************* 2025-08-29 16:56:19.411919 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:19.412028 | orchestrator | 2025-08-29 16:56:19.412044 | orchestrator | TASK [osism.commons.operator : Set language variables in .bashrc configuration file] *** 2025-08-29 16:56:21.110633 | orchestrator | changed: [testbed-manager] => (item=export LANGUAGE=C.UTF-8) 2025-08-29 16:56:21.110708 | orchestrator | changed: [testbed-manager] => (item=export LANG=C.UTF-8) 2025-08-29 16:56:21.110723 | orchestrator | changed: [testbed-manager] => (item=export LC_ALL=C.UTF-8) 2025-08-29 16:56:21.110733 | orchestrator | 2025-08-29 16:56:21.110743 | orchestrator | TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] *** 2025-08-29 16:56:21.162698 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:21.162746 | orchestrator | 2025-08-29 16:56:21.162752 | orchestrator | TASK [osism.commons.operator : Create .ssh directory] ************************** 2025-08-29 16:56:21.691526 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:21.691566 | orchestrator | 2025-08-29 16:56:21.691576 | orchestrator | TASK [osism.commons.operator : Check number of SSH authorized keys] ************ 2025-08-29 16:56:21.770375 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:21.770413 | orchestrator | 2025-08-29 16:56:21.770420 | orchestrator | TASK [osism.commons.operator : Set ssh authorized keys] ************************ 2025-08-29 16:56:22.601288 | orchestrator | changed: [testbed-manager] => (item=None) 2025-08-29 16:56:22.601331 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:22.601340 | orchestrator | 2025-08-29 16:56:22.601347 | orchestrator | TASK [osism.commons.operator : Delete ssh authorized keys] ********************* 2025-08-29 16:56:22.640061 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:22.640100 | orchestrator | 2025-08-29 16:56:22.640108 | orchestrator | TASK [osism.commons.operator : Set authorized GitHub accounts] ***************** 2025-08-29 16:56:22.676942 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:22.677007 | orchestrator | 2025-08-29 16:56:22.677021 | orchestrator | TASK [osism.commons.operator : Delete authorized GitHub accounts] ************** 2025-08-29 16:56:22.716550 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:22.716585 | orchestrator | 2025-08-29 16:56:22.716593 | orchestrator | TASK [osism.commons.operator : Set password] *********************************** 2025-08-29 16:56:22.767299 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:22.767334 | orchestrator | 2025-08-29 16:56:22.767469 | orchestrator | TASK [osism.commons.operator : Unset & lock password] ************************** 2025-08-29 16:56:23.456157 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:23.456192 | orchestrator | 2025-08-29 16:56:23.456198 | orchestrator | PLAY [Run manager part 0] ****************************************************** 2025-08-29 16:56:23.456203 | orchestrator | 2025-08-29 16:56:23.456207 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 16:56:24.835970 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:24.836051 | orchestrator | 2025-08-29 16:56:24.836067 | orchestrator | TASK [Recursively change ownership of /opt/venv] ******************************* 2025-08-29 16:56:25.778183 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:25.778266 | orchestrator | 2025-08-29 16:56:25.778283 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 16:56:25.778297 | orchestrator | testbed-manager : ok=33 changed=23 unreachable=0 failed=0 skipped=13 rescued=0 ignored=0 2025-08-29 16:56:25.778309 | orchestrator | 2025-08-29 16:56:26.190039 | orchestrator | ok: Runtime: 0:06:18.655950 2025-08-29 16:56:26.208132 | 2025-08-29 16:56:26.208258 | TASK [Point out that the log in on the manager is now possible] 2025-08-29 16:56:26.250454 | orchestrator | ok: It is now already possible to log in to the manager with 'make login'. 2025-08-29 16:56:26.259053 | 2025-08-29 16:56:26.259174 | TASK [Point out that the following task takes some time and does not give any output] 2025-08-29 16:56:26.298535 | orchestrator | ok: The task 'Run manager part 1 + 2' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minuts for this task to complete. 2025-08-29 16:56:26.309281 | 2025-08-29 16:56:26.309422 | TASK [Run manager part 1 + 2] 2025-08-29 16:56:27.095838 | orchestrator | [WARNING]: Collection osism.commons does not support Ansible version 2.15.2 2025-08-29 16:56:27.146449 | orchestrator | 2025-08-29 16:56:27.146531 | orchestrator | PLAY [Run manager part 1] ****************************************************** 2025-08-29 16:56:27.146548 | orchestrator | 2025-08-29 16:56:27.146578 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 16:56:29.991025 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:29.991122 | orchestrator | 2025-08-29 16:56:29.991176 | orchestrator | TASK [Set venv_command fact (RedHat)] ****************************************** 2025-08-29 16:56:30.024588 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:30.024659 | orchestrator | 2025-08-29 16:56:30.024681 | orchestrator | TASK [Set venv_command fact (Debian)] ****************************************** 2025-08-29 16:56:30.057829 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:30.057898 | orchestrator | 2025-08-29 16:56:30.057921 | orchestrator | TASK [osism.commons.repository : Gather variables for each operating system] *** 2025-08-29 16:56:30.094194 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:30.094266 | orchestrator | 2025-08-29 16:56:30.094282 | orchestrator | TASK [osism.commons.repository : Set repository_default fact to default value] *** 2025-08-29 16:56:30.156755 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:30.156855 | orchestrator | 2025-08-29 16:56:30.156873 | orchestrator | TASK [osism.commons.repository : Set repositories to default] ****************** 2025-08-29 16:56:30.214491 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:30.214571 | orchestrator | 2025-08-29 16:56:30.214589 | orchestrator | TASK [osism.commons.repository : Include distribution specific repository tasks] *** 2025-08-29 16:56:30.253544 | orchestrator | included: /home/zuul-testbed04/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager 2025-08-29 16:56:30.253614 | orchestrator | 2025-08-29 16:56:30.253628 | orchestrator | TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] ***** 2025-08-29 16:56:30.931702 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:30.931799 | orchestrator | 2025-08-29 16:56:30.931817 | orchestrator | TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] ************* 2025-08-29 16:56:30.980151 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:30.980219 | orchestrator | 2025-08-29 16:56:30.980233 | orchestrator | TASK [osism.commons.repository : Copy 99osism apt configuration] *************** 2025-08-29 16:56:32.330426 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:32.330517 | orchestrator | 2025-08-29 16:56:32.330535 | orchestrator | TASK [osism.commons.repository : Remove sources.list file] ********************* 2025-08-29 16:56:32.901035 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:32.901123 | orchestrator | 2025-08-29 16:56:32.901167 | orchestrator | TASK [osism.commons.repository : Copy ubuntu.sources file] ********************* 2025-08-29 16:56:33.999883 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:33.999945 | orchestrator | 2025-08-29 16:56:33.999961 | orchestrator | TASK [osism.commons.repository : Update package cache] ************************* 2025-08-29 16:56:50.516574 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:50.516671 | orchestrator | 2025-08-29 16:56:50.516688 | orchestrator | TASK [Get home directory of ansible user] ************************************** 2025-08-29 16:56:51.183399 | orchestrator | ok: [testbed-manager] 2025-08-29 16:56:51.183488 | orchestrator | 2025-08-29 16:56:51.183505 | orchestrator | TASK [Set repo_path fact] ****************************************************** 2025-08-29 16:56:51.238656 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:56:51.238731 | orchestrator | 2025-08-29 16:56:51.238745 | orchestrator | TASK [Copy SSH public key] ***************************************************** 2025-08-29 16:56:52.165549 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:52.165620 | orchestrator | 2025-08-29 16:56:52.165631 | orchestrator | TASK [Copy SSH private key] **************************************************** 2025-08-29 16:56:53.137342 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:53.137424 | orchestrator | 2025-08-29 16:56:53.137442 | orchestrator | TASK [Create configuration directory] ****************************************** 2025-08-29 16:56:53.695051 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:53.695132 | orchestrator | 2025-08-29 16:56:53.695148 | orchestrator | TASK [Copy testbed repo] ******************************************************* 2025-08-29 16:56:53.733435 | orchestrator | [DEPRECATION WARNING]: The connection's stdin object is deprecated. Call 2025-08-29 16:56:53.733524 | orchestrator | display.prompt_until(msg) instead. This feature will be removed in version 2025-08-29 16:56:53.733539 | orchestrator | 2.19. Deprecation warnings can be disabled by setting 2025-08-29 16:56:53.733551 | orchestrator | deprecation_warnings=False in ansible.cfg. 2025-08-29 16:56:55.633659 | orchestrator | changed: [testbed-manager] 2025-08-29 16:56:55.633752 | orchestrator | 2025-08-29 16:56:55.633769 | orchestrator | TASK [Install python requirements in venv] ************************************* 2025-08-29 16:57:04.246251 | orchestrator | ok: [testbed-manager] => (item=Jinja2) 2025-08-29 16:57:04.246357 | orchestrator | ok: [testbed-manager] => (item=PyYAML) 2025-08-29 16:57:04.246386 | orchestrator | ok: [testbed-manager] => (item=packaging) 2025-08-29 16:57:04.246406 | orchestrator | changed: [testbed-manager] => (item=python-gilt==1.2.3) 2025-08-29 16:57:04.246426 | orchestrator | ok: [testbed-manager] => (item=requests>=2.32.2) 2025-08-29 16:57:04.246438 | orchestrator | ok: [testbed-manager] => (item=docker>=7.1.0) 2025-08-29 16:57:04.246449 | orchestrator | 2025-08-29 16:57:04.246461 | orchestrator | TASK [Copy testbed custom CA certificate on Debian/Ubuntu] ********************* 2025-08-29 16:57:05.247517 | orchestrator | changed: [testbed-manager] 2025-08-29 16:57:05.247606 | orchestrator | 2025-08-29 16:57:05.247624 | orchestrator | TASK [Copy testbed custom CA certificate on CentOS] **************************** 2025-08-29 16:57:05.286032 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:57:05.286088 | orchestrator | 2025-08-29 16:57:05.286097 | orchestrator | TASK [Run update-ca-certificates on Debian/Ubuntu] ***************************** 2025-08-29 16:57:08.330432 | orchestrator | changed: [testbed-manager] 2025-08-29 16:57:08.331121 | orchestrator | 2025-08-29 16:57:08.331141 | orchestrator | TASK [Run update-ca-trust on RedHat] ******************************************* 2025-08-29 16:57:08.369974 | orchestrator | skipping: [testbed-manager] 2025-08-29 16:57:08.370179 | orchestrator | 2025-08-29 16:57:08.370197 | orchestrator | TASK [Run manager part 2] ****************************************************** 2025-08-29 16:58:41.430530 | orchestrator | changed: [testbed-manager] 2025-08-29 16:58:41.430645 | orchestrator | 2025-08-29 16:58:41.430663 | orchestrator | RUNNING HANDLER [osism.commons.repository : Force update of package cache] ***** 2025-08-29 16:58:42.479532 | orchestrator | ok: [testbed-manager] 2025-08-29 16:58:42.479621 | orchestrator | 2025-08-29 16:58:42.479639 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 16:58:42.479653 | orchestrator | testbed-manager : ok=21 changed=11 unreachable=0 failed=0 skipped=5 rescued=0 ignored=0 2025-08-29 16:58:42.479666 | orchestrator | 2025-08-29 16:58:42.920792 | orchestrator | ok: Runtime: 0:02:15.979995 2025-08-29 16:58:42.937352 | 2025-08-29 16:58:42.937492 | TASK [Reboot manager] 2025-08-29 16:58:44.478157 | orchestrator | ok: Runtime: 0:00:00.950371 2025-08-29 16:58:44.495927 | 2025-08-29 16:58:44.496080 | TASK [Wait up to 300 seconds for port 22 to become open and contain "OpenSSH"] 2025-08-29 16:58:59.138084 | orchestrator | ok 2025-08-29 16:58:59.153018 | 2025-08-29 16:58:59.153244 | TASK [Wait a little longer for the manager so that everything is ready] 2025-08-29 16:59:59.201280 | orchestrator | ok 2025-08-29 16:59:59.212196 | 2025-08-29 16:59:59.212355 | TASK [Deploy manager + bootstrap nodes] 2025-08-29 17:00:01.684673 | orchestrator | 2025-08-29 17:00:01.684792 | orchestrator | # DEPLOY MANAGER 2025-08-29 17:00:01.684802 | orchestrator | 2025-08-29 17:00:01.684808 | orchestrator | + set -e 2025-08-29 17:00:01.684814 | orchestrator | + echo 2025-08-29 17:00:01.684821 | orchestrator | + echo '# DEPLOY MANAGER' 2025-08-29 17:00:01.684828 | orchestrator | + echo 2025-08-29 17:00:01.684851 | orchestrator | + cat /opt/manager-vars.sh 2025-08-29 17:00:01.686720 | orchestrator | export NUMBER_OF_NODES=6 2025-08-29 17:00:01.686732 | orchestrator | 2025-08-29 17:00:01.686737 | orchestrator | export CEPH_VERSION=reef 2025-08-29 17:00:01.686743 | orchestrator | export CONFIGURATION_VERSION=main 2025-08-29 17:00:01.686748 | orchestrator | export MANAGER_VERSION=9.2.0 2025-08-29 17:00:01.686758 | orchestrator | export OPENSTACK_VERSION=2024.2 2025-08-29 17:00:01.686763 | orchestrator | 2025-08-29 17:00:01.686770 | orchestrator | export ARA=false 2025-08-29 17:00:01.686775 | orchestrator | export DEPLOY_MODE=manager 2025-08-29 17:00:01.686781 | orchestrator | export TEMPEST=false 2025-08-29 17:00:01.686786 | orchestrator | export IS_ZUUL=true 2025-08-29 17:00:01.686790 | orchestrator | 2025-08-29 17:00:01.686798 | orchestrator | export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 17:00:01.686803 | orchestrator | export EXTERNAL_API=false 2025-08-29 17:00:01.686807 | orchestrator | 2025-08-29 17:00:01.686811 | orchestrator | export IMAGE_USER=ubuntu 2025-08-29 17:00:01.686830 | orchestrator | export IMAGE_NODE_USER=ubuntu 2025-08-29 17:00:01.686835 | orchestrator | 2025-08-29 17:00:01.686839 | orchestrator | export CEPH_STACK=ceph-ansible 2025-08-29 17:00:01.686868 | orchestrator | 2025-08-29 17:00:01.686874 | orchestrator | + echo 2025-08-29 17:00:01.686879 | orchestrator | + source /opt/configuration/scripts/include.sh 2025-08-29 17:00:01.688625 | orchestrator | ++ export INTERACTIVE=false 2025-08-29 17:00:01.688633 | orchestrator | ++ INTERACTIVE=false 2025-08-29 17:00:01.688637 | orchestrator | ++ export OSISM_APPLY_RETRY=1 2025-08-29 17:00:01.688641 | orchestrator | ++ OSISM_APPLY_RETRY=1 2025-08-29 17:00:01.688823 | orchestrator | + source /opt/manager-vars.sh 2025-08-29 17:00:01.688829 | orchestrator | ++ export NUMBER_OF_NODES=6 2025-08-29 17:00:01.688865 | orchestrator | ++ NUMBER_OF_NODES=6 2025-08-29 17:00:01.688870 | orchestrator | ++ export CEPH_VERSION=reef 2025-08-29 17:00:01.688874 | orchestrator | ++ CEPH_VERSION=reef 2025-08-29 17:00:01.688878 | orchestrator | ++ export CONFIGURATION_VERSION=main 2025-08-29 17:00:01.688882 | orchestrator | ++ CONFIGURATION_VERSION=main 2025-08-29 17:00:01.688898 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 17:00:01.688923 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 17:00:01.688928 | orchestrator | ++ export OPENSTACK_VERSION=2024.2 2025-08-29 17:00:01.688937 | orchestrator | ++ OPENSTACK_VERSION=2024.2 2025-08-29 17:00:01.688941 | orchestrator | ++ export ARA=false 2025-08-29 17:00:01.688955 | orchestrator | ++ ARA=false 2025-08-29 17:00:01.688998 | orchestrator | ++ export DEPLOY_MODE=manager 2025-08-29 17:00:01.689015 | orchestrator | ++ DEPLOY_MODE=manager 2025-08-29 17:00:01.689019 | orchestrator | ++ export TEMPEST=false 2025-08-29 17:00:01.689023 | orchestrator | ++ TEMPEST=false 2025-08-29 17:00:01.689027 | orchestrator | ++ export IS_ZUUL=true 2025-08-29 17:00:01.689031 | orchestrator | ++ IS_ZUUL=true 2025-08-29 17:00:01.689042 | orchestrator | ++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 17:00:01.689049 | orchestrator | ++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 17:00:01.689062 | orchestrator | ++ export EXTERNAL_API=false 2025-08-29 17:00:01.689067 | orchestrator | ++ EXTERNAL_API=false 2025-08-29 17:00:01.689071 | orchestrator | ++ export IMAGE_USER=ubuntu 2025-08-29 17:00:01.689075 | orchestrator | ++ IMAGE_USER=ubuntu 2025-08-29 17:00:01.689174 | orchestrator | ++ export IMAGE_NODE_USER=ubuntu 2025-08-29 17:00:01.689223 | orchestrator | ++ IMAGE_NODE_USER=ubuntu 2025-08-29 17:00:01.689229 | orchestrator | ++ export CEPH_STACK=ceph-ansible 2025-08-29 17:00:01.689233 | orchestrator | ++ CEPH_STACK=ceph-ansible 2025-08-29 17:00:01.689236 | orchestrator | + sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver 2025-08-29 17:00:01.746211 | orchestrator | + docker version 2025-08-29 17:00:02.011470 | orchestrator | Client: Docker Engine - Community 2025-08-29 17:00:02.011532 | orchestrator | Version: 27.5.1 2025-08-29 17:00:02.011544 | orchestrator | API version: 1.47 2025-08-29 17:00:02.011552 | orchestrator | Go version: go1.22.11 2025-08-29 17:00:02.011559 | orchestrator | Git commit: 9f9e405 2025-08-29 17:00:02.011568 | orchestrator | Built: Wed Jan 22 13:41:48 2025 2025-08-29 17:00:02.011577 | orchestrator | OS/Arch: linux/amd64 2025-08-29 17:00:02.011585 | orchestrator | Context: default 2025-08-29 17:00:02.011592 | orchestrator | 2025-08-29 17:00:02.011600 | orchestrator | Server: Docker Engine - Community 2025-08-29 17:00:02.011609 | orchestrator | Engine: 2025-08-29 17:00:02.011617 | orchestrator | Version: 27.5.1 2025-08-29 17:00:02.011625 | orchestrator | API version: 1.47 (minimum version 1.24) 2025-08-29 17:00:02.011654 | orchestrator | Go version: go1.22.11 2025-08-29 17:00:02.011662 | orchestrator | Git commit: 4c9b3b0 2025-08-29 17:00:02.011670 | orchestrator | Built: Wed Jan 22 13:41:48 2025 2025-08-29 17:00:02.011679 | orchestrator | OS/Arch: linux/amd64 2025-08-29 17:00:02.011687 | orchestrator | Experimental: false 2025-08-29 17:00:02.011695 | orchestrator | containerd: 2025-08-29 17:00:02.011703 | orchestrator | Version: 1.7.27 2025-08-29 17:00:02.011721 | orchestrator | GitCommit: 05044ec0a9a75232cad458027ca83437aae3f4da 2025-08-29 17:00:02.011729 | orchestrator | runc: 2025-08-29 17:00:02.011737 | orchestrator | Version: 1.2.5 2025-08-29 17:00:02.011746 | orchestrator | GitCommit: v1.2.5-0-g59923ef 2025-08-29 17:00:02.011754 | orchestrator | docker-init: 2025-08-29 17:00:02.011761 | orchestrator | Version: 0.19.0 2025-08-29 17:00:02.011768 | orchestrator | GitCommit: de40ad0 2025-08-29 17:00:02.015598 | orchestrator | + sh -c /opt/configuration/scripts/deploy/000-manager.sh 2025-08-29 17:00:02.022855 | orchestrator | + set -e 2025-08-29 17:00:02.022876 | orchestrator | + source /opt/manager-vars.sh 2025-08-29 17:00:02.022886 | orchestrator | ++ export NUMBER_OF_NODES=6 2025-08-29 17:00:02.022892 | orchestrator | ++ NUMBER_OF_NODES=6 2025-08-29 17:00:02.022898 | orchestrator | ++ export CEPH_VERSION=reef 2025-08-29 17:00:02.022903 | orchestrator | ++ CEPH_VERSION=reef 2025-08-29 17:00:02.022909 | orchestrator | ++ export CONFIGURATION_VERSION=main 2025-08-29 17:00:02.022915 | orchestrator | ++ CONFIGURATION_VERSION=main 2025-08-29 17:00:02.022942 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 17:00:02.022948 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 17:00:02.022954 | orchestrator | ++ export OPENSTACK_VERSION=2024.2 2025-08-29 17:00:02.022960 | orchestrator | ++ OPENSTACK_VERSION=2024.2 2025-08-29 17:00:02.022981 | orchestrator | ++ export ARA=false 2025-08-29 17:00:02.022988 | orchestrator | ++ ARA=false 2025-08-29 17:00:02.022993 | orchestrator | ++ export DEPLOY_MODE=manager 2025-08-29 17:00:02.023019 | orchestrator | ++ DEPLOY_MODE=manager 2025-08-29 17:00:02.023048 | orchestrator | ++ export TEMPEST=false 2025-08-29 17:00:02.023055 | orchestrator | ++ TEMPEST=false 2025-08-29 17:00:02.023061 | orchestrator | ++ export IS_ZUUL=true 2025-08-29 17:00:02.023066 | orchestrator | ++ IS_ZUUL=true 2025-08-29 17:00:02.023092 | orchestrator | ++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 17:00:02.023098 | orchestrator | ++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 17:00:02.023104 | orchestrator | ++ export EXTERNAL_API=false 2025-08-29 17:00:02.023110 | orchestrator | ++ EXTERNAL_API=false 2025-08-29 17:00:02.023115 | orchestrator | ++ export IMAGE_USER=ubuntu 2025-08-29 17:00:02.023120 | orchestrator | ++ IMAGE_USER=ubuntu 2025-08-29 17:00:02.023126 | orchestrator | ++ export IMAGE_NODE_USER=ubuntu 2025-08-29 17:00:02.023132 | orchestrator | ++ IMAGE_NODE_USER=ubuntu 2025-08-29 17:00:02.023137 | orchestrator | ++ export CEPH_STACK=ceph-ansible 2025-08-29 17:00:02.023143 | orchestrator | ++ CEPH_STACK=ceph-ansible 2025-08-29 17:00:02.023151 | orchestrator | + source /opt/configuration/scripts/include.sh 2025-08-29 17:00:02.023157 | orchestrator | ++ export INTERACTIVE=false 2025-08-29 17:00:02.023162 | orchestrator | ++ INTERACTIVE=false 2025-08-29 17:00:02.023167 | orchestrator | ++ export OSISM_APPLY_RETRY=1 2025-08-29 17:00:02.023174 | orchestrator | ++ OSISM_APPLY_RETRY=1 2025-08-29 17:00:02.023470 | orchestrator | + [[ 9.2.0 != \l\a\t\e\s\t ]] 2025-08-29 17:00:02.023480 | orchestrator | + /opt/configuration/scripts/set-manager-version.sh 9.2.0 2025-08-29 17:00:02.031163 | orchestrator | + set -e 2025-08-29 17:00:02.031222 | orchestrator | + VERSION=9.2.0 2025-08-29 17:00:02.031239 | orchestrator | + sed -i 's/manager_version: .*/manager_version: 9.2.0/g' /opt/configuration/environments/manager/configuration.yml 2025-08-29 17:00:02.041212 | orchestrator | + [[ 9.2.0 != \l\a\t\e\s\t ]] 2025-08-29 17:00:02.041240 | orchestrator | + sed -i /ceph_version:/d /opt/configuration/environments/manager/configuration.yml 2025-08-29 17:00:02.045252 | orchestrator | + sed -i /openstack_version:/d /opt/configuration/environments/manager/configuration.yml 2025-08-29 17:00:02.048629 | orchestrator | + sh -c /opt/configuration/scripts/sync-configuration-repository.sh 2025-08-29 17:00:02.056908 | orchestrator | /opt/configuration ~ 2025-08-29 17:00:02.057022 | orchestrator | + set -e 2025-08-29 17:00:02.057038 | orchestrator | + pushd /opt/configuration 2025-08-29 17:00:02.057047 | orchestrator | + [[ -e /opt/venv/bin/activate ]] 2025-08-29 17:00:02.058545 | orchestrator | + source /opt/venv/bin/activate 2025-08-29 17:00:02.059548 | orchestrator | ++ deactivate nondestructive 2025-08-29 17:00:02.059572 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:02.059582 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:02.059605 | orchestrator | ++ hash -r 2025-08-29 17:00:02.059694 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:02.059704 | orchestrator | ++ unset VIRTUAL_ENV 2025-08-29 17:00:02.059787 | orchestrator | ++ unset VIRTUAL_ENV_PROMPT 2025-08-29 17:00:02.059797 | orchestrator | ++ '[' '!' nondestructive = nondestructive ']' 2025-08-29 17:00:02.060069 | orchestrator | ++ '[' linux-gnu = cygwin ']' 2025-08-29 17:00:02.060088 | orchestrator | ++ '[' linux-gnu = msys ']' 2025-08-29 17:00:02.060104 | orchestrator | ++ export VIRTUAL_ENV=/opt/venv 2025-08-29 17:00:02.060116 | orchestrator | ++ VIRTUAL_ENV=/opt/venv 2025-08-29 17:00:02.060149 | orchestrator | ++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin 2025-08-29 17:00:02.060165 | orchestrator | ++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin 2025-08-29 17:00:02.060230 | orchestrator | ++ export PATH 2025-08-29 17:00:02.060580 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:02.060601 | orchestrator | ++ '[' -z '' ']' 2025-08-29 17:00:02.060612 | orchestrator | ++ _OLD_VIRTUAL_PS1= 2025-08-29 17:00:02.060623 | orchestrator | ++ PS1='(venv) ' 2025-08-29 17:00:02.060634 | orchestrator | ++ export PS1 2025-08-29 17:00:02.060644 | orchestrator | ++ VIRTUAL_ENV_PROMPT='(venv) ' 2025-08-29 17:00:02.060654 | orchestrator | ++ export VIRTUAL_ENV_PROMPT 2025-08-29 17:00:02.060668 | orchestrator | ++ hash -r 2025-08-29 17:00:02.060810 | orchestrator | + pip3 install --no-cache-dir python-gilt==1.2.3 requests Jinja2 PyYAML packaging 2025-08-29 17:00:03.123036 | orchestrator | Requirement already satisfied: python-gilt==1.2.3 in /opt/venv/lib/python3.12/site-packages (1.2.3) 2025-08-29 17:00:03.124363 | orchestrator | Requirement already satisfied: requests in /opt/venv/lib/python3.12/site-packages (2.32.5) 2025-08-29 17:00:03.125755 | orchestrator | Requirement already satisfied: Jinja2 in /opt/venv/lib/python3.12/site-packages (3.1.6) 2025-08-29 17:00:03.127197 | orchestrator | Requirement already satisfied: PyYAML in /opt/venv/lib/python3.12/site-packages (6.0.2) 2025-08-29 17:00:03.128395 | orchestrator | Requirement already satisfied: packaging in /opt/venv/lib/python3.12/site-packages (25.0) 2025-08-29 17:00:03.138326 | orchestrator | Requirement already satisfied: click in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (8.2.1) 2025-08-29 17:00:03.139896 | orchestrator | Requirement already satisfied: colorama in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.4.6) 2025-08-29 17:00:03.141074 | orchestrator | Requirement already satisfied: fasteners in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (0.20) 2025-08-29 17:00:03.142476 | orchestrator | Requirement already satisfied: sh in /opt/venv/lib/python3.12/site-packages (from python-gilt==1.2.3) (2.2.2) 2025-08-29 17:00:03.173909 | orchestrator | Requirement already satisfied: charset_normalizer<4,>=2 in /opt/venv/lib/python3.12/site-packages (from requests) (3.4.3) 2025-08-29 17:00:03.175606 | orchestrator | Requirement already satisfied: idna<4,>=2.5 in /opt/venv/lib/python3.12/site-packages (from requests) (3.10) 2025-08-29 17:00:03.177408 | orchestrator | Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/venv/lib/python3.12/site-packages (from requests) (2.5.0) 2025-08-29 17:00:03.179543 | orchestrator | Requirement already satisfied: certifi>=2017.4.17 in /opt/venv/lib/python3.12/site-packages (from requests) (2025.8.3) 2025-08-29 17:00:03.183056 | orchestrator | Requirement already satisfied: MarkupSafe>=2.0 in /opt/venv/lib/python3.12/site-packages (from Jinja2) (3.0.2) 2025-08-29 17:00:03.385608 | orchestrator | ++ which gilt 2025-08-29 17:00:03.388431 | orchestrator | + GILT=/opt/venv/bin/gilt 2025-08-29 17:00:03.388496 | orchestrator | + /opt/venv/bin/gilt overlay 2025-08-29 17:00:03.637220 | orchestrator | osism.cfg-generics: 2025-08-29 17:00:03.794455 | orchestrator | - copied (v0.20250709.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/environments/manager/images.yml to /opt/configuration/environments/manager/ 2025-08-29 17:00:03.794700 | orchestrator | - copied (v0.20250709.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/render-images.py to /opt/configuration/environments/manager/ 2025-08-29 17:00:03.795035 | orchestrator | - copied (v0.20250709.0) /home/dragon/.gilt/clone/github.com/osism.cfg-generics/src/set-versions.py to /opt/configuration/environments/ 2025-08-29 17:00:03.795151 | orchestrator | - running `/opt/configuration/scripts/wrapper-gilt.sh render-images` in /opt/configuration/environments/manager/ 2025-08-29 17:00:04.537655 | orchestrator | - running `rm render-images.py` in /opt/configuration/environments/manager/ 2025-08-29 17:00:04.547469 | orchestrator | - running `/opt/configuration/scripts/wrapper-gilt.sh set-versions` in /opt/configuration/environments/ 2025-08-29 17:00:04.972295 | orchestrator | - running `rm set-versions.py` in /opt/configuration/environments/ 2025-08-29 17:00:05.015516 | orchestrator | + [[ -e /opt/venv/bin/activate ]] 2025-08-29 17:00:05.015667 | orchestrator | + deactivate 2025-08-29 17:00:05.015686 | orchestrator | + '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']' 2025-08-29 17:00:05.015699 | orchestrator | + PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin 2025-08-29 17:00:05.015722 | orchestrator | + export PATH 2025-08-29 17:00:05.015734 | orchestrator | + unset _OLD_VIRTUAL_PATH 2025-08-29 17:00:05.015746 | orchestrator | + '[' -n '' ']' 2025-08-29 17:00:05.015758 | orchestrator | + hash -r 2025-08-29 17:00:05.015814 | orchestrator | + '[' -n '' ']' 2025-08-29 17:00:05.015828 | orchestrator | + unset VIRTUAL_ENV 2025-08-29 17:00:05.015840 | orchestrator | + unset VIRTUAL_ENV_PROMPT 2025-08-29 17:00:05.015851 | orchestrator | + '[' '!' '' = nondestructive ']' 2025-08-29 17:00:05.016070 | orchestrator | ~ 2025-08-29 17:00:05.016092 | orchestrator | + unset -f deactivate 2025-08-29 17:00:05.016104 | orchestrator | + popd 2025-08-29 17:00:05.017307 | orchestrator | + [[ 9.2.0 == \l\a\t\e\s\t ]] 2025-08-29 17:00:05.017337 | orchestrator | + [[ ceph-ansible == \r\o\o\k ]] 2025-08-29 17:00:05.018355 | orchestrator | ++ semver 9.2.0 7.0.0 2025-08-29 17:00:05.065583 | orchestrator | + [[ 1 -ge 0 ]] 2025-08-29 17:00:05.065680 | orchestrator | + echo 'enable_osism_kubernetes: true' 2025-08-29 17:00:05.065697 | orchestrator | + /opt/configuration/scripts/enable-resource-nodes.sh 2025-08-29 17:00:05.148605 | orchestrator | + [[ -e /opt/venv/bin/activate ]] 2025-08-29 17:00:05.148737 | orchestrator | + source /opt/venv/bin/activate 2025-08-29 17:00:05.148783 | orchestrator | ++ deactivate nondestructive 2025-08-29 17:00:05.148796 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:05.148807 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:05.148818 | orchestrator | ++ hash -r 2025-08-29 17:00:05.149002 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:05.149023 | orchestrator | ++ unset VIRTUAL_ENV 2025-08-29 17:00:05.149039 | orchestrator | ++ unset VIRTUAL_ENV_PROMPT 2025-08-29 17:00:05.149050 | orchestrator | ++ '[' '!' nondestructive = nondestructive ']' 2025-08-29 17:00:05.149683 | orchestrator | ++ '[' linux-gnu = cygwin ']' 2025-08-29 17:00:05.149702 | orchestrator | ++ '[' linux-gnu = msys ']' 2025-08-29 17:00:05.149714 | orchestrator | ++ export VIRTUAL_ENV=/opt/venv 2025-08-29 17:00:05.149725 | orchestrator | ++ VIRTUAL_ENV=/opt/venv 2025-08-29 17:00:05.149770 | orchestrator | ++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin 2025-08-29 17:00:05.149784 | orchestrator | ++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin 2025-08-29 17:00:05.149810 | orchestrator | ++ export PATH 2025-08-29 17:00:05.149822 | orchestrator | ++ '[' -n '' ']' 2025-08-29 17:00:05.149855 | orchestrator | ++ '[' -z '' ']' 2025-08-29 17:00:05.149867 | orchestrator | ++ _OLD_VIRTUAL_PS1= 2025-08-29 17:00:05.149879 | orchestrator | ++ PS1='(venv) ' 2025-08-29 17:00:05.149890 | orchestrator | ++ export PS1 2025-08-29 17:00:05.149902 | orchestrator | ++ VIRTUAL_ENV_PROMPT='(venv) ' 2025-08-29 17:00:05.149918 | orchestrator | ++ export VIRTUAL_ENV_PROMPT 2025-08-29 17:00:05.149952 | orchestrator | ++ hash -r 2025-08-29 17:00:05.149965 | orchestrator | + ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml 2025-08-29 17:00:06.101852 | orchestrator | 2025-08-29 17:00:06.101892 | orchestrator | PLAY [Copy custom facts] ******************************************************* 2025-08-29 17:00:06.101898 | orchestrator | 2025-08-29 17:00:06.101903 | orchestrator | TASK [Create custom facts directory] ******************************************* 2025-08-29 17:00:06.584805 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:06.584852 | orchestrator | 2025-08-29 17:00:06.584856 | orchestrator | TASK [Copy fact files] ********************************************************* 2025-08-29 17:00:07.444118 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:07.444164 | orchestrator | 2025-08-29 17:00:07.444170 | orchestrator | PLAY [Before the deployment of the manager] ************************************ 2025-08-29 17:00:07.444175 | orchestrator | 2025-08-29 17:00:07.444179 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 17:00:09.440282 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:09.441130 | orchestrator | 2025-08-29 17:00:09.441164 | orchestrator | TASK [Get /opt/manager-vars.sh] ************************************************ 2025-08-29 17:00:09.477429 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:09.477479 | orchestrator | 2025-08-29 17:00:09.477493 | orchestrator | TASK [Add ara_server_mariadb_volume_type parameter] **************************** 2025-08-29 17:00:09.848630 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:09.848714 | orchestrator | 2025-08-29 17:00:09.848731 | orchestrator | TASK [Add netbox_enable parameter] ********************************************* 2025-08-29 17:00:09.874433 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:09.874471 | orchestrator | 2025-08-29 17:00:09.874483 | orchestrator | TASK [Install HWE kernel package on Ubuntu] ************************************ 2025-08-29 17:00:10.177743 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:10.177815 | orchestrator | 2025-08-29 17:00:10.177830 | orchestrator | TASK [Use insecure glance configuration] *************************************** 2025-08-29 17:00:10.223473 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:10.223528 | orchestrator | 2025-08-29 17:00:10.223542 | orchestrator | TASK [Check if /etc/OTC_region exist] ****************************************** 2025-08-29 17:00:10.520932 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:10.521034 | orchestrator | 2025-08-29 17:00:10.521050 | orchestrator | TASK [Add nova_compute_virt_type parameter] ************************************ 2025-08-29 17:00:10.606550 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:10.606593 | orchestrator | 2025-08-29 17:00:10.606609 | orchestrator | PLAY [Apply role traefik] ****************************************************** 2025-08-29 17:00:10.606621 | orchestrator | 2025-08-29 17:00:10.606633 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 17:00:12.119120 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:12.119200 | orchestrator | 2025-08-29 17:00:12.119215 | orchestrator | TASK [Apply traefik role] ****************************************************** 2025-08-29 17:00:12.238686 | orchestrator | included: osism.services.traefik for testbed-manager 2025-08-29 17:00:12.238730 | orchestrator | 2025-08-29 17:00:12.238743 | orchestrator | TASK [osism.services.traefik : Include config tasks] *************************** 2025-08-29 17:00:12.303603 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager 2025-08-29 17:00:12.303637 | orchestrator | 2025-08-29 17:00:12.303642 | orchestrator | TASK [osism.services.traefik : Create required directories] ******************** 2025-08-29 17:00:13.307609 | orchestrator | changed: [testbed-manager] => (item=/opt/traefik) 2025-08-29 17:00:13.307743 | orchestrator | changed: [testbed-manager] => (item=/opt/traefik/certificates) 2025-08-29 17:00:13.307760 | orchestrator | changed: [testbed-manager] => (item=/opt/traefik/configuration) 2025-08-29 17:00:13.307773 | orchestrator | 2025-08-29 17:00:13.307785 | orchestrator | TASK [osism.services.traefik : Copy configuration files] *********************** 2025-08-29 17:00:14.977834 | orchestrator | changed: [testbed-manager] => (item=traefik.yml) 2025-08-29 17:00:14.977929 | orchestrator | changed: [testbed-manager] => (item=traefik.env) 2025-08-29 17:00:14.977945 | orchestrator | changed: [testbed-manager] => (item=certificates.yml) 2025-08-29 17:00:14.977958 | orchestrator | 2025-08-29 17:00:14.978005 | orchestrator | TASK [osism.services.traefik : Copy certificate cert files] ******************** 2025-08-29 17:00:15.607569 | orchestrator | changed: [testbed-manager] => (item=None) 2025-08-29 17:00:15.607644 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:15.607659 | orchestrator | 2025-08-29 17:00:15.607671 | orchestrator | TASK [osism.services.traefik : Copy certificate key files] ********************* 2025-08-29 17:00:16.266785 | orchestrator | changed: [testbed-manager] => (item=None) 2025-08-29 17:00:16.266870 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:16.266885 | orchestrator | 2025-08-29 17:00:16.266898 | orchestrator | TASK [osism.services.traefik : Copy dynamic configuration] ********************* 2025-08-29 17:00:16.315463 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:16.315520 | orchestrator | 2025-08-29 17:00:16.315527 | orchestrator | TASK [osism.services.traefik : Remove dynamic configuration] ******************* 2025-08-29 17:00:16.659553 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:16.659627 | orchestrator | 2025-08-29 17:00:16.659644 | orchestrator | TASK [osism.services.traefik : Include service tasks] ************************** 2025-08-29 17:00:16.733210 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager 2025-08-29 17:00:16.733276 | orchestrator | 2025-08-29 17:00:16.733290 | orchestrator | TASK [osism.services.traefik : Create traefik external network] **************** 2025-08-29 17:00:17.764639 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:17.764730 | orchestrator | 2025-08-29 17:00:17.764746 | orchestrator | TASK [osism.services.traefik : Copy docker-compose.yml file] ******************* 2025-08-29 17:00:18.529936 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:18.530116 | orchestrator | 2025-08-29 17:00:18.530135 | orchestrator | TASK [osism.services.traefik : Manage traefik service] ************************* 2025-08-29 17:00:29.489538 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:29.489625 | orchestrator | 2025-08-29 17:00:29.489657 | orchestrator | RUNNING HANDLER [osism.services.traefik : Restart traefik service] ************* 2025-08-29 17:00:29.533221 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:29.533291 | orchestrator | 2025-08-29 17:00:29.533306 | orchestrator | PLAY [Deploy manager service] ************************************************** 2025-08-29 17:00:29.533318 | orchestrator | 2025-08-29 17:00:29.533329 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 17:00:31.215026 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:31.215125 | orchestrator | 2025-08-29 17:00:31.215141 | orchestrator | TASK [Apply manager role] ****************************************************** 2025-08-29 17:00:31.314796 | orchestrator | included: osism.services.manager for testbed-manager 2025-08-29 17:00:31.314869 | orchestrator | 2025-08-29 17:00:31.314881 | orchestrator | TASK [osism.services.manager : Include install tasks] ************************** 2025-08-29 17:00:31.365783 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager 2025-08-29 17:00:31.365850 | orchestrator | 2025-08-29 17:00:31.365864 | orchestrator | TASK [osism.services.manager : Install required packages] ********************** 2025-08-29 17:00:33.752481 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:33.752651 | orchestrator | 2025-08-29 17:00:33.752673 | orchestrator | TASK [osism.services.manager : Gather variables for each operating system] ***** 2025-08-29 17:00:33.806373 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:33.806446 | orchestrator | 2025-08-29 17:00:33.806459 | orchestrator | TASK [osism.services.manager : Include config tasks] *************************** 2025-08-29 17:00:33.933758 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager 2025-08-29 17:00:33.933927 | orchestrator | 2025-08-29 17:00:33.933945 | orchestrator | TASK [osism.services.manager : Create required directories] ******************** 2025-08-29 17:00:36.732273 | orchestrator | changed: [testbed-manager] => (item=/opt/ansible) 2025-08-29 17:00:36.732379 | orchestrator | changed: [testbed-manager] => (item=/opt/archive) 2025-08-29 17:00:36.732389 | orchestrator | changed: [testbed-manager] => (item=/opt/manager/configuration) 2025-08-29 17:00:36.732397 | orchestrator | changed: [testbed-manager] => (item=/opt/manager/data) 2025-08-29 17:00:36.732403 | orchestrator | ok: [testbed-manager] => (item=/opt/manager) 2025-08-29 17:00:36.732410 | orchestrator | changed: [testbed-manager] => (item=/opt/manager/secrets) 2025-08-29 17:00:36.732438 | orchestrator | changed: [testbed-manager] => (item=/opt/ansible/secrets) 2025-08-29 17:00:36.732447 | orchestrator | changed: [testbed-manager] => (item=/opt/state) 2025-08-29 17:00:36.732454 | orchestrator | 2025-08-29 17:00:36.732463 | orchestrator | TASK [osism.services.manager : Copy all environment file] ********************** 2025-08-29 17:00:37.345711 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:37.345796 | orchestrator | 2025-08-29 17:00:37.345813 | orchestrator | TASK [osism.services.manager : Copy client environment file] ******************* 2025-08-29 17:00:37.952324 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:37.952405 | orchestrator | 2025-08-29 17:00:37.952420 | orchestrator | TASK [osism.services.manager : Include ara config tasks] *********************** 2025-08-29 17:00:38.028662 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager 2025-08-29 17:00:38.028735 | orchestrator | 2025-08-29 17:00:38.028749 | orchestrator | TASK [osism.services.manager : Copy ARA environment files] ********************* 2025-08-29 17:00:39.198383 | orchestrator | changed: [testbed-manager] => (item=ara) 2025-08-29 17:00:39.198523 | orchestrator | changed: [testbed-manager] => (item=ara-server) 2025-08-29 17:00:39.198551 | orchestrator | 2025-08-29 17:00:39.198572 | orchestrator | TASK [osism.services.manager : Copy MariaDB environment file] ****************** 2025-08-29 17:00:39.801858 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:39.801954 | orchestrator | 2025-08-29 17:00:39.802011 | orchestrator | TASK [osism.services.manager : Include vault config tasks] ********************* 2025-08-29 17:00:39.855377 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:39.855460 | orchestrator | 2025-08-29 17:00:39.855484 | orchestrator | TASK [osism.services.manager : Include frontend config tasks] ****************** 2025-08-29 17:00:39.903817 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:39.903884 | orchestrator | 2025-08-29 17:00:39.903907 | orchestrator | TASK [osism.services.manager : Include ansible config tasks] ******************* 2025-08-29 17:00:39.956850 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager 2025-08-29 17:00:39.956900 | orchestrator | 2025-08-29 17:00:39.956918 | orchestrator | TASK [osism.services.manager : Copy private ssh keys] ************************** 2025-08-29 17:00:41.293915 | orchestrator | changed: [testbed-manager] => (item=None) 2025-08-29 17:00:41.294076 | orchestrator | changed: [testbed-manager] => (item=None) 2025-08-29 17:00:41.294094 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:41.294107 | orchestrator | 2025-08-29 17:00:41.294119 | orchestrator | TASK [osism.services.manager : Copy ansible environment file] ****************** 2025-08-29 17:00:41.886407 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:41.886477 | orchestrator | 2025-08-29 17:00:41.886488 | orchestrator | TASK [osism.services.manager : Include netbox config tasks] ******************** 2025-08-29 17:00:41.937795 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:41.937857 | orchestrator | 2025-08-29 17:00:41.937871 | orchestrator | TASK [osism.services.manager : Include celery config tasks] ******************** 2025-08-29 17:00:42.032260 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager 2025-08-29 17:00:42.032333 | orchestrator | 2025-08-29 17:00:42.032350 | orchestrator | TASK [osism.services.manager : Set fs.inotify.max_user_watches] **************** 2025-08-29 17:00:42.524211 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:42.524293 | orchestrator | 2025-08-29 17:00:42.524308 | orchestrator | TASK [osism.services.manager : Set fs.inotify.max_user_instances] ************** 2025-08-29 17:00:42.910691 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:42.910781 | orchestrator | 2025-08-29 17:00:42.910797 | orchestrator | TASK [osism.services.manager : Copy celery environment files] ****************** 2025-08-29 17:00:44.085756 | orchestrator | changed: [testbed-manager] => (item=conductor) 2025-08-29 17:00:44.085851 | orchestrator | changed: [testbed-manager] => (item=openstack) 2025-08-29 17:00:44.085866 | orchestrator | 2025-08-29 17:00:44.085879 | orchestrator | TASK [osism.services.manager : Copy listener environment file] ***************** 2025-08-29 17:00:44.721737 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:45.009060 | orchestrator | 2025-08-29 17:00:45.009124 | orchestrator | TASK [osism.services.manager : Check for conductor.yml] ************************ 2025-08-29 17:00:45.103646 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:45.103698 | orchestrator | 2025-08-29 17:00:45.103704 | orchestrator | TASK [osism.services.manager : Copy conductor configuration file] ************** 2025-08-29 17:00:45.444090 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:45.444179 | orchestrator | 2025-08-29 17:00:45.444193 | orchestrator | TASK [osism.services.manager : Copy empty conductor configuration file] ******** 2025-08-29 17:00:45.494724 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:45.494786 | orchestrator | 2025-08-29 17:00:45.494798 | orchestrator | TASK [osism.services.manager : Include wrapper config tasks] ******************* 2025-08-29 17:00:45.553676 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager 2025-08-29 17:00:45.553746 | orchestrator | 2025-08-29 17:00:45.553764 | orchestrator | TASK [osism.services.manager : Include wrapper vars file] ********************** 2025-08-29 17:00:45.592398 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:45.592487 | orchestrator | 2025-08-29 17:00:45.592501 | orchestrator | TASK [osism.services.manager : Copy wrapper scripts] *************************** 2025-08-29 17:00:47.521439 | orchestrator | changed: [testbed-manager] => (item=osism) 2025-08-29 17:00:47.521547 | orchestrator | changed: [testbed-manager] => (item=osism-update-docker) 2025-08-29 17:00:47.521563 | orchestrator | changed: [testbed-manager] => (item=osism-update-manager) 2025-08-29 17:00:47.521575 | orchestrator | 2025-08-29 17:00:47.521587 | orchestrator | TASK [osism.services.manager : Copy cilium wrapper script] ********************* 2025-08-29 17:00:48.205940 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:48.206123 | orchestrator | 2025-08-29 17:00:48.206143 | orchestrator | TASK [osism.services.manager : Copy hubble wrapper script] ********************* 2025-08-29 17:00:48.879514 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:48.879608 | orchestrator | 2025-08-29 17:00:48.879625 | orchestrator | TASK [osism.services.manager : Copy flux wrapper script] *********************** 2025-08-29 17:00:49.551268 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:49.551354 | orchestrator | 2025-08-29 17:00:49.551369 | orchestrator | TASK [osism.services.manager : Include scripts config tasks] ******************* 2025-08-29 17:00:49.620446 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager 2025-08-29 17:00:49.620513 | orchestrator | 2025-08-29 17:00:49.620527 | orchestrator | TASK [osism.services.manager : Include scripts vars file] ********************** 2025-08-29 17:00:49.661732 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:49.661797 | orchestrator | 2025-08-29 17:00:49.661811 | orchestrator | TASK [osism.services.manager : Copy scripts] *********************************** 2025-08-29 17:00:50.365144 | orchestrator | changed: [testbed-manager] => (item=osism-include) 2025-08-29 17:00:50.365239 | orchestrator | 2025-08-29 17:00:50.365255 | orchestrator | TASK [osism.services.manager : Include service tasks] ************************** 2025-08-29 17:00:50.431491 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager 2025-08-29 17:00:50.431569 | orchestrator | 2025-08-29 17:00:50.431584 | orchestrator | TASK [osism.services.manager : Copy manager systemd unit file] ***************** 2025-08-29 17:00:51.113203 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:51.113304 | orchestrator | 2025-08-29 17:00:51.113320 | orchestrator | TASK [osism.services.manager : Create traefik external network] **************** 2025-08-29 17:00:51.675064 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:51.675166 | orchestrator | 2025-08-29 17:00:51.675188 | orchestrator | TASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] *** 2025-08-29 17:00:51.716403 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:00:51.716457 | orchestrator | 2025-08-29 17:00:51.716476 | orchestrator | TASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] *** 2025-08-29 17:00:51.762308 | orchestrator | ok: [testbed-manager] 2025-08-29 17:00:51.762374 | orchestrator | 2025-08-29 17:00:51.762391 | orchestrator | TASK [osism.services.manager : Copy docker-compose.yml file] ******************* 2025-08-29 17:00:52.555900 | orchestrator | changed: [testbed-manager] 2025-08-29 17:00:52.556006 | orchestrator | 2025-08-29 17:00:52.556023 | orchestrator | TASK [osism.services.manager : Pull container images] ************************** 2025-08-29 17:01:53.013447 | orchestrator | changed: [testbed-manager] 2025-08-29 17:01:53.013550 | orchestrator | 2025-08-29 17:01:53.013567 | orchestrator | TASK [osism.services.manager : Stop and disable old service docker-compose@manager] *** 2025-08-29 17:01:53.968787 | orchestrator | ok: [testbed-manager] 2025-08-29 17:01:53.968895 | orchestrator | 2025-08-29 17:01:53.968912 | orchestrator | TASK [osism.services.manager : Do a manual start of the manager service] ******* 2025-08-29 17:01:54.033564 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:01:54.033681 | orchestrator | 2025-08-29 17:01:54.033707 | orchestrator | TASK [osism.services.manager : Manage manager service] ************************* 2025-08-29 17:01:56.654223 | orchestrator | changed: [testbed-manager] 2025-08-29 17:01:56.654317 | orchestrator | 2025-08-29 17:01:56.654334 | orchestrator | TASK [osism.services.manager : Register that manager service was started] ****** 2025-08-29 17:01:56.706705 | orchestrator | ok: [testbed-manager] 2025-08-29 17:01:56.706791 | orchestrator | 2025-08-29 17:01:56.706808 | orchestrator | TASK [osism.services.manager : Flush handlers] ********************************* 2025-08-29 17:01:56.706828 | orchestrator | 2025-08-29 17:01:56.706887 | orchestrator | RUNNING HANDLER [osism.services.manager : Restart manager service] ************* 2025-08-29 17:01:56.755390 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:01:56.755481 | orchestrator | 2025-08-29 17:01:56.755496 | orchestrator | RUNNING HANDLER [osism.services.manager : Wait for manager service to start] *** 2025-08-29 17:02:56.845037 | orchestrator | Pausing for 60 seconds 2025-08-29 17:02:56.845164 | orchestrator | changed: [testbed-manager] 2025-08-29 17:02:56.845181 | orchestrator | 2025-08-29 17:02:56.845195 | orchestrator | RUNNING HANDLER [osism.services.manager : Ensure that all containers are up] *** 2025-08-29 17:03:00.441610 | orchestrator | changed: [testbed-manager] 2025-08-29 17:03:00.441711 | orchestrator | 2025-08-29 17:03:00.441727 | orchestrator | RUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] *** 2025-08-29 17:03:42.015044 | orchestrator | FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left). 2025-08-29 17:03:42.015165 | orchestrator | FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left). 2025-08-29 17:03:42.015181 | orchestrator | changed: [testbed-manager] 2025-08-29 17:03:42.015194 | orchestrator | 2025-08-29 17:03:42.015227 | orchestrator | RUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] *** 2025-08-29 17:03:51.381222 | orchestrator | changed: [testbed-manager] 2025-08-29 17:03:51.381320 | orchestrator | 2025-08-29 17:03:51.381338 | orchestrator | TASK [osism.services.manager : Include initialize tasks] *********************** 2025-08-29 17:03:51.452097 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager 2025-08-29 17:03:51.452163 | orchestrator | 2025-08-29 17:03:51.452178 | orchestrator | TASK [osism.services.manager : Flush handlers] ********************************* 2025-08-29 17:03:51.452190 | orchestrator | 2025-08-29 17:03:51.452202 | orchestrator | TASK [osism.services.manager : Include vault initialize tasks] ***************** 2025-08-29 17:03:51.502418 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:03:51.502512 | orchestrator | 2025-08-29 17:03:51.502538 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:03:51.502563 | orchestrator | testbed-manager : ok=64 changed=35 unreachable=0 failed=0 skipped=13 rescued=0 ignored=0 2025-08-29 17:03:51.502584 | orchestrator | 2025-08-29 17:03:51.590841 | orchestrator | + [[ -e /opt/venv/bin/activate ]] 2025-08-29 17:03:51.590910 | orchestrator | + deactivate 2025-08-29 17:03:51.590927 | orchestrator | + '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']' 2025-08-29 17:03:51.590939 | orchestrator | + PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin 2025-08-29 17:03:51.590949 | orchestrator | + export PATH 2025-08-29 17:03:51.590959 | orchestrator | + unset _OLD_VIRTUAL_PATH 2025-08-29 17:03:51.590969 | orchestrator | + '[' -n '' ']' 2025-08-29 17:03:51.591002 | orchestrator | + hash -r 2025-08-29 17:03:51.591012 | orchestrator | + '[' -n '' ']' 2025-08-29 17:03:51.591022 | orchestrator | + unset VIRTUAL_ENV 2025-08-29 17:03:51.591031 | orchestrator | + unset VIRTUAL_ENV_PROMPT 2025-08-29 17:03:51.591041 | orchestrator | + '[' '!' '' = nondestructive ']' 2025-08-29 17:03:51.591051 | orchestrator | + unset -f deactivate 2025-08-29 17:03:51.591062 | orchestrator | + cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub 2025-08-29 17:03:51.596370 | orchestrator | + [[ ceph-ansible == \c\e\p\h\-\a\n\s\i\b\l\e ]] 2025-08-29 17:03:51.596419 | orchestrator | + wait_for_container_healthy 60 ceph-ansible 2025-08-29 17:03:51.596433 | orchestrator | + local max_attempts=60 2025-08-29 17:03:51.596445 | orchestrator | + local name=ceph-ansible 2025-08-29 17:03:51.596455 | orchestrator | + local attempt_num=1 2025-08-29 17:03:51.597456 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:03:51.627437 | orchestrator | + [[ healthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:03:51.627568 | orchestrator | + wait_for_container_healthy 60 kolla-ansible 2025-08-29 17:03:51.627583 | orchestrator | + local max_attempts=60 2025-08-29 17:03:51.627593 | orchestrator | + local name=kolla-ansible 2025-08-29 17:03:51.627603 | orchestrator | + local attempt_num=1 2025-08-29 17:03:51.628330 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible 2025-08-29 17:03:51.665820 | orchestrator | + [[ healthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:03:51.665862 | orchestrator | + wait_for_container_healthy 60 osism-ansible 2025-08-29 17:03:51.665898 | orchestrator | + local max_attempts=60 2025-08-29 17:03:51.665908 | orchestrator | + local name=osism-ansible 2025-08-29 17:03:51.665918 | orchestrator | + local attempt_num=1 2025-08-29 17:03:51.667000 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible 2025-08-29 17:03:51.701690 | orchestrator | + [[ healthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:03:51.701752 | orchestrator | + [[ true == \t\r\u\e ]] 2025-08-29 17:03:51.701765 | orchestrator | + sh -c /opt/configuration/scripts/disable-ara.sh 2025-08-29 17:03:52.389902 | orchestrator | + docker compose --project-directory /opt/manager ps 2025-08-29 17:03:52.594232 | orchestrator | NAME IMAGE COMMAND SERVICE CREATED STATUS PORTS 2025-08-29 17:03:52.594311 | orchestrator | ceph-ansible registry.osism.tech/osism/ceph-ansible:0.20250711.0 "/entrypoint.sh osis…" ceph-ansible About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594328 | orchestrator | kolla-ansible registry.osism.tech/osism/kolla-ansible:0.20250711.0 "/entrypoint.sh osis…" kolla-ansible About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594340 | orchestrator | manager-api-1 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" api About a minute ago Up About a minute (healthy) 192.168.16.5:8000->8000/tcp 2025-08-29 17:03:52.594354 | orchestrator | manager-ara-server-1 registry.osism.tech/osism/ara-server:1.7.2 "sh -c '/wait && /ru…" ara-server About a minute ago Up About a minute (healthy) 8000/tcp 2025-08-29 17:03:52.594365 | orchestrator | manager-beat-1 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" beat About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594376 | orchestrator | manager-flower-1 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" flower About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594387 | orchestrator | manager-inventory_reconciler-1 registry.osism.tech/osism/inventory-reconciler:0.20250711.0 "/sbin/tini -- /entr…" inventory_reconciler About a minute ago Up 52 seconds (healthy) 2025-08-29 17:03:52.594398 | orchestrator | manager-listener-1 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" listener About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594409 | orchestrator | manager-mariadb-1 registry.osism.tech/dockerhub/library/mariadb:11.8.2 "docker-entrypoint.s…" mariadb About a minute ago Up About a minute (healthy) 3306/tcp 2025-08-29 17:03:52.594420 | orchestrator | manager-openstack-1 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" openstack About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594431 | orchestrator | manager-redis-1 registry.osism.tech/dockerhub/library/redis:7.4.5-alpine "docker-entrypoint.s…" redis About a minute ago Up About a minute (healthy) 6379/tcp 2025-08-29 17:03:52.594442 | orchestrator | osism-ansible registry.osism.tech/osism/osism-ansible:0.20250711.0 "/entrypoint.sh osis…" osism-ansible About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594453 | orchestrator | osism-kubernetes registry.osism.tech/osism/osism-kubernetes:0.20250711.0 "/entrypoint.sh osis…" osism-kubernetes About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.594465 | orchestrator | osismclient registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- sleep…" osismclient About a minute ago Up About a minute (healthy) 2025-08-29 17:03:52.604289 | orchestrator | ++ semver 9.2.0 7.0.0 2025-08-29 17:03:52.664712 | orchestrator | + [[ 1 -ge 0 ]] 2025-08-29 17:03:52.664792 | orchestrator | + sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg 2025-08-29 17:03:52.669390 | orchestrator | + osism apply resolvconf -l testbed-manager 2025-08-29 17:04:04.746205 | orchestrator | 2025-08-29 17:04:04 | INFO  | Task 7a1d3c3e-e787-4916-858f-a6ffcc5dc193 (resolvconf) was prepared for execution. 2025-08-29 17:04:04.746308 | orchestrator | 2025-08-29 17:04:04 | INFO  | It takes a moment until task 7a1d3c3e-e787-4916-858f-a6ffcc5dc193 (resolvconf) has been started and output is visible here. 2025-08-29 17:04:16.631455 | orchestrator | 2025-08-29 17:04:16.631547 | orchestrator | PLAY [Apply role resolvconf] *************************************************** 2025-08-29 17:04:16.631561 | orchestrator | 2025-08-29 17:04:16.631572 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 17:04:16.631583 | orchestrator | Friday 29 August 2025 17:04:08 +0000 (0:00:00.108) 0:00:00.108 ********* 2025-08-29 17:04:16.631593 | orchestrator | ok: [testbed-manager] 2025-08-29 17:04:16.631604 | orchestrator | 2025-08-29 17:04:16.631613 | orchestrator | TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] *** 2025-08-29 17:04:16.631624 | orchestrator | Friday 29 August 2025 17:04:11 +0000 (0:00:03.202) 0:00:03.311 ********* 2025-08-29 17:04:16.631634 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:04:16.631644 | orchestrator | 2025-08-29 17:04:16.631653 | orchestrator | TASK [osism.commons.resolvconf : Include resolvconf tasks] ********************* 2025-08-29 17:04:16.631663 | orchestrator | Friday 29 August 2025 17:04:11 +0000 (0:00:00.057) 0:00:03.369 ********* 2025-08-29 17:04:16.631673 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager 2025-08-29 17:04:16.631684 | orchestrator | 2025-08-29 17:04:16.631695 | orchestrator | TASK [osism.commons.resolvconf : Include distribution specific installation tasks] *** 2025-08-29 17:04:16.631704 | orchestrator | Friday 29 August 2025 17:04:11 +0000 (0:00:00.072) 0:00:03.442 ********* 2025-08-29 17:04:16.631714 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager 2025-08-29 17:04:16.631724 | orchestrator | 2025-08-29 17:04:16.631733 | orchestrator | TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] *** 2025-08-29 17:04:16.631743 | orchestrator | Friday 29 August 2025 17:04:11 +0000 (0:00:00.050) 0:00:03.492 ********* 2025-08-29 17:04:16.631753 | orchestrator | ok: [testbed-manager] 2025-08-29 17:04:16.631762 | orchestrator | 2025-08-29 17:04:16.631772 | orchestrator | TASK [osism.commons.resolvconf : Install package systemd-resolved] ************* 2025-08-29 17:04:16.631781 | orchestrator | Friday 29 August 2025 17:04:12 +0000 (0:00:00.758) 0:00:04.250 ********* 2025-08-29 17:04:16.631791 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:04:16.631800 | orchestrator | 2025-08-29 17:04:16.631810 | orchestrator | TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] ***** 2025-08-29 17:04:16.631820 | orchestrator | Friday 29 August 2025 17:04:12 +0000 (0:00:00.057) 0:00:04.308 ********* 2025-08-29 17:04:16.631829 | orchestrator | ok: [testbed-manager] 2025-08-29 17:04:16.631839 | orchestrator | 2025-08-29 17:04:16.631849 | orchestrator | TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] ******* 2025-08-29 17:04:16.631859 | orchestrator | Friday 29 August 2025 17:04:12 +0000 (0:00:00.441) 0:00:04.750 ********* 2025-08-29 17:04:16.631869 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:04:16.631878 | orchestrator | 2025-08-29 17:04:16.631888 | orchestrator | TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] *** 2025-08-29 17:04:16.631899 | orchestrator | Friday 29 August 2025 17:04:12 +0000 (0:00:00.078) 0:00:04.828 ********* 2025-08-29 17:04:16.631908 | orchestrator | changed: [testbed-manager] 2025-08-29 17:04:16.631918 | orchestrator | 2025-08-29 17:04:16.631927 | orchestrator | TASK [osism.commons.resolvconf : Copy configuration files] ********************* 2025-08-29 17:04:16.631937 | orchestrator | Friday 29 August 2025 17:04:13 +0000 (0:00:00.503) 0:00:05.332 ********* 2025-08-29 17:04:16.631947 | orchestrator | changed: [testbed-manager] 2025-08-29 17:04:16.632010 | orchestrator | 2025-08-29 17:04:16.632023 | orchestrator | TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ******** 2025-08-29 17:04:16.632035 | orchestrator | Friday 29 August 2025 17:04:14 +0000 (0:00:01.029) 0:00:06.361 ********* 2025-08-29 17:04:16.632046 | orchestrator | ok: [testbed-manager] 2025-08-29 17:04:16.632056 | orchestrator | 2025-08-29 17:04:16.632068 | orchestrator | TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] *** 2025-08-29 17:04:16.632078 | orchestrator | Friday 29 August 2025 17:04:15 +0000 (0:00:00.927) 0:00:07.289 ********* 2025-08-29 17:04:16.632090 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager 2025-08-29 17:04:16.632101 | orchestrator | 2025-08-29 17:04:16.632112 | orchestrator | TASK [osism.commons.resolvconf : Restart systemd-resolved service] ************* 2025-08-29 17:04:16.632122 | orchestrator | Friday 29 August 2025 17:04:15 +0000 (0:00:00.069) 0:00:07.359 ********* 2025-08-29 17:04:16.632133 | orchestrator | changed: [testbed-manager] 2025-08-29 17:04:16.632143 | orchestrator | 2025-08-29 17:04:16.632155 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:04:16.632176 | orchestrator | testbed-manager : ok=10  changed=3  unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 17:04:16.632187 | orchestrator | 2025-08-29 17:04:16.632198 | orchestrator | 2025-08-29 17:04:16.632209 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:04:16.632220 | orchestrator | Friday 29 August 2025 17:04:16 +0000 (0:00:01.042) 0:00:08.402 ********* 2025-08-29 17:04:16.632231 | orchestrator | =============================================================================== 2025-08-29 17:04:16.632242 | orchestrator | Gathering Facts --------------------------------------------------------- 3.20s 2025-08-29 17:04:16.632252 | orchestrator | osism.commons.resolvconf : Restart systemd-resolved service ------------- 1.04s 2025-08-29 17:04:16.632263 | orchestrator | osism.commons.resolvconf : Copy configuration files --------------------- 1.03s 2025-08-29 17:04:16.632274 | orchestrator | osism.commons.resolvconf : Start/enable systemd-resolved service -------- 0.93s 2025-08-29 17:04:16.632284 | orchestrator | osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 0.76s 2025-08-29 17:04:16.632295 | orchestrator | osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.50s 2025-08-29 17:04:16.632320 | orchestrator | osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 0.44s 2025-08-29 17:04:16.632332 | orchestrator | osism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.08s 2025-08-29 17:04:16.632342 | orchestrator | osism.commons.resolvconf : Include resolvconf tasks --------------------- 0.07s 2025-08-29 17:04:16.632353 | orchestrator | osism.commons.resolvconf : Include distribution specific configuration tasks --- 0.07s 2025-08-29 17:04:16.632364 | orchestrator | osism.commons.resolvconf : Install package systemd-resolved ------------- 0.06s 2025-08-29 17:04:16.632374 | orchestrator | osism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.06s 2025-08-29 17:04:16.632383 | orchestrator | osism.commons.resolvconf : Include distribution specific installation tasks --- 0.05s 2025-08-29 17:04:16.867319 | orchestrator | + osism apply sshconfig 2025-08-29 17:04:28.847259 | orchestrator | 2025-08-29 17:04:28 | INFO  | Task aa037f34-fb14-445d-b8a3-1bf82bf37b60 (sshconfig) was prepared for execution. 2025-08-29 17:04:28.847364 | orchestrator | 2025-08-29 17:04:28 | INFO  | It takes a moment until task aa037f34-fb14-445d-b8a3-1bf82bf37b60 (sshconfig) has been started and output is visible here. 2025-08-29 17:04:39.270550 | orchestrator | 2025-08-29 17:04:39.270647 | orchestrator | PLAY [Apply role sshconfig] **************************************************** 2025-08-29 17:04:39.270662 | orchestrator | 2025-08-29 17:04:39.270673 | orchestrator | TASK [osism.commons.sshconfig : Get home directory of operator user] *********** 2025-08-29 17:04:39.270683 | orchestrator | Friday 29 August 2025 17:04:32 +0000 (0:00:00.118) 0:00:00.118 ********* 2025-08-29 17:04:39.270721 | orchestrator | ok: [testbed-manager] 2025-08-29 17:04:39.270733 | orchestrator | 2025-08-29 17:04:39.270743 | orchestrator | TASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ******************** 2025-08-29 17:04:39.270753 | orchestrator | Friday 29 August 2025 17:04:32 +0000 (0:00:00.475) 0:00:00.594 ********* 2025-08-29 17:04:39.270763 | orchestrator | changed: [testbed-manager] 2025-08-29 17:04:39.270773 | orchestrator | 2025-08-29 17:04:39.270783 | orchestrator | TASK [osism.commons.sshconfig : Ensure config for each host exist] ************* 2025-08-29 17:04:39.270793 | orchestrator | Friday 29 August 2025 17:04:33 +0000 (0:00:00.441) 0:00:01.036 ********* 2025-08-29 17:04:39.270803 | orchestrator | changed: [testbed-manager] => (item=testbed-manager) 2025-08-29 17:04:39.270812 | orchestrator | changed: [testbed-manager] => (item=testbed-node-0) 2025-08-29 17:04:39.270822 | orchestrator | changed: [testbed-manager] => (item=testbed-node-1) 2025-08-29 17:04:39.270832 | orchestrator | changed: [testbed-manager] => (item=testbed-node-2) 2025-08-29 17:04:39.270842 | orchestrator | changed: [testbed-manager] => (item=testbed-node-3) 2025-08-29 17:04:39.270851 | orchestrator | changed: [testbed-manager] => (item=testbed-node-4) 2025-08-29 17:04:39.270861 | orchestrator | changed: [testbed-manager] => (item=testbed-node-5) 2025-08-29 17:04:39.270870 | orchestrator | 2025-08-29 17:04:39.270880 | orchestrator | TASK [osism.commons.sshconfig : Add extra config] ****************************** 2025-08-29 17:04:39.270890 | orchestrator | Friday 29 August 2025 17:04:38 +0000 (0:00:05.045) 0:00:06.081 ********* 2025-08-29 17:04:39.270899 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:04:39.270909 | orchestrator | 2025-08-29 17:04:39.270919 | orchestrator | TASK [osism.commons.sshconfig : Assemble ssh config] *************************** 2025-08-29 17:04:39.270929 | orchestrator | Friday 29 August 2025 17:04:38 +0000 (0:00:00.056) 0:00:06.137 ********* 2025-08-29 17:04:39.270938 | orchestrator | changed: [testbed-manager] 2025-08-29 17:04:39.270948 | orchestrator | 2025-08-29 17:04:39.270958 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:04:39.270969 | orchestrator | testbed-manager : ok=4  changed=3  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:04:39.271015 | orchestrator | 2025-08-29 17:04:39.271026 | orchestrator | 2025-08-29 17:04:39.271035 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:04:39.271045 | orchestrator | Friday 29 August 2025 17:04:39 +0000 (0:00:00.565) 0:00:06.703 ********* 2025-08-29 17:04:39.271071 | orchestrator | =============================================================================== 2025-08-29 17:04:39.271082 | orchestrator | osism.commons.sshconfig : Ensure config for each host exist ------------- 5.05s 2025-08-29 17:04:39.271092 | orchestrator | osism.commons.sshconfig : Assemble ssh config --------------------------- 0.57s 2025-08-29 17:04:39.271102 | orchestrator | osism.commons.sshconfig : Get home directory of operator user ----------- 0.48s 2025-08-29 17:04:39.271111 | orchestrator | osism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.44s 2025-08-29 17:04:39.271121 | orchestrator | osism.commons.sshconfig : Add extra config ------------------------------ 0.06s 2025-08-29 17:04:39.500764 | orchestrator | + osism apply known-hosts 2025-08-29 17:04:51.322903 | orchestrator | 2025-08-29 17:04:51 | INFO  | Task 831fefd6-a3f2-4c72-bd35-dd7cdc00ea47 (known-hosts) was prepared for execution. 2025-08-29 17:04:51.323038 | orchestrator | 2025-08-29 17:04:51 | INFO  | It takes a moment until task 831fefd6-a3f2-4c72-bd35-dd7cdc00ea47 (known-hosts) has been started and output is visible here. 2025-08-29 17:05:06.910155 | orchestrator | 2025-08-29 17:05:06.910241 | orchestrator | PLAY [Apply role known_hosts] ************************************************** 2025-08-29 17:05:06.910256 | orchestrator | 2025-08-29 17:05:06.910267 | orchestrator | TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] *** 2025-08-29 17:05:06.910278 | orchestrator | Friday 29 August 2025 17:04:55 +0000 (0:00:00.121) 0:00:00.121 ********* 2025-08-29 17:05:06.910288 | orchestrator | ok: [testbed-manager] => (item=testbed-manager) 2025-08-29 17:05:06.910314 | orchestrator | ok: [testbed-manager] => (item=testbed-node-0) 2025-08-29 17:05:06.910325 | orchestrator | ok: [testbed-manager] => (item=testbed-node-1) 2025-08-29 17:05:06.910335 | orchestrator | ok: [testbed-manager] => (item=testbed-node-2) 2025-08-29 17:05:06.910345 | orchestrator | ok: [testbed-manager] => (item=testbed-node-3) 2025-08-29 17:05:06.910354 | orchestrator | ok: [testbed-manager] => (item=testbed-node-4) 2025-08-29 17:05:06.910364 | orchestrator | ok: [testbed-manager] => (item=testbed-node-5) 2025-08-29 17:05:06.910373 | orchestrator | 2025-08-29 17:05:06.910383 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] *** 2025-08-29 17:05:06.910393 | orchestrator | Friday 29 August 2025 17:05:00 +0000 (0:00:05.624) 0:00:05.745 ********* 2025-08-29 17:05:06.910407 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager) 2025-08-29 17:05:06.910425 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0) 2025-08-29 17:05:06.910441 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1) 2025-08-29 17:05:06.910457 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2) 2025-08-29 17:05:06.910473 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3) 2025-08-29 17:05:06.910489 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4) 2025-08-29 17:05:06.910506 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5) 2025-08-29 17:05:06.910524 | orchestrator | 2025-08-29 17:05:06.910541 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:06.910557 | orchestrator | Friday 29 August 2025 17:05:00 +0000 (0:00:00.152) 0:00:05.897 ********* 2025-08-29 17:05:06.910567 | orchestrator | changed: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKR3IaKLyatpCdsNoV34bcOpyFG03VIeBKhrZY91iC/o22DorzWeaOUQU3Fg/jfjC+ZLVtUWzqx0WrLr/mbE9yM=) 2025-08-29 17:05:06.910582 | orchestrator | changed: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDmFTBfhSLekk7+FkkGBpNqziTQbbEEO1033GR+jkFtwejA7RJWLb4KCTsiAhX5A1702rVkXLgo2+aSroA6YdzlyIlRmw2FfWynrbynBDpnbQS9qhG46964UMhTpje9PX+NdSTKT7iSDyQZTxGbbrVXLcPaM2VkyqgM9zT5+MoCNUQWl95PhWzxT2FrIxX8IQdff7gwv/GfK7KmFqreIoxzov5IJjwb9zjlBJCgjaQisHRqB1e7UgjFtMKMJaOuXnDZHiYyJc5JsAqT46O3lVz7+FOoKjSTRU3kGD7RO7aB90PweP7H7tnelzMUqzsozUqAjzlGrkwmI1c1MjksCUynaLiM2bbfd+J13OCQ2oSnFJeVdh/ivtE2W/b5NcrWhKux9zgybfqoEaTaFeDF3LXkEVZtrpfrLJ7E8FvdlAUVO5QCl/kLCCsIenO7BLu4VC5MSTpYlDlttt4qmEXYp481mJv5kNH23Nuh8E7z5LmWRMXczePdb/EMln3+7QQf1nU=) 2025-08-29 17:05:06.910650 | orchestrator | changed: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDiLhlJDC6jRL31+/yBNhor+6qXLrRpckZDSO8rfvTEK) 2025-08-29 17:05:06.910663 | orchestrator | 2025-08-29 17:05:06.910677 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:06.910689 | orchestrator | Friday 29 August 2025 17:05:01 +0000 (0:00:01.131) 0:00:07.028 ********* 2025-08-29 17:05:06.910721 | orchestrator | changed: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDjuuKs+oeNnGfKYXIr/bQIDKEpOsxaZhjpzBXST7IMl7bc2Yni9kh9nQcSRreTtILOMNHg8UKVwtPoOgFqPmmkYztO6naZZlnrkB4Y3jgW1snyun5xrTMsgzy3E3HDMt6xYVPWziX5qZMWXIEPIheTRW+rvXQoFbKCio4ZCXfR35UQ/RuvGUAr3yLPI4mg4a7GZwOsdeAraEIgKAAEg0kuwtSJs/6/3aXe1TWB+SHtOcLI490Cl3vUKXpjnPvZsc+TOqvPIYWw+MSyrsZQLt0MxJmm1yOJt0ovKsikAD4r1Sl8TqcsUz2cI/cJUPJZKO5FmB0CtmVQReYPDQWTrF/PL2dsAI+qpArz3TjtoRKC/j5h5yy1oLnt9QQaBwyUsd3IIXG4GUM2Sik0CKgpmVmhfJAGH96peWM9gr2VbwWosYT/Dh+T0iZbvPYWOKXt9HTaaEaOYc4b7tNw43A6ul/qqIDMmdCs5ZDS7/bNCOA1i+Cui2XYBHsF0ZaHq/nakzs=) 2025-08-29 17:05:06.910746 | orchestrator | changed: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILm9vQJFXvj97+4ChHl/wu5krM7jCfqQwEtXcQXhPhqZ) 2025-08-29 17:05:06.910760 | orchestrator | changed: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBaJRUCuPnBLRE7wJN7uWGxCChq+XMOzQGMCPxP/vImhychnDurpDKstR4QprFAGteMGrZGw9T2mB4oCooDDlG4=) 2025-08-29 17:05:06.910773 | orchestrator | 2025-08-29 17:05:06.910786 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:06.910799 | orchestrator | Friday 29 August 2025 17:05:02 +0000 (0:00:01.005) 0:00:08.033 ********* 2025-08-29 17:05:06.910812 | orchestrator | changed: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC8INcuoqWh7sNzrvLej+fnmgB7G8qJ1VxMwx4Y9oQJhCxM7VQA3lH3PyzrR7BkZeT4M5NLT1P1mBCCbskbn0pHppqhRTb+dFVWMl53/36tpEs7r59Ascgtd4Nno3zoaW/m/kYXsFKXfYIcCjbaQdSQUKkws99SilWD10H/oglvs9WBtKtwvF8GVbZkNx3wOzpuclISjbCLzDVKzf6Cb5IB2ymsTTVKCq/zRmPQJvAy48dL0jXyK993c1nCyZC9+V7hNZJIK25ScyZRcWIKCF4fgV9fiMNovP97MJl8XWBgfNfepp62J0IQ9lI5KbKKCeD+blGrq3wFubO73/JlUJtgXEwvr0UV8hAiKxWUvswxzOLxA9YabKFL1rSTqoLdmDPp3IaHkLzfXQKYK6gW8RwRoIJEzm/dXVERCrXMYiSrUaGAHa5Yx4rmmM5/XmMQH4VBXC0kwHY4UqOBTUSoeRUUthFMqlhg4mhsxVpMI6CG09OsIvOvBs4omhyXv0F2IIU=) 2025-08-29 17:05:06.910825 | orchestrator | changed: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBrPZBhI6GBERrCCvn3GuwErzW4m64UY3mhbtWIzeKYC) 2025-08-29 17:05:06.910838 | orchestrator | changed: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBK5enSWKxi/QNFV2Wp3Kx85z4urSf5goy0LEFizQq5j6SQCf22WxJOSEdxVeL4KdqPKxfnjR3kfCsm7bEOGu/UU=) 2025-08-29 17:05:06.910850 | orchestrator | 2025-08-29 17:05:06.910863 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:06.910876 | orchestrator | Friday 29 August 2025 17:05:04 +0000 (0:00:01.032) 0:00:09.066 ********* 2025-08-29 17:05:06.910889 | orchestrator | changed: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCoZfovDbqmefMoLjKYkPjELdMSgYZGmiHQ/w8shtI4wIbdvWs0f6mLErpDYinly6s1Ykf7j7L6ZuHPzGQfhmFqGMEAGd7PKSgqQuXTVOoeXuKb3gY1z0xQIpLceDnQD1DavfiGCt/LnHct5mvQjfU5KI+23XmRtqrDL4RDDuV3l3hrIE35R0UpefcE7aS+Vdziw/2y4goDsiv2WQuosy14UIrI7qFpy+LN/nEho9gF0yh7RnGm0Mh9VERSQyUopgeET5mjSMBxt8A226gjVzcZxKuXbAlpmkTTbkN++/0SI1e8IzU5qwUM5bjBQU1YQ6HzVeTWK7yYp/funar5qqgJNRSah/BaySvK9pyhNrP85SUJNbiE8kFpoRakYhKRhbMH3JrPS7+97zNwVCM92PZaTjwVnvIKPn9ipqH3jrIOZuERLL5x/ALVKaQsnO7W9WsDpC+ZB9GiiWlU1IyO2T8+xvTz74KY0hSE/gqvuBqMAbBHoEl13vmnM3/yR5hsQw8=) 2025-08-29 17:05:06.910903 | orchestrator | changed: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGTwiQROaZweOIJivZLs2fNPoVC7zSOWtsdaHYUvP4GMp6LjHwzEVyAOD7pi98lCgXQpVrLKlsRXqVP5xUqcVh0=) 2025-08-29 17:05:06.910916 | orchestrator | changed: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBuAn23/yGgNEL6kUivAVFIIVNlys2VO0MJLQQumdc8d) 2025-08-29 17:05:06.910929 | orchestrator | 2025-08-29 17:05:06.910942 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:06.910955 | orchestrator | Friday 29 August 2025 17:05:05 +0000 (0:00:01.055) 0:00:10.122 ********* 2025-08-29 17:05:06.910970 | orchestrator | changed: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDzoHEeALAumuU+HWAxoHhJMhk+mZYgyh0kNh/caDg1aIiVNiPGyMkZUeMRmAy1bLA5mDoedj1SktxGTS2a7q3xcvBp6wl2wsI6AjjP/l/aKzjbk5+aCBfLSuV0Am6bPQV5aFixzLrlSMD89vf8LZ/N2rfLiipL4rSYTT1KQxj22WNxNH7USRupG//i3vKviYoIxOnvrKtr2HruCLTWg0vzG5T3JS5hc1u+V+Nuc6OrWW8lSrfq6TfVX5luo62A7P1ydwnlX7p0mZLLK9bvEKzfjzSt/yycHqqUK6NUzd3777Z7bN6RjyXq7UNLjBRv4Z2eUgYEjoEMjZSrSG559je8Bl4AI5E69gc1Fp6hlTg6shcTYTpFYMbFAJ1nt42+2X/Ao0rVjsBDE5FZ5YDX1Vpn0OdYm557GXeN8WBzD6iYK3J8/wBU6k6O8ivCBfLTVR5bDOS2/21XPrwndoSBJVTJcguYvOmu7C46P9DAO+AS1gXVBsT/bFz0fb9N07curu8=) 2025-08-29 17:05:06.911024 | orchestrator | changed: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDzMwKTO1lkH1O5ZXEMKWPwrDv3sKMh+DydvjIXDtLoF2lJQeoVpYwc1i4p+AmBCQfFgNsbgrlxo2MHimvv3TfQ=) 2025-08-29 17:05:06.911042 | orchestrator | changed: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFG0clm2IAW/XYR3BS1zuNqZugRhEbvY7vp6LnO9IIl5) 2025-08-29 17:05:06.911053 | orchestrator | 2025-08-29 17:05:06.911064 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:06.911075 | orchestrator | Friday 29 August 2025 17:05:05 +0000 (0:00:00.911) 0:00:11.033 ********* 2025-08-29 17:05:06.911094 | orchestrator | changed: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIjKc+DWWBRIo3PF3WwuabSoDWU1o9KIUwl+4h8klSeNkovqMX/cS42Kaypbr/HzJN5jqqLNz/LQvXIGK6fn/y4=) 2025-08-29 17:05:16.873934 | orchestrator | changed: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCuN/tICMyQMpzTG8GI5cMlw1cMBacHeX1bqZrYuSDQO+ITVqyLfodK6t6p7UHC77ji8hDVzanfuzogOBB+5NZmgbnDGNliM55Rdd0esIwnVDvk7EwDj3YL4jdHXKtkbq+bXuLvW+bxbB8hD1NXq8GQe4fIyw8hVmNXFb7C7v5SOAB2eQSe/M/lzWFKMZP6U4EocawvvXVczDaYkV2Pnjx5BKVQ2S2pP95DgQrZOCJhb2HNjjkwTQzFEittOttgA5/ZiHRRtwi89GCusbYSmx9kUdWtrz/EKl5ELB2PrtLZteH+lb+l3wVNeMqyOAhqEj4o52gQKKBv3K2HbNfBOx69CglckQRl8bfCBXDa+QGOtBt264QP/TCh93R1GDjVK56rd7VwEvZok88he9W5oIxnNcjwQO1zckq8QorKGMExuB9lsFOvGzwTKFuay2jRYnFHrMVMxdUMIHXZ22fGEpKkpa/KjNev4/VvJ4IVapA55E6bHXxx0ZFFxeKVUaWeTQU=) 2025-08-29 17:05:16.874182 | orchestrator | changed: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOLootPoNs+u7I7m5ZSvFcAtAm8qVJtj/IxHHr4g6UP4) 2025-08-29 17:05:16.874206 | orchestrator | 2025-08-29 17:05:16.874220 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:16.874233 | orchestrator | Friday 29 August 2025 17:05:06 +0000 (0:00:00.918) 0:00:11.952 ********* 2025-08-29 17:05:16.874246 | orchestrator | changed: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDB0HvCKs5kGR4JXJXVkhK4wPI06+/R9SWhUcFLLtULI/Atc6/0ug8Xdi3iW189x1xc8St8/T/jlb/fwSZnvBERJWF38CQ0OQtwbmTjWx8PoOGWPESMv/uWV32ERp6w8F7l2IvRuinUhUSkvJxWo0/WKfA8SZ/XdKFVrZdYRmCkpwKBkzSDCOGhCMKoYUi0tlKR5EWsBrTNLPcnqnwT/QibOCCSDI9UclaXR1MUQ/oF7F6TddkFm5WuORnjpkiYZlntQKAkvMRP4J3/pbEpYNfB1ETfBiJZ0kVsehnXBEPMvfQ0a0ju1iVG0+MCCR2UJZXnjAXow5JP2W7kxYdCuWP6jInMzEzLGiEFcxbpt2noVOJNtDF/vCa9lUQ60yuD6p127FFRTrCZxQchMVKhaeKCua9vt6EkjZuEYFummQMa0r87PajeejZlM8qMHJ1ikb956o+QKy1jTLW7Q9BFic1g9kWZQLSW58sGRtmZbu5QJGFqKdZPioc44dowyOgnw2E=) 2025-08-29 17:05:16.874259 | orchestrator | changed: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEoyeWc8CpWlK6hRRk/tva9wYhWpBeUSGJl1HSxkkjuPsQX/YlZMWnfwSNMq17DoXInlpYIB14zwXgh7oc1pi+E=) 2025-08-29 17:05:16.874272 | orchestrator | changed: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAICN3yuPKx9eiPscoGb67XBh2jqcdiCsaMvMaXvYlcRk7) 2025-08-29 17:05:16.874284 | orchestrator | 2025-08-29 17:05:16.874295 | orchestrator | TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] *** 2025-08-29 17:05:16.874307 | orchestrator | Friday 29 August 2025 17:05:07 +0000 (0:00:00.933) 0:00:12.886 ********* 2025-08-29 17:05:16.874318 | orchestrator | ok: [testbed-manager] => (item=testbed-manager) 2025-08-29 17:05:16.874330 | orchestrator | ok: [testbed-manager] => (item=testbed-node-0) 2025-08-29 17:05:16.874341 | orchestrator | ok: [testbed-manager] => (item=testbed-node-1) 2025-08-29 17:05:16.874352 | orchestrator | ok: [testbed-manager] => (item=testbed-node-2) 2025-08-29 17:05:16.874363 | orchestrator | ok: [testbed-manager] => (item=testbed-node-3) 2025-08-29 17:05:16.874396 | orchestrator | ok: [testbed-manager] => (item=testbed-node-4) 2025-08-29 17:05:16.874407 | orchestrator | ok: [testbed-manager] => (item=testbed-node-5) 2025-08-29 17:05:16.874418 | orchestrator | 2025-08-29 17:05:16.874429 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] *** 2025-08-29 17:05:16.874441 | orchestrator | Friday 29 August 2025 17:05:12 +0000 (0:00:04.994) 0:00:17.880 ********* 2025-08-29 17:05:16.874454 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager) 2025-08-29 17:05:16.874470 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0) 2025-08-29 17:05:16.874482 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1) 2025-08-29 17:05:16.874495 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2) 2025-08-29 17:05:16.874507 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3) 2025-08-29 17:05:16.874519 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4) 2025-08-29 17:05:16.874532 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5) 2025-08-29 17:05:16.874544 | orchestrator | 2025-08-29 17:05:16.874575 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:16.874589 | orchestrator | Friday 29 August 2025 17:05:12 +0000 (0:00:00.146) 0:00:18.027 ********* 2025-08-29 17:05:16.874600 | orchestrator | changed: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDiLhlJDC6jRL31+/yBNhor+6qXLrRpckZDSO8rfvTEK) 2025-08-29 17:05:16.874622 | orchestrator | changed: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDmFTBfhSLekk7+FkkGBpNqziTQbbEEO1033GR+jkFtwejA7RJWLb4KCTsiAhX5A1702rVkXLgo2+aSroA6YdzlyIlRmw2FfWynrbynBDpnbQS9qhG46964UMhTpje9PX+NdSTKT7iSDyQZTxGbbrVXLcPaM2VkyqgM9zT5+MoCNUQWl95PhWzxT2FrIxX8IQdff7gwv/GfK7KmFqreIoxzov5IJjwb9zjlBJCgjaQisHRqB1e7UgjFtMKMJaOuXnDZHiYyJc5JsAqT46O3lVz7+FOoKjSTRU3kGD7RO7aB90PweP7H7tnelzMUqzsozUqAjzlGrkwmI1c1MjksCUynaLiM2bbfd+J13OCQ2oSnFJeVdh/ivtE2W/b5NcrWhKux9zgybfqoEaTaFeDF3LXkEVZtrpfrLJ7E8FvdlAUVO5QCl/kLCCsIenO7BLu4VC5MSTpYlDlttt4qmEXYp481mJv5kNH23Nuh8E7z5LmWRMXczePdb/EMln3+7QQf1nU=) 2025-08-29 17:05:16.874635 | orchestrator | changed: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBKR3IaKLyatpCdsNoV34bcOpyFG03VIeBKhrZY91iC/o22DorzWeaOUQU3Fg/jfjC+ZLVtUWzqx0WrLr/mbE9yM=) 2025-08-29 17:05:16.874646 | orchestrator | 2025-08-29 17:05:16.874657 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:16.874668 | orchestrator | Friday 29 August 2025 17:05:13 +0000 (0:00:00.902) 0:00:18.930 ********* 2025-08-29 17:05:16.874679 | orchestrator | changed: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBBaJRUCuPnBLRE7wJN7uWGxCChq+XMOzQGMCPxP/vImhychnDurpDKstR4QprFAGteMGrZGw9T2mB4oCooDDlG4=) 2025-08-29 17:05:16.874691 | orchestrator | changed: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDjuuKs+oeNnGfKYXIr/bQIDKEpOsxaZhjpzBXST7IMl7bc2Yni9kh9nQcSRreTtILOMNHg8UKVwtPoOgFqPmmkYztO6naZZlnrkB4Y3jgW1snyun5xrTMsgzy3E3HDMt6xYVPWziX5qZMWXIEPIheTRW+rvXQoFbKCio4ZCXfR35UQ/RuvGUAr3yLPI4mg4a7GZwOsdeAraEIgKAAEg0kuwtSJs/6/3aXe1TWB+SHtOcLI490Cl3vUKXpjnPvZsc+TOqvPIYWw+MSyrsZQLt0MxJmm1yOJt0ovKsikAD4r1Sl8TqcsUz2cI/cJUPJZKO5FmB0CtmVQReYPDQWTrF/PL2dsAI+qpArz3TjtoRKC/j5h5yy1oLnt9QQaBwyUsd3IIXG4GUM2Sik0CKgpmVmhfJAGH96peWM9gr2VbwWosYT/Dh+T0iZbvPYWOKXt9HTaaEaOYc4b7tNw43A6ul/qqIDMmdCs5ZDS7/bNCOA1i+Cui2XYBHsF0ZaHq/nakzs=) 2025-08-29 17:05:16.874713 | orchestrator | changed: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAILm9vQJFXvj97+4ChHl/wu5krM7jCfqQwEtXcQXhPhqZ) 2025-08-29 17:05:16.874724 | orchestrator | 2025-08-29 17:05:16.874735 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:16.874746 | orchestrator | Friday 29 August 2025 17:05:14 +0000 (0:00:00.998) 0:00:19.928 ********* 2025-08-29 17:05:16.874757 | orchestrator | changed: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBK5enSWKxi/QNFV2Wp3Kx85z4urSf5goy0LEFizQq5j6SQCf22WxJOSEdxVeL4KdqPKxfnjR3kfCsm7bEOGu/UU=) 2025-08-29 17:05:16.874768 | orchestrator | changed: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC8INcuoqWh7sNzrvLej+fnmgB7G8qJ1VxMwx4Y9oQJhCxM7VQA3lH3PyzrR7BkZeT4M5NLT1P1mBCCbskbn0pHppqhRTb+dFVWMl53/36tpEs7r59Ascgtd4Nno3zoaW/m/kYXsFKXfYIcCjbaQdSQUKkws99SilWD10H/oglvs9WBtKtwvF8GVbZkNx3wOzpuclISjbCLzDVKzf6Cb5IB2ymsTTVKCq/zRmPQJvAy48dL0jXyK993c1nCyZC9+V7hNZJIK25ScyZRcWIKCF4fgV9fiMNovP97MJl8XWBgfNfepp62J0IQ9lI5KbKKCeD+blGrq3wFubO73/JlUJtgXEwvr0UV8hAiKxWUvswxzOLxA9YabKFL1rSTqoLdmDPp3IaHkLzfXQKYK6gW8RwRoIJEzm/dXVERCrXMYiSrUaGAHa5Yx4rmmM5/XmMQH4VBXC0kwHY4UqOBTUSoeRUUthFMqlhg4mhsxVpMI6CG09OsIvOvBs4omhyXv0F2IIU=) 2025-08-29 17:05:16.874779 | orchestrator | changed: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBrPZBhI6GBERrCCvn3GuwErzW4m64UY3mhbtWIzeKYC) 2025-08-29 17:05:16.874790 | orchestrator | 2025-08-29 17:05:16.874801 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:16.874812 | orchestrator | Friday 29 August 2025 17:05:15 +0000 (0:00:00.999) 0:00:20.928 ********* 2025-08-29 17:05:16.874823 | orchestrator | changed: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGTwiQROaZweOIJivZLs2fNPoVC7zSOWtsdaHYUvP4GMp6LjHwzEVyAOD7pi98lCgXQpVrLKlsRXqVP5xUqcVh0=) 2025-08-29 17:05:16.874853 | orchestrator | changed: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCoZfovDbqmefMoLjKYkPjELdMSgYZGmiHQ/w8shtI4wIbdvWs0f6mLErpDYinly6s1Ykf7j7L6ZuHPzGQfhmFqGMEAGd7PKSgqQuXTVOoeXuKb3gY1z0xQIpLceDnQD1DavfiGCt/LnHct5mvQjfU5KI+23XmRtqrDL4RDDuV3l3hrIE35R0UpefcE7aS+Vdziw/2y4goDsiv2WQuosy14UIrI7qFpy+LN/nEho9gF0yh7RnGm0Mh9VERSQyUopgeET5mjSMBxt8A226gjVzcZxKuXbAlpmkTTbkN++/0SI1e8IzU5qwUM5bjBQU1YQ6HzVeTWK7yYp/funar5qqgJNRSah/BaySvK9pyhNrP85SUJNbiE8kFpoRakYhKRhbMH3JrPS7+97zNwVCM92PZaTjwVnvIKPn9ipqH3jrIOZuERLL5x/ALVKaQsnO7W9WsDpC+ZB9GiiWlU1IyO2T8+xvTz74KY0hSE/gqvuBqMAbBHoEl13vmnM3/yR5hsQw8=) 2025-08-29 17:05:20.903851 | orchestrator | changed: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIBuAn23/yGgNEL6kUivAVFIIVNlys2VO0MJLQQumdc8d) 2025-08-29 17:05:20.903967 | orchestrator | 2025-08-29 17:05:20.904022 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:20.904043 | orchestrator | Friday 29 August 2025 17:05:16 +0000 (0:00:00.992) 0:00:21.921 ********* 2025-08-29 17:05:20.904064 | orchestrator | changed: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDzMwKTO1lkH1O5ZXEMKWPwrDv3sKMh+DydvjIXDtLoF2lJQeoVpYwc1i4p+AmBCQfFgNsbgrlxo2MHimvv3TfQ=) 2025-08-29 17:05:20.904088 | orchestrator | changed: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDzoHEeALAumuU+HWAxoHhJMhk+mZYgyh0kNh/caDg1aIiVNiPGyMkZUeMRmAy1bLA5mDoedj1SktxGTS2a7q3xcvBp6wl2wsI6AjjP/l/aKzjbk5+aCBfLSuV0Am6bPQV5aFixzLrlSMD89vf8LZ/N2rfLiipL4rSYTT1KQxj22WNxNH7USRupG//i3vKviYoIxOnvrKtr2HruCLTWg0vzG5T3JS5hc1u+V+Nuc6OrWW8lSrfq6TfVX5luo62A7P1ydwnlX7p0mZLLK9bvEKzfjzSt/yycHqqUK6NUzd3777Z7bN6RjyXq7UNLjBRv4Z2eUgYEjoEMjZSrSG559je8Bl4AI5E69gc1Fp6hlTg6shcTYTpFYMbFAJ1nt42+2X/Ao0rVjsBDE5FZ5YDX1Vpn0OdYm557GXeN8WBzD6iYK3J8/wBU6k6O8ivCBfLTVR5bDOS2/21XPrwndoSBJVTJcguYvOmu7C46P9DAO+AS1gXVBsT/bFz0fb9N07curu8=) 2025-08-29 17:05:20.904143 | orchestrator | changed: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIFG0clm2IAW/XYR3BS1zuNqZugRhEbvY7vp6LnO9IIl5) 2025-08-29 17:05:20.904163 | orchestrator | 2025-08-29 17:05:20.904181 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:20.904192 | orchestrator | Friday 29 August 2025 17:05:17 +0000 (0:00:01.032) 0:00:22.953 ********* 2025-08-29 17:05:20.904204 | orchestrator | changed: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCuN/tICMyQMpzTG8GI5cMlw1cMBacHeX1bqZrYuSDQO+ITVqyLfodK6t6p7UHC77ji8hDVzanfuzogOBB+5NZmgbnDGNliM55Rdd0esIwnVDvk7EwDj3YL4jdHXKtkbq+bXuLvW+bxbB8hD1NXq8GQe4fIyw8hVmNXFb7C7v5SOAB2eQSe/M/lzWFKMZP6U4EocawvvXVczDaYkV2Pnjx5BKVQ2S2pP95DgQrZOCJhb2HNjjkwTQzFEittOttgA5/ZiHRRtwi89GCusbYSmx9kUdWtrz/EKl5ELB2PrtLZteH+lb+l3wVNeMqyOAhqEj4o52gQKKBv3K2HbNfBOx69CglckQRl8bfCBXDa+QGOtBt264QP/TCh93R1GDjVK56rd7VwEvZok88he9W5oIxnNcjwQO1zckq8QorKGMExuB9lsFOvGzwTKFuay2jRYnFHrMVMxdUMIHXZ22fGEpKkpa/KjNev4/VvJ4IVapA55E6bHXxx0ZFFxeKVUaWeTQU=) 2025-08-29 17:05:20.904232 | orchestrator | changed: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBIjKc+DWWBRIo3PF3WwuabSoDWU1o9KIUwl+4h8klSeNkovqMX/cS42Kaypbr/HzJN5jqqLNz/LQvXIGK6fn/y4=) 2025-08-29 17:05:20.904244 | orchestrator | changed: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIOLootPoNs+u7I7m5ZSvFcAtAm8qVJtj/IxHHr4g6UP4) 2025-08-29 17:05:20.904255 | orchestrator | 2025-08-29 17:05:20.904267 | orchestrator | TASK [osism.commons.known_hosts : Write scanned known_hosts entries] *********** 2025-08-29 17:05:20.904278 | orchestrator | Friday 29 August 2025 17:05:18 +0000 (0:00:01.010) 0:00:23.964 ********* 2025-08-29 17:05:20.904289 | orchestrator | changed: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDB0HvCKs5kGR4JXJXVkhK4wPI06+/R9SWhUcFLLtULI/Atc6/0ug8Xdi3iW189x1xc8St8/T/jlb/fwSZnvBERJWF38CQ0OQtwbmTjWx8PoOGWPESMv/uWV32ERp6w8F7l2IvRuinUhUSkvJxWo0/WKfA8SZ/XdKFVrZdYRmCkpwKBkzSDCOGhCMKoYUi0tlKR5EWsBrTNLPcnqnwT/QibOCCSDI9UclaXR1MUQ/oF7F6TddkFm5WuORnjpkiYZlntQKAkvMRP4J3/pbEpYNfB1ETfBiJZ0kVsehnXBEPMvfQ0a0ju1iVG0+MCCR2UJZXnjAXow5JP2W7kxYdCuWP6jInMzEzLGiEFcxbpt2noVOJNtDF/vCa9lUQ60yuD6p127FFRTrCZxQchMVKhaeKCua9vt6EkjZuEYFummQMa0r87PajeejZlM8qMHJ1ikb956o+QKy1jTLW7Q9BFic1g9kWZQLSW58sGRtmZbu5QJGFqKdZPioc44dowyOgnw2E=) 2025-08-29 17:05:20.904300 | orchestrator | changed: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBEoyeWc8CpWlK6hRRk/tva9wYhWpBeUSGJl1HSxkkjuPsQX/YlZMWnfwSNMq17DoXInlpYIB14zwXgh7oc1pi+E=) 2025-08-29 17:05:20.904311 | orchestrator | changed: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAICN3yuPKx9eiPscoGb67XBh2jqcdiCsaMvMaXvYlcRk7) 2025-08-29 17:05:20.904322 | orchestrator | 2025-08-29 17:05:20.904333 | orchestrator | TASK [osism.commons.known_hosts : Write static known_hosts entries] ************ 2025-08-29 17:05:20.904343 | orchestrator | Friday 29 August 2025 17:05:19 +0000 (0:00:01.049) 0:00:25.014 ********* 2025-08-29 17:05:20.904355 | orchestrator | skipping: [testbed-manager] => (item=testbed-manager)  2025-08-29 17:05:20.904367 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-0)  2025-08-29 17:05:20.904378 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-1)  2025-08-29 17:05:20.904397 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-2)  2025-08-29 17:05:20.904411 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-3)  2025-08-29 17:05:20.904441 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-4)  2025-08-29 17:05:20.904455 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-5)  2025-08-29 17:05:20.904467 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:05:20.904480 | orchestrator | 2025-08-29 17:05:20.904492 | orchestrator | TASK [osism.commons.known_hosts : Write extra known_hosts entries] ************* 2025-08-29 17:05:20.904513 | orchestrator | Friday 29 August 2025 17:05:20 +0000 (0:00:00.159) 0:00:25.173 ********* 2025-08-29 17:05:20.904527 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:05:20.904540 | orchestrator | 2025-08-29 17:05:20.904552 | orchestrator | TASK [osism.commons.known_hosts : Delete known_hosts entries] ****************** 2025-08-29 17:05:20.904564 | orchestrator | Friday 29 August 2025 17:05:20 +0000 (0:00:00.054) 0:00:25.227 ********* 2025-08-29 17:05:20.904577 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:05:20.904589 | orchestrator | 2025-08-29 17:05:20.904602 | orchestrator | TASK [osism.commons.known_hosts : Set file permissions] ************************ 2025-08-29 17:05:20.904614 | orchestrator | Friday 29 August 2025 17:05:20 +0000 (0:00:00.047) 0:00:25.275 ********* 2025-08-29 17:05:20.904626 | orchestrator | changed: [testbed-manager] 2025-08-29 17:05:20.904639 | orchestrator | 2025-08-29 17:05:20.904651 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:05:20.904664 | orchestrator | testbed-manager : ok=31  changed=15  unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 17:05:20.904678 | orchestrator | 2025-08-29 17:05:20.904690 | orchestrator | 2025-08-29 17:05:20.904702 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:05:20.904715 | orchestrator | Friday 29 August 2025 17:05:20 +0000 (0:00:00.470) 0:00:25.745 ********* 2025-08-29 17:05:20.904728 | orchestrator | =============================================================================== 2025-08-29 17:05:20.904740 | orchestrator | osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 5.62s 2025-08-29 17:05:20.904751 | orchestrator | osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 4.99s 2025-08-29 17:05:20.904763 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.13s 2025-08-29 17:05:20.904773 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.06s 2025-08-29 17:05:20.904785 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.05s 2025-08-29 17:05:20.904796 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s 2025-08-29 17:05:20.904807 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.03s 2025-08-29 17:05:20.904818 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s 2025-08-29 17:05:20.904829 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s 2025-08-29 17:05:20.904840 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.00s 2025-08-29 17:05:20.904851 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.00s 2025-08-29 17:05:20.904862 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.99s 2025-08-29 17:05:20.904873 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.93s 2025-08-29 17:05:20.904884 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.92s 2025-08-29 17:05:20.904895 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.91s 2025-08-29 17:05:20.904906 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.90s 2025-08-29 17:05:20.904917 | orchestrator | osism.commons.known_hosts : Set file permissions ------------------------ 0.47s 2025-08-29 17:05:20.904927 | orchestrator | osism.commons.known_hosts : Write static known_hosts entries ------------ 0.16s 2025-08-29 17:05:20.904938 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.15s 2025-08-29 17:05:20.904950 | orchestrator | osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.15s 2025-08-29 17:05:21.149649 | orchestrator | + osism apply squid 2025-08-29 17:05:32.987211 | orchestrator | 2025-08-29 17:05:32 | INFO  | Task d28f06b7-28c7-4eb6-9728-8df97fd9d6b8 (squid) was prepared for execution. 2025-08-29 17:05:32.987319 | orchestrator | 2025-08-29 17:05:32 | INFO  | It takes a moment until task d28f06b7-28c7-4eb6-9728-8df97fd9d6b8 (squid) has been started and output is visible here. 2025-08-29 17:07:26.650136 | orchestrator | 2025-08-29 17:07:26.650277 | orchestrator | PLAY [Apply role squid] ******************************************************** 2025-08-29 17:07:26.650310 | orchestrator | 2025-08-29 17:07:26.650323 | orchestrator | TASK [osism.services.squid : Include install tasks] **************************** 2025-08-29 17:07:26.650345 | orchestrator | Friday 29 August 2025 17:05:36 +0000 (0:00:00.176) 0:00:00.176 ********* 2025-08-29 17:07:26.650357 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager 2025-08-29 17:07:26.650369 | orchestrator | 2025-08-29 17:07:26.650381 | orchestrator | TASK [osism.services.squid : Install required packages] ************************ 2025-08-29 17:07:26.650392 | orchestrator | Friday 29 August 2025 17:05:36 +0000 (0:00:00.089) 0:00:00.265 ********* 2025-08-29 17:07:26.650403 | orchestrator | ok: [testbed-manager] 2025-08-29 17:07:26.650415 | orchestrator | 2025-08-29 17:07:26.650427 | orchestrator | TASK [osism.services.squid : Create required directories] ********************** 2025-08-29 17:07:26.650438 | orchestrator | Friday 29 August 2025 17:05:38 +0000 (0:00:01.293) 0:00:01.559 ********* 2025-08-29 17:07:26.650449 | orchestrator | changed: [testbed-manager] => (item=/opt/squid/configuration) 2025-08-29 17:07:26.650460 | orchestrator | changed: [testbed-manager] => (item=/opt/squid/configuration/conf.d) 2025-08-29 17:07:26.650472 | orchestrator | ok: [testbed-manager] => (item=/opt/squid) 2025-08-29 17:07:26.650483 | orchestrator | 2025-08-29 17:07:26.650494 | orchestrator | TASK [osism.services.squid : Copy squid configuration files] ******************* 2025-08-29 17:07:26.650505 | orchestrator | Friday 29 August 2025 17:05:39 +0000 (0:00:01.069) 0:00:02.629 ********* 2025-08-29 17:07:26.650516 | orchestrator | changed: [testbed-manager] => (item=osism.conf) 2025-08-29 17:07:26.650527 | orchestrator | 2025-08-29 17:07:26.650538 | orchestrator | TASK [osism.services.squid : Remove osism_allow_list.conf configuration file] *** 2025-08-29 17:07:26.650549 | orchestrator | Friday 29 August 2025 17:05:40 +0000 (0:00:01.037) 0:00:03.667 ********* 2025-08-29 17:07:26.650560 | orchestrator | ok: [testbed-manager] 2025-08-29 17:07:26.650572 | orchestrator | 2025-08-29 17:07:26.650583 | orchestrator | TASK [osism.services.squid : Copy docker-compose.yml file] ********************* 2025-08-29 17:07:26.650616 | orchestrator | Friday 29 August 2025 17:05:40 +0000 (0:00:00.347) 0:00:04.014 ********* 2025-08-29 17:07:26.650629 | orchestrator | changed: [testbed-manager] 2025-08-29 17:07:26.650642 | orchestrator | 2025-08-29 17:07:26.650654 | orchestrator | TASK [osism.services.squid : Manage squid service] ***************************** 2025-08-29 17:07:26.650667 | orchestrator | Friday 29 August 2025 17:05:41 +0000 (0:00:00.895) 0:00:04.910 ********* 2025-08-29 17:07:26.650679 | orchestrator | FAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left). 2025-08-29 17:07:26.650692 | orchestrator | ok: [testbed-manager] 2025-08-29 17:07:26.650705 | orchestrator | 2025-08-29 17:07:26.650717 | orchestrator | RUNNING HANDLER [osism.services.squid : Restart squid service] ***************** 2025-08-29 17:07:26.650729 | orchestrator | Friday 29 August 2025 17:06:13 +0000 (0:00:32.340) 0:00:37.250 ********* 2025-08-29 17:07:26.650741 | orchestrator | changed: [testbed-manager] 2025-08-29 17:07:26.650753 | orchestrator | 2025-08-29 17:07:26.650766 | orchestrator | RUNNING HANDLER [osism.services.squid : Wait for squid service to start] ******* 2025-08-29 17:07:26.650779 | orchestrator | Friday 29 August 2025 17:06:25 +0000 (0:00:11.829) 0:00:49.080 ********* 2025-08-29 17:07:26.650791 | orchestrator | Pausing for 60 seconds 2025-08-29 17:07:26.650804 | orchestrator | changed: [testbed-manager] 2025-08-29 17:07:26.650817 | orchestrator | 2025-08-29 17:07:26.650829 | orchestrator | RUNNING HANDLER [osism.services.squid : Register that squid service was restarted] *** 2025-08-29 17:07:26.650841 | orchestrator | Friday 29 August 2025 17:07:25 +0000 (0:01:00.062) 0:01:49.143 ********* 2025-08-29 17:07:26.650854 | orchestrator | ok: [testbed-manager] 2025-08-29 17:07:26.650866 | orchestrator | 2025-08-29 17:07:26.650879 | orchestrator | RUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] ***** 2025-08-29 17:07:26.650914 | orchestrator | Friday 29 August 2025 17:07:25 +0000 (0:00:00.063) 0:01:49.206 ********* 2025-08-29 17:07:26.650927 | orchestrator | changed: [testbed-manager] 2025-08-29 17:07:26.650940 | orchestrator | 2025-08-29 17:07:26.650953 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:07:26.650982 | orchestrator | testbed-manager : ok=11  changed=6  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:07:26.650994 | orchestrator | 2025-08-29 17:07:26.651005 | orchestrator | 2025-08-29 17:07:26.651015 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:07:26.651026 | orchestrator | Friday 29 August 2025 17:07:26 +0000 (0:00:00.603) 0:01:49.810 ********* 2025-08-29 17:07:26.651037 | orchestrator | =============================================================================== 2025-08-29 17:07:26.651048 | orchestrator | osism.services.squid : Wait for squid service to start ----------------- 60.06s 2025-08-29 17:07:26.651059 | orchestrator | osism.services.squid : Manage squid service ---------------------------- 32.34s 2025-08-29 17:07:26.651069 | orchestrator | osism.services.squid : Restart squid service --------------------------- 11.83s 2025-08-29 17:07:26.651080 | orchestrator | osism.services.squid : Install required packages ------------------------ 1.29s 2025-08-29 17:07:26.651091 | orchestrator | osism.services.squid : Create required directories ---------------------- 1.07s 2025-08-29 17:07:26.651101 | orchestrator | osism.services.squid : Copy squid configuration files ------------------- 1.04s 2025-08-29 17:07:26.651112 | orchestrator | osism.services.squid : Copy docker-compose.yml file --------------------- 0.90s 2025-08-29 17:07:26.651123 | orchestrator | osism.services.squid : Wait for an healthy squid service ---------------- 0.60s 2025-08-29 17:07:26.651134 | orchestrator | osism.services.squid : Remove osism_allow_list.conf configuration file --- 0.35s 2025-08-29 17:07:26.651145 | orchestrator | osism.services.squid : Include install tasks ---------------------------- 0.09s 2025-08-29 17:07:26.651156 | orchestrator | osism.services.squid : Register that squid service was restarted -------- 0.06s 2025-08-29 17:07:26.884635 | orchestrator | + [[ 9.2.0 != \l\a\t\e\s\t ]] 2025-08-29 17:07:26.884717 | orchestrator | + sed -i 's#docker_namespace: kolla#docker_namespace: kolla/release#' /opt/configuration/inventory/group_vars/all/kolla.yml 2025-08-29 17:07:26.888155 | orchestrator | ++ semver 9.2.0 9.0.0 2025-08-29 17:07:26.948430 | orchestrator | + [[ 1 -lt 0 ]] 2025-08-29 17:07:26.949017 | orchestrator | + osism apply operator -u ubuntu -l testbed-nodes 2025-08-29 17:07:38.836813 | orchestrator | 2025-08-29 17:07:38 | INFO  | Task 0ab06e29-009d-446e-97e4-f1239251eba9 (operator) was prepared for execution. 2025-08-29 17:07:38.836910 | orchestrator | 2025-08-29 17:07:38 | INFO  | It takes a moment until task 0ab06e29-009d-446e-97e4-f1239251eba9 (operator) has been started and output is visible here. 2025-08-29 17:07:54.902356 | orchestrator | 2025-08-29 17:07:54.902472 | orchestrator | PLAY [Make ssh pipelining working] ********************************************* 2025-08-29 17:07:54.902491 | orchestrator | 2025-08-29 17:07:54.902504 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 17:07:54.902516 | orchestrator | Friday 29 August 2025 17:07:42 +0000 (0:00:00.151) 0:00:00.151 ********* 2025-08-29 17:07:54.902527 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:07:54.902539 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:07:54.902551 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:07:54.902562 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:07:54.902573 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:07:54.902584 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:07:54.902595 | orchestrator | 2025-08-29 17:07:54.902607 | orchestrator | TASK [Do not require tty for all users] **************************************** 2025-08-29 17:07:54.902618 | orchestrator | Friday 29 August 2025 17:07:46 +0000 (0:00:03.761) 0:00:03.912 ********* 2025-08-29 17:07:54.902629 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:07:54.902640 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:07:54.902652 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:07:54.902663 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:07:54.902697 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:07:54.902708 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:07:54.902719 | orchestrator | 2025-08-29 17:07:54.902730 | orchestrator | PLAY [Apply role operator] ***************************************************** 2025-08-29 17:07:54.902741 | orchestrator | 2025-08-29 17:07:54.902752 | orchestrator | TASK [osism.commons.operator : Gather variables for each operating system] ***** 2025-08-29 17:07:54.902763 | orchestrator | Friday 29 August 2025 17:07:47 +0000 (0:00:00.748) 0:00:04.660 ********* 2025-08-29 17:07:54.902774 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:07:54.902785 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:07:54.902796 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:07:54.902807 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:07:54.902817 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:07:54.902828 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:07:54.902839 | orchestrator | 2025-08-29 17:07:54.902850 | orchestrator | TASK [osism.commons.operator : Set operator_groups variable to default value] *** 2025-08-29 17:07:54.902861 | orchestrator | Friday 29 August 2025 17:07:47 +0000 (0:00:00.151) 0:00:04.812 ********* 2025-08-29 17:07:54.902872 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:07:54.902883 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:07:54.902895 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:07:54.902908 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:07:54.902920 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:07:54.902933 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:07:54.902945 | orchestrator | 2025-08-29 17:07:54.902980 | orchestrator | TASK [osism.commons.operator : Create operator group] ************************** 2025-08-29 17:07:54.902993 | orchestrator | Friday 29 August 2025 17:07:47 +0000 (0:00:00.137) 0:00:04.949 ********* 2025-08-29 17:07:54.903006 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:07:54.903019 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:07:54.903032 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:07:54.903044 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:07:54.903062 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:07:54.903082 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:07:54.903103 | orchestrator | 2025-08-29 17:07:54.903120 | orchestrator | TASK [osism.commons.operator : Create user] ************************************ 2025-08-29 17:07:54.903139 | orchestrator | Friday 29 August 2025 17:07:48 +0000 (0:00:00.588) 0:00:05.538 ********* 2025-08-29 17:07:54.903160 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:07:54.903182 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:07:54.903204 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:07:54.903221 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:07:54.903233 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:07:54.903245 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:07:54.903258 | orchestrator | 2025-08-29 17:07:54.903270 | orchestrator | TASK [osism.commons.operator : Add user to additional groups] ****************** 2025-08-29 17:07:54.903282 | orchestrator | Friday 29 August 2025 17:07:48 +0000 (0:00:00.764) 0:00:06.303 ********* 2025-08-29 17:07:54.903293 | orchestrator | changed: [testbed-node-0] => (item=adm) 2025-08-29 17:07:54.903303 | orchestrator | changed: [testbed-node-1] => (item=adm) 2025-08-29 17:07:54.903314 | orchestrator | changed: [testbed-node-2] => (item=adm) 2025-08-29 17:07:54.903325 | orchestrator | changed: [testbed-node-3] => (item=adm) 2025-08-29 17:07:54.903336 | orchestrator | changed: [testbed-node-4] => (item=adm) 2025-08-29 17:07:54.903347 | orchestrator | changed: [testbed-node-5] => (item=adm) 2025-08-29 17:07:54.903357 | orchestrator | changed: [testbed-node-0] => (item=sudo) 2025-08-29 17:07:54.903368 | orchestrator | changed: [testbed-node-3] => (item=sudo) 2025-08-29 17:07:54.903379 | orchestrator | changed: [testbed-node-2] => (item=sudo) 2025-08-29 17:07:54.903390 | orchestrator | changed: [testbed-node-4] => (item=sudo) 2025-08-29 17:07:54.903401 | orchestrator | changed: [testbed-node-5] => (item=sudo) 2025-08-29 17:07:54.903412 | orchestrator | changed: [testbed-node-1] => (item=sudo) 2025-08-29 17:07:54.903423 | orchestrator | 2025-08-29 17:07:54.903434 | orchestrator | TASK [osism.commons.operator : Copy user sudoers file] ************************* 2025-08-29 17:07:54.903455 | orchestrator | Friday 29 August 2025 17:07:49 +0000 (0:00:01.213) 0:00:07.516 ********* 2025-08-29 17:07:54.903466 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:07:54.903477 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:07:54.903492 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:07:54.903503 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:07:54.903514 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:07:54.903524 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:07:54.903535 | orchestrator | 2025-08-29 17:07:54.903546 | orchestrator | TASK [osism.commons.operator : Set language variables in .bashrc configuration file] *** 2025-08-29 17:07:54.903559 | orchestrator | Friday 29 August 2025 17:07:51 +0000 (0:00:01.382) 0:00:08.899 ********* 2025-08-29 17:07:54.903570 | orchestrator | [WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created 2025-08-29 17:07:54.903581 | orchestrator | with a mode of 0700, this may cause issues when running as another user. To 2025-08-29 17:07:54.903592 | orchestrator | avoid this, create the remote_tmp dir with the correct permissions manually 2025-08-29 17:07:54.903603 | orchestrator | changed: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8) 2025-08-29 17:07:54.903631 | orchestrator | changed: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8) 2025-08-29 17:07:54.903642 | orchestrator | changed: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8) 2025-08-29 17:07:54.903653 | orchestrator | changed: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8) 2025-08-29 17:07:54.903682 | orchestrator | changed: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8) 2025-08-29 17:07:54.903698 | orchestrator | changed: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8) 2025-08-29 17:07:54.903710 | orchestrator | changed: [testbed-node-1] => (item=export LANG=C.UTF-8) 2025-08-29 17:07:54.903720 | orchestrator | changed: [testbed-node-0] => (item=export LANG=C.UTF-8) 2025-08-29 17:07:54.903731 | orchestrator | changed: [testbed-node-5] => (item=export LANG=C.UTF-8) 2025-08-29 17:07:54.903742 | orchestrator | changed: [testbed-node-4] => (item=export LANG=C.UTF-8) 2025-08-29 17:07:54.903753 | orchestrator | changed: [testbed-node-2] => (item=export LANG=C.UTF-8) 2025-08-29 17:07:54.903763 | orchestrator | changed: [testbed-node-3] => (item=export LANG=C.UTF-8) 2025-08-29 17:07:54.903774 | orchestrator | changed: [testbed-node-5] => (item=export LC_ALL=C.UTF-8) 2025-08-29 17:07:54.903785 | orchestrator | changed: [testbed-node-4] => (item=export LC_ALL=C.UTF-8) 2025-08-29 17:07:54.903796 | orchestrator | changed: [testbed-node-3] => (item=export LC_ALL=C.UTF-8) 2025-08-29 17:07:54.903807 | orchestrator | changed: [testbed-node-2] => (item=export LC_ALL=C.UTF-8) 2025-08-29 17:07:54.903817 | orchestrator | changed: [testbed-node-1] => (item=export LC_ALL=C.UTF-8) 2025-08-29 17:07:54.903828 | orchestrator | changed: [testbed-node-0] => (item=export LC_ALL=C.UTF-8) 2025-08-29 17:07:54.903839 | orchestrator | 2025-08-29 17:07:54.903850 | orchestrator | TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] *** 2025-08-29 17:07:54.903862 | orchestrator | Friday 29 August 2025 17:07:52 +0000 (0:00:01.243) 0:00:10.143 ********* 2025-08-29 17:07:54.903873 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:07:54.903884 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:07:54.903895 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:07:54.903906 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:07:54.903916 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:07:54.903927 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:07:54.903938 | orchestrator | 2025-08-29 17:07:54.903980 | orchestrator | TASK [osism.commons.operator : Create .ssh directory] ************************** 2025-08-29 17:07:54.903992 | orchestrator | Friday 29 August 2025 17:07:52 +0000 (0:00:00.164) 0:00:10.307 ********* 2025-08-29 17:07:54.904003 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:07:54.904014 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:07:54.904025 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:07:54.904035 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:07:54.904053 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:07:54.904064 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:07:54.904075 | orchestrator | 2025-08-29 17:07:54.904086 | orchestrator | TASK [osism.commons.operator : Check number of SSH authorized keys] ************ 2025-08-29 17:07:54.904097 | orchestrator | Friday 29 August 2025 17:07:53 +0000 (0:00:00.610) 0:00:10.918 ********* 2025-08-29 17:07:54.904108 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:07:54.904118 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:07:54.904129 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:07:54.904140 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:07:54.904150 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:07:54.904161 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:07:54.904171 | orchestrator | 2025-08-29 17:07:54.904182 | orchestrator | TASK [osism.commons.operator : Set ssh authorized keys] ************************ 2025-08-29 17:07:54.904193 | orchestrator | Friday 29 August 2025 17:07:53 +0000 (0:00:00.194) 0:00:11.113 ********* 2025-08-29 17:07:54.904204 | orchestrator | changed: [testbed-node-3] => (item=None) 2025-08-29 17:07:54.904215 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:07:54.904225 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:07:54.904236 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:07:54.904247 | orchestrator | changed: [testbed-node-5] => (item=None) 2025-08-29 17:07:54.904257 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:07:54.904268 | orchestrator | changed: [testbed-node-1] => (item=None) 2025-08-29 17:07:54.904278 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:07:54.904289 | orchestrator | changed: [testbed-node-2] => (item=None) 2025-08-29 17:07:54.904300 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:07:54.904310 | orchestrator | changed: [testbed-node-4] => (item=None) 2025-08-29 17:07:54.904321 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:07:54.904332 | orchestrator | 2025-08-29 17:07:54.904343 | orchestrator | TASK [osism.commons.operator : Delete ssh authorized keys] ********************* 2025-08-29 17:07:54.904353 | orchestrator | Friday 29 August 2025 17:07:54 +0000 (0:00:00.765) 0:00:11.878 ********* 2025-08-29 17:07:54.904364 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:07:54.904375 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:07:54.904386 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:07:54.904396 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:07:54.904407 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:07:54.904417 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:07:54.904428 | orchestrator | 2025-08-29 17:07:54.904439 | orchestrator | TASK [osism.commons.operator : Set authorized GitHub accounts] ***************** 2025-08-29 17:07:54.904450 | orchestrator | Friday 29 August 2025 17:07:54 +0000 (0:00:00.184) 0:00:12.063 ********* 2025-08-29 17:07:54.904461 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:07:54.904472 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:07:54.904482 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:07:54.904493 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:07:54.904504 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:07:54.904514 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:07:54.904525 | orchestrator | 2025-08-29 17:07:54.904536 | orchestrator | TASK [osism.commons.operator : Delete authorized GitHub accounts] ************** 2025-08-29 17:07:54.904547 | orchestrator | Friday 29 August 2025 17:07:54 +0000 (0:00:00.169) 0:00:12.233 ********* 2025-08-29 17:07:54.904558 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:07:54.904569 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:07:54.904580 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:07:54.904591 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:07:54.904609 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:07:56.009073 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:07:56.009139 | orchestrator | 2025-08-29 17:07:56.009148 | orchestrator | TASK [osism.commons.operator : Set password] *********************************** 2025-08-29 17:07:56.009156 | orchestrator | Friday 29 August 2025 17:07:54 +0000 (0:00:00.187) 0:00:12.420 ********* 2025-08-29 17:07:56.009195 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:07:56.009201 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:07:56.009207 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:07:56.009212 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:07:56.009218 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:07:56.009224 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:07:56.009229 | orchestrator | 2025-08-29 17:07:56.009235 | orchestrator | TASK [osism.commons.operator : Unset & lock password] ************************** 2025-08-29 17:07:56.009241 | orchestrator | Friday 29 August 2025 17:07:55 +0000 (0:00:00.670) 0:00:13.091 ********* 2025-08-29 17:07:56.009247 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:07:56.009252 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:07:56.009258 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:07:56.009264 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:07:56.009269 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:07:56.009275 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:07:56.009281 | orchestrator | 2025-08-29 17:07:56.009286 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:07:56.009293 | orchestrator | testbed-node-0 : ok=12  changed=8  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:07:56.009300 | orchestrator | testbed-node-1 : ok=12  changed=8  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:07:56.009306 | orchestrator | testbed-node-2 : ok=12  changed=8  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:07:56.009312 | orchestrator | testbed-node-3 : ok=12  changed=8  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:07:56.009318 | orchestrator | testbed-node-4 : ok=12  changed=8  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:07:56.009324 | orchestrator | testbed-node-5 : ok=12  changed=8  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:07:56.009329 | orchestrator | 2025-08-29 17:07:56.009335 | orchestrator | 2025-08-29 17:07:56.009341 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:07:56.009347 | orchestrator | Friday 29 August 2025 17:07:55 +0000 (0:00:00.221) 0:00:13.313 ********* 2025-08-29 17:07:56.009352 | orchestrator | =============================================================================== 2025-08-29 17:07:56.009358 | orchestrator | Gathering Facts --------------------------------------------------------- 3.76s 2025-08-29 17:07:56.009364 | orchestrator | osism.commons.operator : Copy user sudoers file ------------------------- 1.38s 2025-08-29 17:07:56.009369 | orchestrator | osism.commons.operator : Set language variables in .bashrc configuration file --- 1.24s 2025-08-29 17:07:56.009376 | orchestrator | osism.commons.operator : Add user to additional groups ------------------ 1.21s 2025-08-29 17:07:56.009382 | orchestrator | osism.commons.operator : Set ssh authorized keys ------------------------ 0.77s 2025-08-29 17:07:56.009388 | orchestrator | osism.commons.operator : Create user ------------------------------------ 0.77s 2025-08-29 17:07:56.009393 | orchestrator | Do not require tty for all users ---------------------------------------- 0.75s 2025-08-29 17:07:56.009399 | orchestrator | osism.commons.operator : Set password ----------------------------------- 0.67s 2025-08-29 17:07:56.009404 | orchestrator | osism.commons.operator : Create .ssh directory -------------------------- 0.61s 2025-08-29 17:07:56.009410 | orchestrator | osism.commons.operator : Create operator group -------------------------- 0.59s 2025-08-29 17:07:56.009415 | orchestrator | osism.commons.operator : Unset & lock password -------------------------- 0.22s 2025-08-29 17:07:56.009421 | orchestrator | osism.commons.operator : Check number of SSH authorized keys ------------ 0.19s 2025-08-29 17:07:56.009431 | orchestrator | osism.commons.operator : Delete authorized GitHub accounts -------------- 0.19s 2025-08-29 17:07:56.009436 | orchestrator | osism.commons.operator : Delete ssh authorized keys --------------------- 0.19s 2025-08-29 17:07:56.009442 | orchestrator | osism.commons.operator : Set authorized GitHub accounts ----------------- 0.17s 2025-08-29 17:07:56.009447 | orchestrator | osism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.16s 2025-08-29 17:07:56.009453 | orchestrator | osism.commons.operator : Gather variables for each operating system ----- 0.15s 2025-08-29 17:07:56.009459 | orchestrator | osism.commons.operator : Set operator_groups variable to default value --- 0.14s 2025-08-29 17:07:56.275906 | orchestrator | + osism apply --environment custom facts 2025-08-29 17:07:57.994412 | orchestrator | 2025-08-29 17:07:57 | INFO  | Trying to run play facts in environment custom 2025-08-29 17:08:08.089804 | orchestrator | 2025-08-29 17:08:08 | INFO  | Task b959a503-d37c-44ae-8dab-62014cfc1e72 (facts) was prepared for execution. 2025-08-29 17:08:08.089909 | orchestrator | 2025-08-29 17:08:08 | INFO  | It takes a moment until task b959a503-d37c-44ae-8dab-62014cfc1e72 (facts) has been started and output is visible here. 2025-08-29 17:08:53.769932 | orchestrator | 2025-08-29 17:08:53.770188 | orchestrator | PLAY [Copy custom network devices fact] **************************************** 2025-08-29 17:08:53.770210 | orchestrator | 2025-08-29 17:08:53.770222 | orchestrator | TASK [Create custom facts directory] ******************************************* 2025-08-29 17:08:53.770234 | orchestrator | Friday 29 August 2025 17:08:11 +0000 (0:00:00.099) 0:00:00.099 ********* 2025-08-29 17:08:53.770245 | orchestrator | ok: [testbed-manager] 2025-08-29 17:08:53.770275 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:08:53.770287 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:08:53.770298 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:08:53.770309 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:08:53.770320 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:08:53.770331 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:08:53.770342 | orchestrator | 2025-08-29 17:08:53.770354 | orchestrator | TASK [Copy fact file] ********************************************************** 2025-08-29 17:08:53.770365 | orchestrator | Friday 29 August 2025 17:08:13 +0000 (0:00:01.444) 0:00:01.544 ********* 2025-08-29 17:08:53.770376 | orchestrator | ok: [testbed-manager] 2025-08-29 17:08:53.770387 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:08:53.770398 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:08:53.770409 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:08:53.770420 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:08:53.770430 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:08:53.770441 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:08:53.770452 | orchestrator | 2025-08-29 17:08:53.770463 | orchestrator | PLAY [Copy custom ceph devices facts] ****************************************** 2025-08-29 17:08:53.770475 | orchestrator | 2025-08-29 17:08:53.770488 | orchestrator | TASK [osism.commons.repository : Gather variables for each operating system] *** 2025-08-29 17:08:53.770501 | orchestrator | Friday 29 August 2025 17:08:14 +0000 (0:00:01.192) 0:00:02.737 ********* 2025-08-29 17:08:53.770514 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.770526 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.770538 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.770551 | orchestrator | 2025-08-29 17:08:53.770563 | orchestrator | TASK [osism.commons.repository : Set repository_default fact to default value] *** 2025-08-29 17:08:53.770576 | orchestrator | Friday 29 August 2025 17:08:14 +0000 (0:00:00.112) 0:00:02.849 ********* 2025-08-29 17:08:53.770589 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.770601 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.770613 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.770626 | orchestrator | 2025-08-29 17:08:53.770638 | orchestrator | TASK [osism.commons.repository : Set repositories to default] ****************** 2025-08-29 17:08:53.770651 | orchestrator | Friday 29 August 2025 17:08:14 +0000 (0:00:00.206) 0:00:03.056 ********* 2025-08-29 17:08:53.770663 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.770696 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.770710 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.770722 | orchestrator | 2025-08-29 17:08:53.770734 | orchestrator | TASK [osism.commons.repository : Include distribution specific repository tasks] *** 2025-08-29 17:08:53.770747 | orchestrator | Friday 29 August 2025 17:08:14 +0000 (0:00:00.200) 0:00:03.256 ********* 2025-08-29 17:08:53.770760 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:08:53.770775 | orchestrator | 2025-08-29 17:08:53.770787 | orchestrator | TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] ***** 2025-08-29 17:08:53.770800 | orchestrator | Friday 29 August 2025 17:08:15 +0000 (0:00:00.137) 0:00:03.393 ********* 2025-08-29 17:08:53.770812 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.770825 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.770836 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.770847 | orchestrator | 2025-08-29 17:08:53.770858 | orchestrator | TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] ************* 2025-08-29 17:08:53.770869 | orchestrator | Friday 29 August 2025 17:08:15 +0000 (0:00:00.602) 0:00:03.996 ********* 2025-08-29 17:08:53.770880 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:08:53.770891 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:08:53.770902 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:08:53.770912 | orchestrator | 2025-08-29 17:08:53.770923 | orchestrator | TASK [osism.commons.repository : Copy 99osism apt configuration] *************** 2025-08-29 17:08:53.770934 | orchestrator | Friday 29 August 2025 17:08:15 +0000 (0:00:00.099) 0:00:04.096 ********* 2025-08-29 17:08:53.770945 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:08:53.770976 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:08:53.770988 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:08:53.770999 | orchestrator | 2025-08-29 17:08:53.771010 | orchestrator | TASK [osism.commons.repository : Remove sources.list file] ********************* 2025-08-29 17:08:53.771021 | orchestrator | Friday 29 August 2025 17:08:16 +0000 (0:00:01.097) 0:00:05.194 ********* 2025-08-29 17:08:53.771032 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.771043 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.771054 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.771065 | orchestrator | 2025-08-29 17:08:53.771076 | orchestrator | TASK [osism.commons.repository : Copy ubuntu.sources file] ********************* 2025-08-29 17:08:53.771087 | orchestrator | Friday 29 August 2025 17:08:17 +0000 (0:00:00.475) 0:00:05.669 ********* 2025-08-29 17:08:53.771098 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:08:53.771108 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:08:53.771119 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:08:53.771130 | orchestrator | 2025-08-29 17:08:53.771141 | orchestrator | TASK [osism.commons.repository : Update package cache] ************************* 2025-08-29 17:08:53.771152 | orchestrator | Friday 29 August 2025 17:08:18 +0000 (0:00:01.077) 0:00:06.747 ********* 2025-08-29 17:08:53.771163 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:08:53.771173 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:08:53.771184 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:08:53.771195 | orchestrator | 2025-08-29 17:08:53.771206 | orchestrator | TASK [Install required packages (RedHat)] ************************************** 2025-08-29 17:08:53.771217 | orchestrator | Friday 29 August 2025 17:08:36 +0000 (0:00:17.844) 0:00:24.591 ********* 2025-08-29 17:08:53.771227 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:08:53.771238 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:08:53.771249 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:08:53.771260 | orchestrator | 2025-08-29 17:08:53.771270 | orchestrator | TASK [Install required packages (Debian)] ************************************** 2025-08-29 17:08:53.771300 | orchestrator | Friday 29 August 2025 17:08:36 +0000 (0:00:00.093) 0:00:24.685 ********* 2025-08-29 17:08:53.771312 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:08:53.771322 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:08:53.771341 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:08:53.771352 | orchestrator | 2025-08-29 17:08:53.771363 | orchestrator | TASK [Create custom facts directory] ******************************************* 2025-08-29 17:08:53.771379 | orchestrator | Friday 29 August 2025 17:08:44 +0000 (0:00:08.026) 0:00:32.711 ********* 2025-08-29 17:08:53.771390 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.771401 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.771412 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.771423 | orchestrator | 2025-08-29 17:08:53.771433 | orchestrator | TASK [Copy fact files] ********************************************************* 2025-08-29 17:08:53.771445 | orchestrator | Friday 29 August 2025 17:08:44 +0000 (0:00:00.441) 0:00:33.153 ********* 2025-08-29 17:08:53.771455 | orchestrator | changed: [testbed-node-3] => (item=testbed_ceph_devices) 2025-08-29 17:08:53.771466 | orchestrator | changed: [testbed-node-4] => (item=testbed_ceph_devices) 2025-08-29 17:08:53.771477 | orchestrator | changed: [testbed-node-5] => (item=testbed_ceph_devices) 2025-08-29 17:08:53.771488 | orchestrator | changed: [testbed-node-3] => (item=testbed_ceph_devices_all) 2025-08-29 17:08:53.771499 | orchestrator | changed: [testbed-node-5] => (item=testbed_ceph_devices_all) 2025-08-29 17:08:53.771510 | orchestrator | changed: [testbed-node-4] => (item=testbed_ceph_devices_all) 2025-08-29 17:08:53.771521 | orchestrator | changed: [testbed-node-3] => (item=testbed_ceph_osd_devices) 2025-08-29 17:08:53.771531 | orchestrator | changed: [testbed-node-5] => (item=testbed_ceph_osd_devices) 2025-08-29 17:08:53.771542 | orchestrator | changed: [testbed-node-4] => (item=testbed_ceph_osd_devices) 2025-08-29 17:08:53.771553 | orchestrator | changed: [testbed-node-3] => (item=testbed_ceph_osd_devices_all) 2025-08-29 17:08:53.771564 | orchestrator | changed: [testbed-node-5] => (item=testbed_ceph_osd_devices_all) 2025-08-29 17:08:53.771574 | orchestrator | changed: [testbed-node-4] => (item=testbed_ceph_osd_devices_all) 2025-08-29 17:08:53.771585 | orchestrator | 2025-08-29 17:08:53.771596 | orchestrator | RUNNING HANDLER [osism.commons.repository : Force update of package cache] ***** 2025-08-29 17:08:53.771607 | orchestrator | Friday 29 August 2025 17:08:48 +0000 (0:00:03.693) 0:00:36.846 ********* 2025-08-29 17:08:53.771618 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.771629 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.771640 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.771651 | orchestrator | 2025-08-29 17:08:53.771661 | orchestrator | PLAY [Gather facts for all hosts] ********************************************** 2025-08-29 17:08:53.771672 | orchestrator | 2025-08-29 17:08:53.771683 | orchestrator | TASK [Gathers facts about hosts] *********************************************** 2025-08-29 17:08:53.771694 | orchestrator | Friday 29 August 2025 17:08:49 +0000 (0:00:01.262) 0:00:38.109 ********* 2025-08-29 17:08:53.771705 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:08:53.771716 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:08:53.771727 | orchestrator | ok: [testbed-manager] 2025-08-29 17:08:53.771738 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:08:53.771749 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:08:53.771759 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:08:53.771770 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:08:53.771781 | orchestrator | 2025-08-29 17:08:53.771792 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:08:53.771803 | orchestrator | testbed-manager : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:08:53.771815 | orchestrator | testbed-node-0 : ok=3  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:08:53.771826 | orchestrator | testbed-node-1 : ok=3  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:08:53.771837 | orchestrator | testbed-node-2 : ok=3  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:08:53.771849 | orchestrator | testbed-node-3 : ok=16  changed=7  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:08:53.771866 | orchestrator | testbed-node-4 : ok=16  changed=7  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:08:53.771877 | orchestrator | testbed-node-5 : ok=16  changed=7  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:08:53.771888 | orchestrator | 2025-08-29 17:08:53.771899 | orchestrator | 2025-08-29 17:08:53.771910 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:08:53.771921 | orchestrator | Friday 29 August 2025 17:08:53 +0000 (0:00:03.916) 0:00:42.026 ********* 2025-08-29 17:08:53.771932 | orchestrator | =============================================================================== 2025-08-29 17:08:53.771943 | orchestrator | osism.commons.repository : Update package cache ------------------------ 17.84s 2025-08-29 17:08:53.771968 | orchestrator | Install required packages (Debian) -------------------------------------- 8.03s 2025-08-29 17:08:53.771979 | orchestrator | Gathers facts about hosts ----------------------------------------------- 3.92s 2025-08-29 17:08:53.771990 | orchestrator | Copy fact files --------------------------------------------------------- 3.69s 2025-08-29 17:08:53.772001 | orchestrator | Create custom facts directory ------------------------------------------- 1.44s 2025-08-29 17:08:53.772012 | orchestrator | osism.commons.repository : Force update of package cache ---------------- 1.26s 2025-08-29 17:08:53.772029 | orchestrator | Copy fact file ---------------------------------------------------------- 1.19s 2025-08-29 17:08:54.028989 | orchestrator | osism.commons.repository : Copy 99osism apt configuration --------------- 1.10s 2025-08-29 17:08:54.029072 | orchestrator | osism.commons.repository : Copy ubuntu.sources file --------------------- 1.08s 2025-08-29 17:08:54.029084 | orchestrator | osism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.60s 2025-08-29 17:08:54.029094 | orchestrator | osism.commons.repository : Remove sources.list file --------------------- 0.48s 2025-08-29 17:08:54.029103 | orchestrator | Create custom facts directory ------------------------------------------- 0.44s 2025-08-29 17:08:54.029113 | orchestrator | osism.commons.repository : Set repository_default fact to default value --- 0.21s 2025-08-29 17:08:54.029123 | orchestrator | osism.commons.repository : Set repositories to default ------------------ 0.20s 2025-08-29 17:08:54.029133 | orchestrator | osism.commons.repository : Include distribution specific repository tasks --- 0.14s 2025-08-29 17:08:54.029143 | orchestrator | osism.commons.repository : Gather variables for each operating system --- 0.11s 2025-08-29 17:08:54.029153 | orchestrator | osism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.10s 2025-08-29 17:08:54.029162 | orchestrator | Install required packages (RedHat) -------------------------------------- 0.09s 2025-08-29 17:08:54.282244 | orchestrator | + osism apply bootstrap 2025-08-29 17:09:06.117497 | orchestrator | 2025-08-29 17:09:06 | INFO  | Task 9a3e8d0c-0102-4e3c-84b5-b7c6fa34963c (bootstrap) was prepared for execution. 2025-08-29 17:09:06.117602 | orchestrator | 2025-08-29 17:09:06 | INFO  | It takes a moment until task 9a3e8d0c-0102-4e3c-84b5-b7c6fa34963c (bootstrap) has been started and output is visible here. 2025-08-29 17:09:20.957121 | orchestrator | 2025-08-29 17:09:20.957228 | orchestrator | PLAY [Group hosts based on state bootstrap] ************************************ 2025-08-29 17:09:20.957245 | orchestrator | 2025-08-29 17:09:20.957273 | orchestrator | TASK [Group hosts based on state bootstrap] ************************************ 2025-08-29 17:09:20.957284 | orchestrator | Friday 29 August 2025 17:09:09 +0000 (0:00:00.120) 0:00:00.120 ********* 2025-08-29 17:09:20.957294 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:20.957304 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:20.957314 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:20.957323 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:20.957333 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:20.957342 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:20.957352 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:20.957381 | orchestrator | 2025-08-29 17:09:20.957392 | orchestrator | PLAY [Gather facts for all hosts] ********************************************** 2025-08-29 17:09:20.957402 | orchestrator | 2025-08-29 17:09:20.957411 | orchestrator | TASK [Gathers facts about hosts] *********************************************** 2025-08-29 17:09:20.957421 | orchestrator | Friday 29 August 2025 17:09:10 +0000 (0:00:00.180) 0:00:00.301 ********* 2025-08-29 17:09:20.957431 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:20.957441 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:20.957450 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:20.957460 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:20.957469 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:20.957478 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:20.957488 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:20.957497 | orchestrator | 2025-08-29 17:09:20.957507 | orchestrator | PLAY [Gather facts for all hosts (if using --limit)] *************************** 2025-08-29 17:09:20.957516 | orchestrator | 2025-08-29 17:09:20.957526 | orchestrator | TASK [Gathers facts about hosts] *********************************************** 2025-08-29 17:09:20.957536 | orchestrator | Friday 29 August 2025 17:09:13 +0000 (0:00:03.590) 0:00:03.891 ********* 2025-08-29 17:09:20.957546 | orchestrator | skipping: [testbed-manager] => (item=testbed-manager)  2025-08-29 17:09:20.957556 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-0)  2025-08-29 17:09:20.957565 | orchestrator | skipping: [testbed-node-0] => (item=testbed-manager)  2025-08-29 17:09:20.957575 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-1)  2025-08-29 17:09:20.957585 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-0)  2025-08-29 17:09:20.957594 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-2)  2025-08-29 17:09:20.957604 | orchestrator | skipping: [testbed-node-1] => (item=testbed-manager)  2025-08-29 17:09:20.957613 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-3)  2025-08-29 17:09:20.957623 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-1)  2025-08-29 17:09:20.957633 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-0)  2025-08-29 17:09:20.957642 | orchestrator | skipping: [testbed-node-2] => (item=testbed-manager)  2025-08-29 17:09:20.957652 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-4)  2025-08-29 17:09:20.957661 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-1)  2025-08-29 17:09:20.957672 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-2)  2025-08-29 17:09:20.957684 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-0)  2025-08-29 17:09:20.957694 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-5)  2025-08-29 17:09:20.957705 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:20.957715 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-2)  2025-08-29 17:09:20.957726 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-3)  2025-08-29 17:09:20.957737 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-3)  2025-08-29 17:09:20.957748 | orchestrator | skipping: [testbed-node-3] => (item=testbed-manager)  2025-08-29 17:09:20.957759 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-1)  2025-08-29 17:09:20.957770 | orchestrator | skipping: [testbed-node-4] => (item=testbed-manager)  2025-08-29 17:09:20.957781 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-4)  2025-08-29 17:09:20.957792 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-0)  2025-08-29 17:09:20.957802 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-4)  2025-08-29 17:09:20.957813 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-0)  2025-08-29 17:09:20.957824 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-5)  2025-08-29 17:09:20.957834 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-2)  2025-08-29 17:09:20.957845 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:09:20.957860 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-5)  2025-08-29 17:09:20.957878 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:09:20.957889 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-1)  2025-08-29 17:09:20.957900 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-1)  2025-08-29 17:09:20.957911 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-3)  2025-08-29 17:09:20.957921 | orchestrator | skipping: [testbed-node-5] => (item=testbed-manager)  2025-08-29 17:09:20.957931 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-2)  2025-08-29 17:09:20.957941 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-2)  2025-08-29 17:09:20.957950 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-4)  2025-08-29 17:09:20.957980 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-0)  2025-08-29 17:09:20.957990 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-3)  2025-08-29 17:09:20.958000 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:09:20.958009 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-5)  2025-08-29 17:09:20.958058 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:09:20.958070 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-1)  2025-08-29 17:09:20.958080 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-4)  2025-08-29 17:09:20.958105 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:09:20.958115 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-2)  2025-08-29 17:09:20.958125 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-5)  2025-08-29 17:09:20.958134 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:09:20.958144 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-3)  2025-08-29 17:09:20.958153 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:09:20.958162 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:09:20.958172 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-4)  2025-08-29 17:09:20.958181 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-5)  2025-08-29 17:09:20.958191 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:09:20.958200 | orchestrator | 2025-08-29 17:09:20.958209 | orchestrator | PLAY [Apply bootstrap roles part 1] ******************************************** 2025-08-29 17:09:20.958219 | orchestrator | 2025-08-29 17:09:20.958229 | orchestrator | TASK [osism.commons.hostname : Set hostname] *********************************** 2025-08-29 17:09:20.958238 | orchestrator | Friday 29 August 2025 17:09:13 +0000 (0:00:00.365) 0:00:04.257 ********* 2025-08-29 17:09:20.958248 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:20.958257 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:20.958267 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:20.958276 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:20.958286 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:20.958295 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:20.958305 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:20.958314 | orchestrator | 2025-08-29 17:09:20.958324 | orchestrator | TASK [osism.commons.hostname : Copy /etc/hostname] ***************************** 2025-08-29 17:09:20.958334 | orchestrator | Friday 29 August 2025 17:09:15 +0000 (0:00:01.173) 0:00:05.431 ********* 2025-08-29 17:09:20.958344 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:20.958353 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:20.958362 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:20.958372 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:20.958381 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:20.958390 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:20.958400 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:20.958409 | orchestrator | 2025-08-29 17:09:20.958418 | orchestrator | TASK [osism.commons.hosts : Include type specific tasks] *********************** 2025-08-29 17:09:20.958428 | orchestrator | Friday 29 August 2025 17:09:16 +0000 (0:00:01.132) 0:00:06.563 ********* 2025-08-29 17:09:20.958439 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:09:20.958458 | orchestrator | 2025-08-29 17:09:20.958468 | orchestrator | TASK [osism.commons.hosts : Copy /etc/hosts file] ****************************** 2025-08-29 17:09:20.958477 | orchestrator | Friday 29 August 2025 17:09:16 +0000 (0:00:00.223) 0:00:06.787 ********* 2025-08-29 17:09:20.958487 | orchestrator | changed: [testbed-manager] 2025-08-29 17:09:20.958497 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:09:20.958506 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:09:20.958516 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:09:20.958525 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:09:20.958535 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:09:20.958544 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:09:20.958553 | orchestrator | 2025-08-29 17:09:20.958563 | orchestrator | TASK [osism.commons.proxy : Include distribution specific tasks] *************** 2025-08-29 17:09:20.958572 | orchestrator | Friday 29 August 2025 17:09:18 +0000 (0:00:02.050) 0:00:08.837 ********* 2025-08-29 17:09:20.958582 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:20.958593 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:09:20.958605 | orchestrator | 2025-08-29 17:09:20.958614 | orchestrator | TASK [osism.commons.proxy : Configure proxy parameters for apt] **************** 2025-08-29 17:09:20.958624 | orchestrator | Friday 29 August 2025 17:09:18 +0000 (0:00:00.221) 0:00:09.059 ********* 2025-08-29 17:09:20.958634 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:09:20.958643 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:09:20.958653 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:09:20.958662 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:09:20.958671 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:09:20.958681 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:09:20.958690 | orchestrator | 2025-08-29 17:09:20.958699 | orchestrator | TASK [osism.commons.proxy : Set system wide settings in environment file] ****** 2025-08-29 17:09:20.958709 | orchestrator | Friday 29 August 2025 17:09:19 +0000 (0:00:01.038) 0:00:10.098 ********* 2025-08-29 17:09:20.958719 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:20.958728 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:09:20.958737 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:09:20.958747 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:09:20.958756 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:09:20.958765 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:09:20.958775 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:09:20.958784 | orchestrator | 2025-08-29 17:09:20.958794 | orchestrator | TASK [osism.commons.proxy : Remove system wide settings in environment file] *** 2025-08-29 17:09:20.958803 | orchestrator | Friday 29 August 2025 17:09:20 +0000 (0:00:00.590) 0:00:10.688 ********* 2025-08-29 17:09:20.958813 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:09:20.958822 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:09:20.958832 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:09:20.958841 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:09:20.958851 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:09:20.958860 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:09:20.958870 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:20.958879 | orchestrator | 2025-08-29 17:09:20.958888 | orchestrator | TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] *** 2025-08-29 17:09:20.958899 | orchestrator | Friday 29 August 2025 17:09:20 +0000 (0:00:00.406) 0:00:11.094 ********* 2025-08-29 17:09:20.958908 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:20.958918 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:09:20.958933 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:09:32.941539 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:09:32.941643 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:09:32.941657 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:09:32.941690 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:09:32.941702 | orchestrator | 2025-08-29 17:09:32.941715 | orchestrator | TASK [osism.commons.resolvconf : Include resolvconf tasks] ********************* 2025-08-29 17:09:32.941727 | orchestrator | Friday 29 August 2025 17:09:21 +0000 (0:00:00.182) 0:00:11.276 ********* 2025-08-29 17:09:32.941740 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:09:32.941767 | orchestrator | 2025-08-29 17:09:32.941778 | orchestrator | TASK [osism.commons.resolvconf : Include distribution specific installation tasks] *** 2025-08-29 17:09:32.941790 | orchestrator | Friday 29 August 2025 17:09:21 +0000 (0:00:00.269) 0:00:11.545 ********* 2025-08-29 17:09:32.941802 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:09:32.941813 | orchestrator | 2025-08-29 17:09:32.941824 | orchestrator | TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] *** 2025-08-29 17:09:32.941835 | orchestrator | Friday 29 August 2025 17:09:21 +0000 (0:00:00.294) 0:00:11.840 ********* 2025-08-29 17:09:32.941862 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.941874 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.941885 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.941896 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.941907 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.941918 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.941929 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.941940 | orchestrator | 2025-08-29 17:09:32.941951 | orchestrator | TASK [osism.commons.resolvconf : Install package systemd-resolved] ************* 2025-08-29 17:09:32.941982 | orchestrator | Friday 29 August 2025 17:09:22 +0000 (0:00:01.278) 0:00:13.119 ********* 2025-08-29 17:09:32.941994 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:32.942004 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:09:32.942064 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:09:32.942078 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:09:32.942090 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:09:32.942100 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:09:32.942111 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:09:32.942122 | orchestrator | 2025-08-29 17:09:32.942133 | orchestrator | TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] ***** 2025-08-29 17:09:32.942144 | orchestrator | Friday 29 August 2025 17:09:23 +0000 (0:00:00.207) 0:00:13.326 ********* 2025-08-29 17:09:32.942155 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.942165 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.942176 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.942187 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.942198 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.942208 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.942219 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.942230 | orchestrator | 2025-08-29 17:09:32.942241 | orchestrator | TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] ******* 2025-08-29 17:09:32.942251 | orchestrator | Friday 29 August 2025 17:09:23 +0000 (0:00:00.539) 0:00:13.865 ********* 2025-08-29 17:09:32.942301 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:32.942313 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:09:32.942324 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:09:32.942335 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:09:32.942346 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:09:32.942356 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:09:32.942367 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:09:32.942378 | orchestrator | 2025-08-29 17:09:32.942389 | orchestrator | TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] *** 2025-08-29 17:09:32.942418 | orchestrator | Friday 29 August 2025 17:09:23 +0000 (0:00:00.201) 0:00:14.067 ********* 2025-08-29 17:09:32.942429 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.942440 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:09:32.942451 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:09:32.942462 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:09:32.942472 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:09:32.942483 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:09:32.942498 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:09:32.942509 | orchestrator | 2025-08-29 17:09:32.942521 | orchestrator | TASK [osism.commons.resolvconf : Copy configuration files] ********************* 2025-08-29 17:09:32.942532 | orchestrator | Friday 29 August 2025 17:09:24 +0000 (0:00:00.538) 0:00:14.605 ********* 2025-08-29 17:09:32.942542 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.942553 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:09:32.942564 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:09:32.942575 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:09:32.942586 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:09:32.942597 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:09:32.942608 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:09:32.942619 | orchestrator | 2025-08-29 17:09:32.942629 | orchestrator | TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ******** 2025-08-29 17:09:32.942641 | orchestrator | Friday 29 August 2025 17:09:25 +0000 (0:00:01.104) 0:00:15.709 ********* 2025-08-29 17:09:32.942652 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.942663 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.942674 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.942684 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.942695 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.942706 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.942717 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.942727 | orchestrator | 2025-08-29 17:09:32.942739 | orchestrator | TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] *** 2025-08-29 17:09:32.942750 | orchestrator | Friday 29 August 2025 17:09:26 +0000 (0:00:01.211) 0:00:16.921 ********* 2025-08-29 17:09:32.942779 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:09:32.942790 | orchestrator | 2025-08-29 17:09:32.942801 | orchestrator | TASK [osism.commons.resolvconf : Restart systemd-resolved service] ************* 2025-08-29 17:09:32.942812 | orchestrator | Friday 29 August 2025 17:09:26 +0000 (0:00:00.312) 0:00:17.233 ********* 2025-08-29 17:09:32.942823 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:32.942834 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:09:32.942845 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:09:32.942856 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:09:32.942867 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:09:32.942877 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:09:32.942888 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:09:32.942899 | orchestrator | 2025-08-29 17:09:32.942910 | orchestrator | TASK [osism.commons.repository : Gather variables for each operating system] *** 2025-08-29 17:09:32.942921 | orchestrator | Friday 29 August 2025 17:09:28 +0000 (0:00:01.495) 0:00:18.729 ********* 2025-08-29 17:09:32.942932 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.942943 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.942953 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.942982 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.942994 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.943004 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.943015 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.943026 | orchestrator | 2025-08-29 17:09:32.943037 | orchestrator | TASK [osism.commons.repository : Set repository_default fact to default value] *** 2025-08-29 17:09:32.943048 | orchestrator | Friday 29 August 2025 17:09:28 +0000 (0:00:00.215) 0:00:18.944 ********* 2025-08-29 17:09:32.943068 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.943079 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.943090 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.943101 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.943111 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.943122 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.943132 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.943143 | orchestrator | 2025-08-29 17:09:32.943154 | orchestrator | TASK [osism.commons.repository : Set repositories to default] ****************** 2025-08-29 17:09:32.943165 | orchestrator | Friday 29 August 2025 17:09:28 +0000 (0:00:00.278) 0:00:19.222 ********* 2025-08-29 17:09:32.943176 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.943187 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.943198 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.943208 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.943219 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.943230 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.943240 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.943251 | orchestrator | 2025-08-29 17:09:32.943262 | orchestrator | TASK [osism.commons.repository : Include distribution specific repository tasks] *** 2025-08-29 17:09:32.943273 | orchestrator | Friday 29 August 2025 17:09:29 +0000 (0:00:00.246) 0:00:19.469 ********* 2025-08-29 17:09:32.943285 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:09:32.943297 | orchestrator | 2025-08-29 17:09:32.943308 | orchestrator | TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] ***** 2025-08-29 17:09:32.943319 | orchestrator | Friday 29 August 2025 17:09:29 +0000 (0:00:00.285) 0:00:19.755 ********* 2025-08-29 17:09:32.943330 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.943341 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.943351 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.943362 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.943373 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.943384 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.943394 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.943405 | orchestrator | 2025-08-29 17:09:32.943416 | orchestrator | TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] ************* 2025-08-29 17:09:32.943427 | orchestrator | Friday 29 August 2025 17:09:30 +0000 (0:00:00.541) 0:00:20.297 ********* 2025-08-29 17:09:32.943437 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:09:32.943448 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:09:32.943459 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:09:32.943470 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:09:32.943481 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:09:32.943492 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:09:32.943502 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:09:32.943513 | orchestrator | 2025-08-29 17:09:32.943530 | orchestrator | TASK [osism.commons.repository : Copy 99osism apt configuration] *************** 2025-08-29 17:09:32.943541 | orchestrator | Friday 29 August 2025 17:09:30 +0000 (0:00:00.233) 0:00:20.530 ********* 2025-08-29 17:09:32.943552 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.943563 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.943574 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:09:32.943585 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:09:32.943596 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.943606 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:09:32.943617 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.943628 | orchestrator | 2025-08-29 17:09:32.943639 | orchestrator | TASK [osism.commons.repository : Remove sources.list file] ********************* 2025-08-29 17:09:32.943650 | orchestrator | Friday 29 August 2025 17:09:31 +0000 (0:00:01.005) 0:00:21.536 ********* 2025-08-29 17:09:32.943660 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.943678 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:09:32.943688 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:09:32.943699 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:09:32.943710 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.943721 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.943731 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.943742 | orchestrator | 2025-08-29 17:09:32.943753 | orchestrator | TASK [osism.commons.repository : Copy ubuntu.sources file] ********************* 2025-08-29 17:09:32.943764 | orchestrator | Friday 29 August 2025 17:09:31 +0000 (0:00:00.577) 0:00:22.113 ********* 2025-08-29 17:09:32.943775 | orchestrator | ok: [testbed-manager] 2025-08-29 17:09:32.943786 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:09:32.943796 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:09:32.943807 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:09:32.943825 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:10:12.769603 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:10:12.769714 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:10:12.769730 | orchestrator | 2025-08-29 17:10:12.769743 | orchestrator | TASK [osism.commons.repository : Update package cache] ************************* 2025-08-29 17:10:12.769756 | orchestrator | Friday 29 August 2025 17:09:32 +0000 (0:00:01.075) 0:00:23.189 ********* 2025-08-29 17:10:12.769768 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.769779 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.769790 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.769801 | orchestrator | changed: [testbed-manager] 2025-08-29 17:10:12.769812 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:10:12.769824 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:10:12.769836 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:10:12.769846 | orchestrator | 2025-08-29 17:10:12.769857 | orchestrator | TASK [osism.services.rsyslog : Gather variables for each operating system] ***** 2025-08-29 17:10:12.769868 | orchestrator | Friday 29 August 2025 17:09:50 +0000 (0:00:17.801) 0:00:40.990 ********* 2025-08-29 17:10:12.769879 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.769890 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.769901 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.769912 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.769922 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.769933 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.769944 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.769955 | orchestrator | 2025-08-29 17:10:12.769966 | orchestrator | TASK [osism.services.rsyslog : Set rsyslog_user variable to default value] ***** 2025-08-29 17:10:12.770006 | orchestrator | Friday 29 August 2025 17:09:50 +0000 (0:00:00.217) 0:00:41.207 ********* 2025-08-29 17:10:12.770060 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.770075 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.770086 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.770097 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.770107 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.770118 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.770129 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.770141 | orchestrator | 2025-08-29 17:10:12.770154 | orchestrator | TASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] *** 2025-08-29 17:10:12.770166 | orchestrator | Friday 29 August 2025 17:09:51 +0000 (0:00:00.205) 0:00:41.413 ********* 2025-08-29 17:10:12.770178 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.770191 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.770202 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.770215 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.770227 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.770238 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.770248 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.770259 | orchestrator | 2025-08-29 17:10:12.770270 | orchestrator | TASK [osism.services.rsyslog : Include distribution specific install tasks] **** 2025-08-29 17:10:12.770281 | orchestrator | Friday 29 August 2025 17:09:51 +0000 (0:00:00.183) 0:00:41.597 ********* 2025-08-29 17:10:12.770294 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:10:12.770332 | orchestrator | 2025-08-29 17:10:12.770343 | orchestrator | TASK [osism.services.rsyslog : Install rsyslog package] ************************ 2025-08-29 17:10:12.770354 | orchestrator | Friday 29 August 2025 17:09:51 +0000 (0:00:00.256) 0:00:41.853 ********* 2025-08-29 17:10:12.770365 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.770375 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.770386 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.770397 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.770407 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.770418 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.770428 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.770439 | orchestrator | 2025-08-29 17:10:12.770450 | orchestrator | TASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] *********** 2025-08-29 17:10:12.770461 | orchestrator | Friday 29 August 2025 17:09:53 +0000 (0:00:01.775) 0:00:43.629 ********* 2025-08-29 17:10:12.770472 | orchestrator | changed: [testbed-manager] 2025-08-29 17:10:12.770483 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:10:12.770494 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:10:12.770505 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:10:12.770515 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:10:12.770526 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:10:12.770537 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:10:12.770548 | orchestrator | 2025-08-29 17:10:12.770559 | orchestrator | TASK [osism.services.rsyslog : Manage rsyslog service] ************************* 2025-08-29 17:10:12.770570 | orchestrator | Friday 29 August 2025 17:09:54 +0000 (0:00:01.159) 0:00:44.788 ********* 2025-08-29 17:10:12.770580 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.770591 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.770602 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.770613 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.770624 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.770634 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.770645 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.770656 | orchestrator | 2025-08-29 17:10:12.770667 | orchestrator | TASK [osism.services.rsyslog : Include fluentd tasks] ************************** 2025-08-29 17:10:12.770677 | orchestrator | Friday 29 August 2025 17:09:55 +0000 (0:00:00.831) 0:00:45.619 ********* 2025-08-29 17:10:12.770689 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:10:12.770701 | orchestrator | 2025-08-29 17:10:12.770712 | orchestrator | TASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] *** 2025-08-29 17:10:12.770724 | orchestrator | Friday 29 August 2025 17:09:55 +0000 (0:00:00.284) 0:00:45.904 ********* 2025-08-29 17:10:12.770735 | orchestrator | changed: [testbed-manager] 2025-08-29 17:10:12.770745 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:10:12.770756 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:10:12.770767 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:10:12.770778 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:10:12.770788 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:10:12.770799 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:10:12.770810 | orchestrator | 2025-08-29 17:10:12.770840 | orchestrator | TASK [osism.services.rsyslog : Include additional log server tasks] ************ 2025-08-29 17:10:12.770851 | orchestrator | Friday 29 August 2025 17:09:56 +0000 (0:00:01.072) 0:00:46.976 ********* 2025-08-29 17:10:12.770862 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:10:12.770873 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:10:12.770884 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:10:12.770895 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:10:12.770906 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:10:12.770925 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:10:12.770935 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:10:12.770946 | orchestrator | 2025-08-29 17:10:12.770957 | orchestrator | TASK [osism.commons.systohc : Install util-linux-extra package] **************** 2025-08-29 17:10:12.770968 | orchestrator | Friday 29 August 2025 17:09:57 +0000 (0:00:00.291) 0:00:47.268 ********* 2025-08-29 17:10:12.771048 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:10:12.771062 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:10:12.771072 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:10:12.771083 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:10:12.771094 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:10:12.771104 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:10:12.771115 | orchestrator | changed: [testbed-manager] 2025-08-29 17:10:12.771125 | orchestrator | 2025-08-29 17:10:12.771136 | orchestrator | TASK [osism.commons.systohc : Sync hardware clock] ***************************** 2025-08-29 17:10:12.771147 | orchestrator | Friday 29 August 2025 17:10:07 +0000 (0:00:10.839) 0:00:58.107 ********* 2025-08-29 17:10:12.771158 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.771168 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.771179 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.771189 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.771200 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.771211 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.771221 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.771232 | orchestrator | 2025-08-29 17:10:12.771241 | orchestrator | TASK [osism.commons.configfs : Start sys-kernel-config mount] ****************** 2025-08-29 17:10:12.771251 | orchestrator | Friday 29 August 2025 17:10:08 +0000 (0:00:00.804) 0:00:58.912 ********* 2025-08-29 17:10:12.771260 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.771270 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.771279 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.771288 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.771298 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.771307 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.771316 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.771326 | orchestrator | 2025-08-29 17:10:12.771335 | orchestrator | TASK [osism.commons.packages : Gather variables for each operating system] ***** 2025-08-29 17:10:12.771345 | orchestrator | Friday 29 August 2025 17:10:09 +0000 (0:00:00.878) 0:00:59.790 ********* 2025-08-29 17:10:12.771354 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.771380 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.771390 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.771399 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.771409 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.771418 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.771428 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.771437 | orchestrator | 2025-08-29 17:10:12.771447 | orchestrator | TASK [osism.commons.packages : Set required_packages_distribution variable to default value] *** 2025-08-29 17:10:12.771457 | orchestrator | Friday 29 August 2025 17:10:09 +0000 (0:00:00.204) 0:00:59.995 ********* 2025-08-29 17:10:12.771466 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.771476 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.771485 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.771495 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.771504 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.771513 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.771523 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.771532 | orchestrator | 2025-08-29 17:10:12.771542 | orchestrator | TASK [osism.commons.packages : Include distribution specific package tasks] **** 2025-08-29 17:10:12.771551 | orchestrator | Friday 29 August 2025 17:10:09 +0000 (0:00:00.218) 0:01:00.213 ********* 2025-08-29 17:10:12.771561 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:10:12.771580 | orchestrator | 2025-08-29 17:10:12.771590 | orchestrator | TASK [osism.commons.packages : Install needrestart package] ******************** 2025-08-29 17:10:12.771599 | orchestrator | Friday 29 August 2025 17:10:10 +0000 (0:00:00.286) 0:01:00.500 ********* 2025-08-29 17:10:12.771609 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.771623 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.771633 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.771642 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.771651 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.771661 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.771670 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.771680 | orchestrator | 2025-08-29 17:10:12.771690 | orchestrator | TASK [osism.commons.packages : Set needrestart mode] *************************** 2025-08-29 17:10:12.771699 | orchestrator | Friday 29 August 2025 17:10:12 +0000 (0:00:01.764) 0:01:02.264 ********* 2025-08-29 17:10:12.771709 | orchestrator | changed: [testbed-manager] 2025-08-29 17:10:12.771718 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:10:12.771728 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:10:12.771737 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:10:12.771747 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:10:12.771756 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:10:12.771766 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:10:12.771775 | orchestrator | 2025-08-29 17:10:12.771785 | orchestrator | TASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] *** 2025-08-29 17:10:12.771794 | orchestrator | Friday 29 August 2025 17:10:12 +0000 (0:00:00.554) 0:01:02.819 ********* 2025-08-29 17:10:12.771804 | orchestrator | ok: [testbed-manager] 2025-08-29 17:10:12.771814 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:10:12.771823 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:10:12.771833 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:10:12.771842 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:10:12.771852 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:10:12.771861 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:10:12.771871 | orchestrator | 2025-08-29 17:10:12.771887 | orchestrator | TASK [osism.commons.packages : Update package cache] *************************** 2025-08-29 17:12:33.449685 | orchestrator | Friday 29 August 2025 17:10:12 +0000 (0:00:00.199) 0:01:03.018 ********* 2025-08-29 17:12:33.449792 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:33.449809 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:33.449821 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:33.449832 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:33.449843 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:33.449854 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:33.449865 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:33.449876 | orchestrator | 2025-08-29 17:12:33.449888 | orchestrator | TASK [osism.commons.packages : Download upgrade packages] ********************** 2025-08-29 17:12:33.449900 | orchestrator | Friday 29 August 2025 17:10:13 +0000 (0:00:01.175) 0:01:04.194 ********* 2025-08-29 17:12:33.449911 | orchestrator | changed: [testbed-manager] 2025-08-29 17:12:33.449923 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:12:33.449933 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:12:33.449944 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:12:33.449955 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:12:33.449966 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:12:33.449977 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:12:33.449988 | orchestrator | 2025-08-29 17:12:33.449999 | orchestrator | TASK [osism.commons.packages : Upgrade packages] ******************************* 2025-08-29 17:12:33.450010 | orchestrator | Friday 29 August 2025 17:10:15 +0000 (0:00:01.849) 0:01:06.043 ********* 2025-08-29 17:12:33.450076 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:33.450113 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:33.450126 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:33.450136 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:33.450147 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:33.450158 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:33.450259 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:33.450276 | orchestrator | 2025-08-29 17:12:33.450289 | orchestrator | TASK [osism.commons.packages : Download required packages] ********************* 2025-08-29 17:12:33.450301 | orchestrator | Friday 29 August 2025 17:10:18 +0000 (0:00:02.522) 0:01:08.566 ********* 2025-08-29 17:12:33.450313 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:33.450325 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:33.450337 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:33.450349 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:33.450361 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:33.450373 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:33.450385 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:33.450398 | orchestrator | 2025-08-29 17:12:33.450410 | orchestrator | TASK [osism.commons.packages : Install required packages] ********************** 2025-08-29 17:12:33.450422 | orchestrator | Friday 29 August 2025 17:10:56 +0000 (0:00:38.325) 0:01:46.891 ********* 2025-08-29 17:12:33.450434 | orchestrator | changed: [testbed-manager] 2025-08-29 17:12:33.450446 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:12:33.450457 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:12:33.450470 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:12:33.450482 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:12:33.450494 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:12:33.450506 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:12:33.450519 | orchestrator | 2025-08-29 17:12:33.450530 | orchestrator | TASK [osism.commons.packages : Remove useless packages from the cache] ********* 2025-08-29 17:12:33.450543 | orchestrator | Friday 29 August 2025 17:12:14 +0000 (0:01:17.446) 0:03:04.337 ********* 2025-08-29 17:12:33.450556 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:33.450567 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:33.450580 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:33.450591 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:33.450602 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:33.450612 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:33.450623 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:33.450634 | orchestrator | 2025-08-29 17:12:33.450645 | orchestrator | TASK [osism.commons.packages : Remove dependencies that are no longer required] *** 2025-08-29 17:12:33.450657 | orchestrator | Friday 29 August 2025 17:12:15 +0000 (0:00:01.784) 0:03:06.122 ********* 2025-08-29 17:12:33.450668 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:33.450679 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:33.450689 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:33.450700 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:33.450711 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:33.450722 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:33.450732 | orchestrator | changed: [testbed-manager] 2025-08-29 17:12:33.450743 | orchestrator | 2025-08-29 17:12:33.450754 | orchestrator | TASK [osism.commons.sysctl : Include sysctl tasks] ***************************** 2025-08-29 17:12:33.450765 | orchestrator | Friday 29 August 2025 17:12:26 +0000 (0:00:10.856) 0:03:16.979 ********* 2025-08-29 17:12:33.450800 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]}) 2025-08-29 17:12:33.450817 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]}) 2025-08-29 17:12:33.450865 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]}) 2025-08-29 17:12:33.450879 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]}) 2025-08-29 17:12:33.450891 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]}) 2025-08-29 17:12:33.450903 | orchestrator | 2025-08-29 17:12:33.450914 | orchestrator | TASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] *********** 2025-08-29 17:12:33.450925 | orchestrator | Friday 29 August 2025 17:12:27 +0000 (0:00:00.382) 0:03:17.361 ********* 2025-08-29 17:12:33.450936 | orchestrator | skipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144})  2025-08-29 17:12:33.450947 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:12:33.450958 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144})  2025-08-29 17:12:33.450970 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144})  2025-08-29 17:12:33.450981 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:12:33.450992 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:12:33.451003 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144})  2025-08-29 17:12:33.451014 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:12:33.451025 | orchestrator | changed: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144}) 2025-08-29 17:12:33.451036 | orchestrator | changed: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144}) 2025-08-29 17:12:33.451047 | orchestrator | changed: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144}) 2025-08-29 17:12:33.451058 | orchestrator | 2025-08-29 17:12:33.451070 | orchestrator | TASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] **************** 2025-08-29 17:12:33.451152 | orchestrator | Friday 29 August 2025 17:12:27 +0000 (0:00:00.665) 0:03:18.026 ********* 2025-08-29 17:12:33.451167 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})  2025-08-29 17:12:33.451180 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})  2025-08-29 17:12:33.451191 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})  2025-08-29 17:12:33.451202 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216})  2025-08-29 17:12:33.451213 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216})  2025-08-29 17:12:33.451224 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})  2025-08-29 17:12:33.451235 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})  2025-08-29 17:12:33.451246 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096})  2025-08-29 17:12:33.451257 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})  2025-08-29 17:12:33.451268 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})  2025-08-29 17:12:33.451288 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:12:33.451299 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})  2025-08-29 17:12:33.451310 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})  2025-08-29 17:12:33.451320 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})  2025-08-29 17:12:33.451331 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216})  2025-08-29 17:12:33.451342 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})  2025-08-29 17:12:33.451353 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216})  2025-08-29 17:12:33.451365 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})  2025-08-29 17:12:33.451375 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})  2025-08-29 17:12:33.451386 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})  2025-08-29 17:12:33.451397 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})  2025-08-29 17:12:33.451416 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216})  2025-08-29 17:12:35.817314 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096})  2025-08-29 17:12:35.817417 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216})  2025-08-29 17:12:35.817431 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})  2025-08-29 17:12:35.817442 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})  2025-08-29 17:12:35.817455 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})  2025-08-29 17:12:35.817466 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})  2025-08-29 17:12:35.817478 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:12:35.817489 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096})  2025-08-29 17:12:35.817500 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})  2025-08-29 17:12:35.817511 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})  2025-08-29 17:12:35.817522 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:12:35.817533 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})  2025-08-29 17:12:35.817544 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})  2025-08-29 17:12:35.817555 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})  2025-08-29 17:12:35.817566 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216})  2025-08-29 17:12:35.817577 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216})  2025-08-29 17:12:35.817587 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})  2025-08-29 17:12:35.817598 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})  2025-08-29 17:12:35.817609 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096})  2025-08-29 17:12:35.817620 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})  2025-08-29 17:12:35.817631 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})  2025-08-29 17:12:35.817665 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:12:35.817677 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) 2025-08-29 17:12:35.817688 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) 2025-08-29 17:12:35.817698 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) 2025-08-29 17:12:35.817709 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) 2025-08-29 17:12:35.817720 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) 2025-08-29 17:12:35.817731 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) 2025-08-29 17:12:35.817741 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) 2025-08-29 17:12:35.817770 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) 2025-08-29 17:12:35.817782 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) 2025-08-29 17:12:35.817797 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) 2025-08-29 17:12:35.817808 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) 2025-08-29 17:12:35.817819 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) 2025-08-29 17:12:35.817830 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) 2025-08-29 17:12:35.817840 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) 2025-08-29 17:12:35.817852 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) 2025-08-29 17:12:35.817864 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) 2025-08-29 17:12:35.817877 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) 2025-08-29 17:12:35.817888 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) 2025-08-29 17:12:35.817900 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) 2025-08-29 17:12:35.817913 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096}) 2025-08-29 17:12:35.817925 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) 2025-08-29 17:12:35.817954 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096}) 2025-08-29 17:12:35.817967 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) 2025-08-29 17:12:35.817981 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) 2025-08-29 17:12:35.817993 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) 2025-08-29 17:12:35.818005 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) 2025-08-29 17:12:35.818077 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) 2025-08-29 17:12:35.818139 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096}) 2025-08-29 17:12:35.818150 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) 2025-08-29 17:12:35.818171 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) 2025-08-29 17:12:35.818182 | orchestrator | 2025-08-29 17:12:35.818193 | orchestrator | TASK [osism.commons.sysctl : Set sysctl parameters on generic] ***************** 2025-08-29 17:12:35.818205 | orchestrator | Friday 29 August 2025 17:12:33 +0000 (0:00:05.672) 0:03:23.698 ********* 2025-08-29 17:12:35.818225 | orchestrator | changed: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1}) 2025-08-29 17:12:35.818236 | orchestrator | changed: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1}) 2025-08-29 17:12:35.818247 | orchestrator | changed: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1}) 2025-08-29 17:12:35.818258 | orchestrator | changed: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1}) 2025-08-29 17:12:35.818269 | orchestrator | changed: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1}) 2025-08-29 17:12:35.818280 | orchestrator | changed: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1}) 2025-08-29 17:12:35.818290 | orchestrator | changed: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1}) 2025-08-29 17:12:35.818301 | orchestrator | 2025-08-29 17:12:35.818312 | orchestrator | TASK [osism.commons.sysctl : Set sysctl parameters on compute] ***************** 2025-08-29 17:12:35.818323 | orchestrator | Friday 29 August 2025 17:12:34 +0000 (0:00:00.833) 0:03:24.532 ********* 2025-08-29 17:12:35.818339 | orchestrator | skipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})  2025-08-29 17:12:35.818351 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})  2025-08-29 17:12:35.818361 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:12:35.818372 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:12:35.818383 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})  2025-08-29 17:12:35.818394 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})  2025-08-29 17:12:35.818405 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:12:35.818416 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:12:35.818426 | orchestrator | changed: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) 2025-08-29 17:12:35.818438 | orchestrator | changed: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) 2025-08-29 17:12:35.818448 | orchestrator | changed: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) 2025-08-29 17:12:35.818459 | orchestrator | 2025-08-29 17:12:35.818470 | orchestrator | TASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] **************** 2025-08-29 17:12:35.818481 | orchestrator | Friday 29 August 2025 17:12:34 +0000 (0:00:00.600) 0:03:25.133 ********* 2025-08-29 17:12:35.818491 | orchestrator | skipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})  2025-08-29 17:12:35.818508 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:12:35.818519 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})  2025-08-29 17:12:35.818530 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})  2025-08-29 17:12:35.818541 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:12:35.818552 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:12:35.818563 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})  2025-08-29 17:12:35.818574 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:12:35.818585 | orchestrator | changed: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) 2025-08-29 17:12:35.818596 | orchestrator | changed: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) 2025-08-29 17:12:35.818607 | orchestrator | changed: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) 2025-08-29 17:12:35.818617 | orchestrator | 2025-08-29 17:12:35.818628 | orchestrator | TASK [osism.commons.limits : Include limits tasks] ***************************** 2025-08-29 17:12:35.818639 | orchestrator | Friday 29 August 2025 17:12:35 +0000 (0:00:00.675) 0:03:25.808 ********* 2025-08-29 17:12:35.818649 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:12:35.818660 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:12:35.818678 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:12:35.818689 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:12:35.818699 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:12:35.818718 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:12:47.356001 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:12:47.356136 | orchestrator | 2025-08-29 17:12:47.356153 | orchestrator | TASK [osism.commons.services : Populate service facts] ************************* 2025-08-29 17:12:47.356166 | orchestrator | Friday 29 August 2025 17:12:35 +0000 (0:00:00.263) 0:03:26.072 ********* 2025-08-29 17:12:47.356178 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:47.356189 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:47.356200 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:47.356211 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:47.356222 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:47.356233 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:47.356244 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:47.356255 | orchestrator | 2025-08-29 17:12:47.356266 | orchestrator | TASK [osism.commons.services : Check services] ********************************* 2025-08-29 17:12:47.356277 | orchestrator | Friday 29 August 2025 17:12:41 +0000 (0:00:05.621) 0:03:31.694 ********* 2025-08-29 17:12:47.356288 | orchestrator | skipping: [testbed-manager] => (item=nscd)  2025-08-29 17:12:47.356299 | orchestrator | skipping: [testbed-node-0] => (item=nscd)  2025-08-29 17:12:47.356311 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:12:47.356322 | orchestrator | skipping: [testbed-node-1] => (item=nscd)  2025-08-29 17:12:47.356332 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:12:47.356344 | orchestrator | skipping: [testbed-node-2] => (item=nscd)  2025-08-29 17:12:47.356354 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:12:47.356365 | orchestrator | skipping: [testbed-node-3] => (item=nscd)  2025-08-29 17:12:47.356376 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:12:47.356387 | orchestrator | skipping: [testbed-node-4] => (item=nscd)  2025-08-29 17:12:47.356397 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:12:47.356408 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:12:47.356419 | orchestrator | skipping: [testbed-node-5] => (item=nscd)  2025-08-29 17:12:47.356430 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:12:47.356440 | orchestrator | 2025-08-29 17:12:47.356451 | orchestrator | TASK [osism.commons.services : Start/enable required services] ***************** 2025-08-29 17:12:47.356463 | orchestrator | Friday 29 August 2025 17:12:41 +0000 (0:00:00.292) 0:03:31.986 ********* 2025-08-29 17:12:47.356474 | orchestrator | ok: [testbed-manager] => (item=cron) 2025-08-29 17:12:47.356485 | orchestrator | ok: [testbed-node-0] => (item=cron) 2025-08-29 17:12:47.356495 | orchestrator | ok: [testbed-node-1] => (item=cron) 2025-08-29 17:12:47.356506 | orchestrator | ok: [testbed-node-2] => (item=cron) 2025-08-29 17:12:47.356517 | orchestrator | ok: [testbed-node-3] => (item=cron) 2025-08-29 17:12:47.356528 | orchestrator | ok: [testbed-node-5] => (item=cron) 2025-08-29 17:12:47.356539 | orchestrator | ok: [testbed-node-4] => (item=cron) 2025-08-29 17:12:47.356549 | orchestrator | 2025-08-29 17:12:47.356561 | orchestrator | TASK [osism.commons.motd : Include distribution specific configure tasks] ****** 2025-08-29 17:12:47.356574 | orchestrator | Friday 29 August 2025 17:12:42 +0000 (0:00:00.999) 0:03:32.986 ********* 2025-08-29 17:12:47.356588 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:12:47.356603 | orchestrator | 2025-08-29 17:12:47.356615 | orchestrator | TASK [osism.commons.motd : Remove update-motd package] ************************* 2025-08-29 17:12:47.356628 | orchestrator | Friday 29 August 2025 17:12:43 +0000 (0:00:00.396) 0:03:33.382 ********* 2025-08-29 17:12:47.356640 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:47.356652 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:47.356664 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:47.356698 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:47.356711 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:47.356723 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:47.356734 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:47.356747 | orchestrator | 2025-08-29 17:12:47.356759 | orchestrator | TASK [osism.commons.motd : Check if /etc/default/motd-news exists] ************* 2025-08-29 17:12:47.356771 | orchestrator | Friday 29 August 2025 17:12:44 +0000 (0:00:01.546) 0:03:34.929 ********* 2025-08-29 17:12:47.356783 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:47.356794 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:47.356806 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:47.356818 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:47.356830 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:47.356841 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:47.356853 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:47.356865 | orchestrator | 2025-08-29 17:12:47.356891 | orchestrator | TASK [osism.commons.motd : Disable the dynamic motd-news service] ************** 2025-08-29 17:12:47.356904 | orchestrator | Friday 29 August 2025 17:12:45 +0000 (0:00:00.587) 0:03:35.516 ********* 2025-08-29 17:12:47.356915 | orchestrator | changed: [testbed-manager] 2025-08-29 17:12:47.356926 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:12:47.356937 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:12:47.356947 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:12:47.356958 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:12:47.356969 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:12:47.356979 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:12:47.356990 | orchestrator | 2025-08-29 17:12:47.357001 | orchestrator | TASK [osism.commons.motd : Get all configuration files in /etc/pam.d] ********** 2025-08-29 17:12:47.357011 | orchestrator | Friday 29 August 2025 17:12:45 +0000 (0:00:00.597) 0:03:36.114 ********* 2025-08-29 17:12:47.357022 | orchestrator | ok: [testbed-manager] 2025-08-29 17:12:47.357033 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:12:47.357044 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:12:47.357054 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:12:47.357065 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:12:47.357075 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:12:47.357086 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:12:47.357113 | orchestrator | 2025-08-29 17:12:47.357125 | orchestrator | TASK [osism.commons.motd : Remove pam_motd.so rule] **************************** 2025-08-29 17:12:47.357136 | orchestrator | Friday 29 August 2025 17:12:46 +0000 (0:00:00.574) 0:03:36.689 ********* 2025-08-29 17:12:47.357167 | orchestrator | changed: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1756486208.7933588, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:12:47.357183 | orchestrator | changed: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1756486231.6307616, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:12:47.357195 | orchestrator | changed: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1756486235.621267, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:12:47.357215 | orchestrator | changed: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1756486245.5538208, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:12:47.357227 | orchestrator | changed: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1756486240.152679, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:12:47.357239 | orchestrator | changed: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1756486239.3295631, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:12:47.357250 | orchestrator | changed: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1756486245.1763003, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:12:47.357278 | orchestrator | changed: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:13:12.551491 | orchestrator | changed: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:13:12.551626 | orchestrator | changed: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:13:12.551667 | orchestrator | changed: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:13:12.551680 | orchestrator | changed: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:13:12.551697 | orchestrator | changed: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:13:12.551709 | orchestrator | changed: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:13:12.551721 | orchestrator | 2025-08-29 17:13:12.551735 | orchestrator | TASK [osism.commons.motd : Copy motd file] ************************************* 2025-08-29 17:13:12.551749 | orchestrator | Friday 29 August 2025 17:12:47 +0000 (0:00:00.915) 0:03:37.604 ********* 2025-08-29 17:13:12.551760 | orchestrator | changed: [testbed-manager] 2025-08-29 17:13:12.551772 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:13:12.551783 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:13:12.551793 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:13:12.551804 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:13:12.551815 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:13:12.551826 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:13:12.551837 | orchestrator | 2025-08-29 17:13:12.551848 | orchestrator | TASK [osism.commons.motd : Copy issue file] ************************************ 2025-08-29 17:13:12.551859 | orchestrator | Friday 29 August 2025 17:12:48 +0000 (0:00:01.069) 0:03:38.674 ********* 2025-08-29 17:13:12.551870 | orchestrator | changed: [testbed-manager] 2025-08-29 17:13:12.551881 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:13:12.551892 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:13:12.551903 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:13:12.551931 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:13:12.551942 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:13:12.551953 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:13:12.551964 | orchestrator | 2025-08-29 17:13:12.551975 | orchestrator | TASK [osism.commons.motd : Copy issue.net file] ******************************** 2025-08-29 17:13:12.551994 | orchestrator | Friday 29 August 2025 17:12:49 +0000 (0:00:01.095) 0:03:39.769 ********* 2025-08-29 17:13:12.552005 | orchestrator | changed: [testbed-manager] 2025-08-29 17:13:12.552017 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:13:12.552029 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:13:12.552041 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:13:12.552054 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:13:12.552065 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:13:12.552078 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:13:12.552090 | orchestrator | 2025-08-29 17:13:12.552102 | orchestrator | TASK [osism.commons.motd : Configure SSH to print the motd] ******************** 2025-08-29 17:13:12.552138 | orchestrator | Friday 29 August 2025 17:12:50 +0000 (0:00:01.175) 0:03:40.944 ********* 2025-08-29 17:13:12.552150 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:13:12.552162 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:13:12.552174 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:13:12.552186 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:13:12.552198 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:13:12.552210 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:13:12.552221 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:13:12.552233 | orchestrator | 2025-08-29 17:13:12.552245 | orchestrator | TASK [osism.commons.motd : Configure SSH to not print the motd] **************** 2025-08-29 17:13:12.552258 | orchestrator | Friday 29 August 2025 17:12:50 +0000 (0:00:00.267) 0:03:41.212 ********* 2025-08-29 17:13:12.552269 | orchestrator | ok: [testbed-manager] 2025-08-29 17:13:12.552282 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:13:12.552294 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:13:12.552306 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:13:12.552317 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:13:12.552330 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:13:12.552342 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:13:12.552354 | orchestrator | 2025-08-29 17:13:12.552366 | orchestrator | TASK [osism.services.rng : Include distribution specific install tasks] ******** 2025-08-29 17:13:12.552377 | orchestrator | Friday 29 August 2025 17:12:51 +0000 (0:00:00.690) 0:03:41.903 ********* 2025-08-29 17:13:12.552391 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:13:12.552405 | orchestrator | 2025-08-29 17:13:12.552416 | orchestrator | TASK [osism.services.rng : Install rng package] ******************************** 2025-08-29 17:13:12.552427 | orchestrator | Friday 29 August 2025 17:12:52 +0000 (0:00:00.382) 0:03:42.286 ********* 2025-08-29 17:13:12.552437 | orchestrator | ok: [testbed-manager] 2025-08-29 17:13:12.552448 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:13:12.552459 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:13:12.552470 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:13:12.552480 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:13:12.552491 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:13:12.552502 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:13:12.552513 | orchestrator | 2025-08-29 17:13:12.552523 | orchestrator | TASK [osism.services.rng : Remove haveged package] ***************************** 2025-08-29 17:13:12.552534 | orchestrator | Friday 29 August 2025 17:13:00 +0000 (0:00:08.251) 0:03:50.537 ********* 2025-08-29 17:13:12.552545 | orchestrator | ok: [testbed-manager] 2025-08-29 17:13:12.552556 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:13:12.552567 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:13:12.552578 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:13:12.552588 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:13:12.552599 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:13:12.552610 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:13:12.552620 | orchestrator | 2025-08-29 17:13:12.552631 | orchestrator | TASK [osism.services.rng : Manage rng service] ********************************* 2025-08-29 17:13:12.552649 | orchestrator | Friday 29 August 2025 17:13:01 +0000 (0:00:01.270) 0:03:51.808 ********* 2025-08-29 17:13:12.552660 | orchestrator | ok: [testbed-manager] 2025-08-29 17:13:12.552676 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:13:12.552687 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:13:12.552698 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:13:12.552708 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:13:12.552719 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:13:12.552730 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:13:12.552741 | orchestrator | 2025-08-29 17:13:12.552751 | orchestrator | TASK [osism.services.smartd : Include distribution specific install tasks] ***** 2025-08-29 17:13:12.552762 | orchestrator | Friday 29 August 2025 17:13:02 +0000 (0:00:00.974) 0:03:52.783 ********* 2025-08-29 17:13:12.552774 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:13:12.552785 | orchestrator | 2025-08-29 17:13:12.552796 | orchestrator | TASK [osism.services.smartd : Install smartmontools package] ******************* 2025-08-29 17:13:12.552807 | orchestrator | Friday 29 August 2025 17:13:02 +0000 (0:00:00.363) 0:03:53.146 ********* 2025-08-29 17:13:12.552818 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:13:12.552829 | orchestrator | changed: [testbed-manager] 2025-08-29 17:13:12.552839 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:13:12.552865 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:13:12.552887 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:13:12.552898 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:13:12.552909 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:13:12.552919 | orchestrator | 2025-08-29 17:13:12.552930 | orchestrator | TASK [osism.services.smartd : Create /var/log/smartd directory] **************** 2025-08-29 17:13:12.552941 | orchestrator | Friday 29 August 2025 17:13:11 +0000 (0:00:09.027) 0:04:02.174 ********* 2025-08-29 17:13:12.552952 | orchestrator | changed: [testbed-manager] 2025-08-29 17:13:12.552962 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:13:12.552973 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:13:12.552991 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:22.668336 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:22.668428 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:22.668442 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:22.668453 | orchestrator | 2025-08-29 17:14:22.668464 | orchestrator | TASK [osism.services.smartd : Copy smartmontools configuration file] *********** 2025-08-29 17:14:22.668475 | orchestrator | Friday 29 August 2025 17:13:12 +0000 (0:00:00.625) 0:04:02.800 ********* 2025-08-29 17:14:22.668485 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:22.668495 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:22.668505 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:22.668515 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:22.668524 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:22.668534 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:22.668544 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:22.668553 | orchestrator | 2025-08-29 17:14:22.668564 | orchestrator | TASK [osism.services.smartd : Manage smartd service] *************************** 2025-08-29 17:14:22.668573 | orchestrator | Friday 29 August 2025 17:13:14 +0000 (0:00:01.774) 0:04:04.574 ********* 2025-08-29 17:14:22.668583 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:22.668593 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:22.668603 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:22.668612 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:22.668622 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:22.668631 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:22.668641 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:22.668650 | orchestrator | 2025-08-29 17:14:22.668660 | orchestrator | TASK [osism.commons.cleanup : Gather variables for each operating system] ****** 2025-08-29 17:14:22.668670 | orchestrator | Friday 29 August 2025 17:13:15 +0000 (0:00:00.938) 0:04:05.513 ********* 2025-08-29 17:14:22.668702 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:22.668713 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:22.668723 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:22.668732 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:22.668742 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:22.668751 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:22.668760 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:22.668770 | orchestrator | 2025-08-29 17:14:22.668780 | orchestrator | TASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] *** 2025-08-29 17:14:22.668790 | orchestrator | Friday 29 August 2025 17:13:15 +0000 (0:00:00.220) 0:04:05.734 ********* 2025-08-29 17:14:22.668799 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:22.668809 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:22.668818 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:22.668827 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:22.668836 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:22.668846 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:22.668855 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:22.668865 | orchestrator | 2025-08-29 17:14:22.668874 | orchestrator | TASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] *** 2025-08-29 17:14:22.668885 | orchestrator | Friday 29 August 2025 17:13:15 +0000 (0:00:00.245) 0:04:05.979 ********* 2025-08-29 17:14:22.668896 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:22.668907 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:22.668917 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:22.668928 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:22.668938 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:22.668949 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:22.668959 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:22.668970 | orchestrator | 2025-08-29 17:14:22.668981 | orchestrator | TASK [osism.commons.cleanup : Populate service facts] ************************** 2025-08-29 17:14:22.668992 | orchestrator | Friday 29 August 2025 17:13:15 +0000 (0:00:00.247) 0:04:06.227 ********* 2025-08-29 17:14:22.669002 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:22.669013 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:22.669023 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:22.669033 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:22.669044 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:22.669054 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:22.669064 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:22.669075 | orchestrator | 2025-08-29 17:14:22.669085 | orchestrator | TASK [osism.commons.cleanup : Include distribution specific timer tasks] ******* 2025-08-29 17:14:22.669096 | orchestrator | Friday 29 August 2025 17:13:21 +0000 (0:00:05.639) 0:04:11.867 ********* 2025-08-29 17:14:22.669122 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:14:22.669136 | orchestrator | 2025-08-29 17:14:22.669147 | orchestrator | TASK [osism.commons.cleanup : Disable apt-daily timers] ************************ 2025-08-29 17:14:22.669158 | orchestrator | Friday 29 August 2025 17:13:21 +0000 (0:00:00.331) 0:04:12.198 ********* 2025-08-29 17:14:22.669169 | orchestrator | skipping: [testbed-manager] => (item=apt-daily-upgrade)  2025-08-29 17:14:22.669179 | orchestrator | skipping: [testbed-manager] => (item=apt-daily)  2025-08-29 17:14:22.669230 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:22.669241 | orchestrator | skipping: [testbed-node-0] => (item=apt-daily-upgrade)  2025-08-29 17:14:22.669252 | orchestrator | skipping: [testbed-node-0] => (item=apt-daily)  2025-08-29 17:14:22.669262 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:22.669271 | orchestrator | skipping: [testbed-node-1] => (item=apt-daily-upgrade)  2025-08-29 17:14:22.669281 | orchestrator | skipping: [testbed-node-1] => (item=apt-daily)  2025-08-29 17:14:22.669291 | orchestrator | skipping: [testbed-node-2] => (item=apt-daily-upgrade)  2025-08-29 17:14:22.669308 | orchestrator | skipping: [testbed-node-2] => (item=apt-daily)  2025-08-29 17:14:22.669318 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:22.669328 | orchestrator | skipping: [testbed-node-3] => (item=apt-daily-upgrade)  2025-08-29 17:14:22.669337 | orchestrator | skipping: [testbed-node-3] => (item=apt-daily)  2025-08-29 17:14:22.669347 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:14:22.669356 | orchestrator | skipping: [testbed-node-4] => (item=apt-daily-upgrade)  2025-08-29 17:14:22.669366 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:14:22.669375 | orchestrator | skipping: [testbed-node-4] => (item=apt-daily)  2025-08-29 17:14:22.669400 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:14:22.669410 | orchestrator | skipping: [testbed-node-5] => (item=apt-daily-upgrade)  2025-08-29 17:14:22.669419 | orchestrator | skipping: [testbed-node-5] => (item=apt-daily)  2025-08-29 17:14:22.669429 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:14:22.669438 | orchestrator | 2025-08-29 17:14:22.669448 | orchestrator | TASK [osism.commons.cleanup : Include service tasks] *************************** 2025-08-29 17:14:22.669458 | orchestrator | Friday 29 August 2025 17:13:22 +0000 (0:00:00.278) 0:04:12.477 ********* 2025-08-29 17:14:22.669468 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:14:22.669478 | orchestrator | 2025-08-29 17:14:22.669487 | orchestrator | TASK [osism.commons.cleanup : Cleanup services] ******************************** 2025-08-29 17:14:22.669497 | orchestrator | Friday 29 August 2025 17:13:22 +0000 (0:00:00.337) 0:04:12.814 ********* 2025-08-29 17:14:22.669506 | orchestrator | skipping: [testbed-manager] => (item=ModemManager.service)  2025-08-29 17:14:22.669516 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:22.669526 | orchestrator | skipping: [testbed-node-0] => (item=ModemManager.service)  2025-08-29 17:14:22.669535 | orchestrator | skipping: [testbed-node-1] => (item=ModemManager.service)  2025-08-29 17:14:22.669544 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:22.669554 | orchestrator | skipping: [testbed-node-2] => (item=ModemManager.service)  2025-08-29 17:14:22.669563 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:22.669573 | orchestrator | skipping: [testbed-node-3] => (item=ModemManager.service)  2025-08-29 17:14:22.669582 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:14:22.669592 | orchestrator | skipping: [testbed-node-4] => (item=ModemManager.service)  2025-08-29 17:14:22.669601 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:14:22.669611 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:14:22.669620 | orchestrator | skipping: [testbed-node-5] => (item=ModemManager.service)  2025-08-29 17:14:22.669630 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:14:22.669640 | orchestrator | 2025-08-29 17:14:22.669649 | orchestrator | TASK [osism.commons.cleanup : Include packages tasks] ************************** 2025-08-29 17:14:22.669659 | orchestrator | Friday 29 August 2025 17:13:22 +0000 (0:00:00.279) 0:04:13.093 ********* 2025-08-29 17:14:22.669669 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:14:22.669679 | orchestrator | 2025-08-29 17:14:22.669688 | orchestrator | TASK [osism.commons.cleanup : Cleanup installed packages] ********************** 2025-08-29 17:14:22.669698 | orchestrator | Friday 29 August 2025 17:13:23 +0000 (0:00:00.341) 0:04:13.435 ********* 2025-08-29 17:14:22.669707 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:22.669717 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:22.669727 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:22.669736 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:22.669746 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:22.669755 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:22.669770 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:22.669780 | orchestrator | 2025-08-29 17:14:22.669789 | orchestrator | TASK [osism.commons.cleanup : Remove cloudinit package] ************************ 2025-08-29 17:14:22.669799 | orchestrator | Friday 29 August 2025 17:13:59 +0000 (0:00:35.898) 0:04:49.334 ********* 2025-08-29 17:14:22.669809 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:22.669818 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:22.669828 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:22.669837 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:22.669847 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:22.669856 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:22.669866 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:22.669875 | orchestrator | 2025-08-29 17:14:22.669885 | orchestrator | TASK [osism.commons.cleanup : Uninstall unattended-upgrades package] *********** 2025-08-29 17:14:22.669895 | orchestrator | Friday 29 August 2025 17:14:07 +0000 (0:00:08.316) 0:04:57.651 ********* 2025-08-29 17:14:22.669904 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:22.669914 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:22.669923 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:22.669933 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:22.669942 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:22.669951 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:22.669961 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:22.669970 | orchestrator | 2025-08-29 17:14:22.669980 | orchestrator | TASK [osism.commons.cleanup : Remove useless packages from the cache] ********** 2025-08-29 17:14:22.669989 | orchestrator | Friday 29 August 2025 17:14:15 +0000 (0:00:07.740) 0:05:05.392 ********* 2025-08-29 17:14:22.669999 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:22.670008 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:22.670092 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:22.670103 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:22.670112 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:22.670122 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:22.670131 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:22.670141 | orchestrator | 2025-08-29 17:14:22.670150 | orchestrator | TASK [osism.commons.cleanup : Remove dependencies that are no longer required] *** 2025-08-29 17:14:22.670160 | orchestrator | Friday 29 August 2025 17:14:16 +0000 (0:00:01.654) 0:05:07.047 ********* 2025-08-29 17:14:22.670170 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:22.670179 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:22.670206 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:22.670216 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:22.670225 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:22.670235 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:22.670245 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:22.670255 | orchestrator | 2025-08-29 17:14:22.670264 | orchestrator | TASK [osism.commons.cleanup : Include cloudinit tasks] ************************* 2025-08-29 17:14:22.670282 | orchestrator | Friday 29 August 2025 17:14:22 +0000 (0:00:05.863) 0:05:12.910 ********* 2025-08-29 17:14:32.690913 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:14:32.690973 | orchestrator | 2025-08-29 17:14:32.690983 | orchestrator | TASK [osism.commons.cleanup : Remove cloud-init configuration directory] ******* 2025-08-29 17:14:32.690990 | orchestrator | Friday 29 August 2025 17:14:23 +0000 (0:00:00.412) 0:05:13.323 ********* 2025-08-29 17:14:32.690997 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:32.691004 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:32.691011 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:32.691018 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:32.691024 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:32.691030 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:32.691037 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:32.691057 | orchestrator | 2025-08-29 17:14:32.691064 | orchestrator | TASK [osism.commons.timezone : Install tzdata package] ************************* 2025-08-29 17:14:32.691070 | orchestrator | Friday 29 August 2025 17:14:23 +0000 (0:00:00.722) 0:05:14.045 ********* 2025-08-29 17:14:32.691076 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:32.691083 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:32.691089 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:32.691095 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:32.691102 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:32.691108 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:32.691114 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:32.691121 | orchestrator | 2025-08-29 17:14:32.691127 | orchestrator | TASK [osism.commons.timezone : Set timezone to UTC] **************************** 2025-08-29 17:14:32.691133 | orchestrator | Friday 29 August 2025 17:14:25 +0000 (0:00:01.695) 0:05:15.741 ********* 2025-08-29 17:14:32.691139 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:14:32.691146 | orchestrator | changed: [testbed-manager] 2025-08-29 17:14:32.691152 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:14:32.691158 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:14:32.691164 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:14:32.691178 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:14:32.691185 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:14:32.691205 | orchestrator | 2025-08-29 17:14:32.691212 | orchestrator | TASK [osism.commons.timezone : Create /etc/adjtime file] *********************** 2025-08-29 17:14:32.691218 | orchestrator | Friday 29 August 2025 17:14:26 +0000 (0:00:00.755) 0:05:16.497 ********* 2025-08-29 17:14:32.691225 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:32.691231 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:32.691237 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:32.691243 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:14:32.691250 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:14:32.691256 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:14:32.691262 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:14:32.691269 | orchestrator | 2025-08-29 17:14:32.691275 | orchestrator | TASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] ********************* 2025-08-29 17:14:32.691281 | orchestrator | Friday 29 August 2025 17:14:26 +0000 (0:00:00.244) 0:05:16.742 ********* 2025-08-29 17:14:32.691287 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:32.691293 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:32.691300 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:32.691306 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:14:32.691312 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:14:32.691319 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:14:32.691325 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:14:32.691331 | orchestrator | 2025-08-29 17:14:32.691338 | orchestrator | TASK [osism.services.docker : Gather variables for each operating system] ****** 2025-08-29 17:14:32.691344 | orchestrator | Friday 29 August 2025 17:14:26 +0000 (0:00:00.358) 0:05:17.100 ********* 2025-08-29 17:14:32.691350 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:32.691357 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:32.691363 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:32.691369 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:32.691375 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:32.691381 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:32.691388 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:32.691394 | orchestrator | 2025-08-29 17:14:32.691404 | orchestrator | TASK [osism.services.docker : Set docker_version variable to default value] **** 2025-08-29 17:14:32.691410 | orchestrator | Friday 29 August 2025 17:14:27 +0000 (0:00:00.279) 0:05:17.380 ********* 2025-08-29 17:14:32.691417 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:32.691423 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:32.691429 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:32.691435 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:14:32.691442 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:14:32.691453 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:14:32.691459 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:14:32.691465 | orchestrator | 2025-08-29 17:14:32.691472 | orchestrator | TASK [osism.services.docker : Set docker_cli_version variable to default value] *** 2025-08-29 17:14:32.691478 | orchestrator | Friday 29 August 2025 17:14:27 +0000 (0:00:00.234) 0:05:17.614 ********* 2025-08-29 17:14:32.691485 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:32.691491 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:32.691497 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:32.691504 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:32.691510 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:32.691517 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:32.691523 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:32.691530 | orchestrator | 2025-08-29 17:14:32.691537 | orchestrator | TASK [osism.services.docker : Print used docker version] *********************** 2025-08-29 17:14:32.691544 | orchestrator | Friday 29 August 2025 17:14:27 +0000 (0:00:00.221) 0:05:17.836 ********* 2025-08-29 17:14:32.691550 | orchestrator | ok: [testbed-manager] =>  2025-08-29 17:14:32.691557 | orchestrator |  docker_version: 5:27.5.1 2025-08-29 17:14:32.691564 | orchestrator | ok: [testbed-node-0] =>  2025-08-29 17:14:32.691571 | orchestrator |  docker_version: 5:27.5.1 2025-08-29 17:14:32.691577 | orchestrator | ok: [testbed-node-1] =>  2025-08-29 17:14:32.691584 | orchestrator |  docker_version: 5:27.5.1 2025-08-29 17:14:32.691591 | orchestrator | ok: [testbed-node-2] =>  2025-08-29 17:14:32.691598 | orchestrator |  docker_version: 5:27.5.1 2025-08-29 17:14:32.691604 | orchestrator | ok: [testbed-node-3] =>  2025-08-29 17:14:32.691611 | orchestrator |  docker_version: 5:27.5.1 2025-08-29 17:14:32.691627 | orchestrator | ok: [testbed-node-4] =>  2025-08-29 17:14:32.691634 | orchestrator |  docker_version: 5:27.5.1 2025-08-29 17:14:32.691641 | orchestrator | ok: [testbed-node-5] =>  2025-08-29 17:14:32.691647 | orchestrator |  docker_version: 5:27.5.1 2025-08-29 17:14:32.691654 | orchestrator | 2025-08-29 17:14:32.691661 | orchestrator | TASK [osism.services.docker : Print used docker cli version] ******************* 2025-08-29 17:14:32.691668 | orchestrator | Friday 29 August 2025 17:14:27 +0000 (0:00:00.218) 0:05:18.054 ********* 2025-08-29 17:14:32.691675 | orchestrator | ok: [testbed-manager] =>  2025-08-29 17:14:32.691682 | orchestrator |  docker_cli_version: 5:27.5.1 2025-08-29 17:14:32.691688 | orchestrator | ok: [testbed-node-0] =>  2025-08-29 17:14:32.691695 | orchestrator |  docker_cli_version: 5:27.5.1 2025-08-29 17:14:32.691702 | orchestrator | ok: [testbed-node-1] =>  2025-08-29 17:14:32.691708 | orchestrator |  docker_cli_version: 5:27.5.1 2025-08-29 17:14:32.691715 | orchestrator | ok: [testbed-node-2] =>  2025-08-29 17:14:32.691722 | orchestrator |  docker_cli_version: 5:27.5.1 2025-08-29 17:14:32.691729 | orchestrator | ok: [testbed-node-3] =>  2025-08-29 17:14:32.691735 | orchestrator |  docker_cli_version: 5:27.5.1 2025-08-29 17:14:32.691742 | orchestrator | ok: [testbed-node-4] =>  2025-08-29 17:14:32.691749 | orchestrator |  docker_cli_version: 5:27.5.1 2025-08-29 17:14:32.691755 | orchestrator | ok: [testbed-node-5] =>  2025-08-29 17:14:32.691762 | orchestrator |  docker_cli_version: 5:27.5.1 2025-08-29 17:14:32.691769 | orchestrator | 2025-08-29 17:14:32.691776 | orchestrator | TASK [osism.services.docker : Include block storage tasks] ********************* 2025-08-29 17:14:32.691782 | orchestrator | Friday 29 August 2025 17:14:28 +0000 (0:00:00.267) 0:05:18.322 ********* 2025-08-29 17:14:32.691789 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:32.691795 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:32.691802 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:32.691809 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:14:32.691815 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:14:32.691822 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:14:32.691828 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:14:32.691835 | orchestrator | 2025-08-29 17:14:32.691841 | orchestrator | TASK [osism.services.docker : Include zram storage tasks] ********************** 2025-08-29 17:14:32.691852 | orchestrator | Friday 29 August 2025 17:14:28 +0000 (0:00:00.220) 0:05:18.542 ********* 2025-08-29 17:14:32.691859 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:32.691866 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:32.691873 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:32.691879 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:14:32.691885 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:14:32.691891 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:14:32.691897 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:14:32.691903 | orchestrator | 2025-08-29 17:14:32.691910 | orchestrator | TASK [osism.services.docker : Include docker install tasks] ******************** 2025-08-29 17:14:32.691916 | orchestrator | Friday 29 August 2025 17:14:28 +0000 (0:00:00.223) 0:05:18.766 ********* 2025-08-29 17:14:32.691924 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:14:32.691931 | orchestrator | 2025-08-29 17:14:32.691937 | orchestrator | TASK [osism.services.docker : Remove old architecture-dependent repository] **** 2025-08-29 17:14:32.691943 | orchestrator | Friday 29 August 2025 17:14:28 +0000 (0:00:00.350) 0:05:19.117 ********* 2025-08-29 17:14:32.691949 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:32.691955 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:32.691962 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:32.691968 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:32.691974 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:32.691980 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:32.691986 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:32.691993 | orchestrator | 2025-08-29 17:14:32.691999 | orchestrator | TASK [osism.services.docker : Gather package facts] **************************** 2025-08-29 17:14:32.692005 | orchestrator | Friday 29 August 2025 17:14:29 +0000 (0:00:00.775) 0:05:19.892 ********* 2025-08-29 17:14:32.692012 | orchestrator | ok: [testbed-manager] 2025-08-29 17:14:32.692020 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:14:32.692027 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:14:32.692033 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:14:32.692039 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:14:32.692045 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:14:32.692051 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:14:32.692057 | orchestrator | 2025-08-29 17:14:32.692063 | orchestrator | TASK [osism.services.docker : Check whether packages are installed that should not be installed] *** 2025-08-29 17:14:32.692070 | orchestrator | Friday 29 August 2025 17:14:32 +0000 (0:00:02.626) 0:05:22.519 ********* 2025-08-29 17:14:32.692077 | orchestrator | skipping: [testbed-manager] => (item=containerd)  2025-08-29 17:14:32.692083 | orchestrator | skipping: [testbed-manager] => (item=docker.io)  2025-08-29 17:14:32.692090 | orchestrator | skipping: [testbed-manager] => (item=docker-engine)  2025-08-29 17:14:32.692095 | orchestrator | skipping: [testbed-node-0] => (item=containerd)  2025-08-29 17:14:32.692101 | orchestrator | skipping: [testbed-node-0] => (item=docker.io)  2025-08-29 17:14:32.692107 | orchestrator | skipping: [testbed-node-0] => (item=docker-engine)  2025-08-29 17:14:32.692113 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:14:32.692119 | orchestrator | skipping: [testbed-node-1] => (item=containerd)  2025-08-29 17:14:32.692126 | orchestrator | skipping: [testbed-node-1] => (item=docker.io)  2025-08-29 17:14:32.692132 | orchestrator | skipping: [testbed-node-1] => (item=docker-engine)  2025-08-29 17:14:32.692138 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:14:32.692144 | orchestrator | skipping: [testbed-node-2] => (item=containerd)  2025-08-29 17:14:32.692150 | orchestrator | skipping: [testbed-node-2] => (item=docker.io)  2025-08-29 17:14:32.692156 | orchestrator | skipping: [testbed-node-2] => (item=docker-engine)  2025-08-29 17:14:32.692163 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:14:32.692169 | orchestrator | skipping: [testbed-node-3] => (item=containerd)  2025-08-29 17:14:32.692179 | orchestrator | skipping: [testbed-node-3] => (item=docker.io)  2025-08-29 17:14:32.692201 | orchestrator | skipping: [testbed-node-3] => (item=docker-engine)  2025-08-29 17:15:32.760443 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:32.760610 | orchestrator | skipping: [testbed-node-4] => (item=containerd)  2025-08-29 17:15:32.760629 | orchestrator | skipping: [testbed-node-4] => (item=docker.io)  2025-08-29 17:15:32.760641 | orchestrator | skipping: [testbed-node-4] => (item=docker-engine)  2025-08-29 17:15:32.760692 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:32.760703 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:32.760748 | orchestrator | skipping: [testbed-node-5] => (item=containerd)  2025-08-29 17:15:32.760761 | orchestrator | skipping: [testbed-node-5] => (item=docker.io)  2025-08-29 17:15:32.760772 | orchestrator | skipping: [testbed-node-5] => (item=docker-engine)  2025-08-29 17:15:32.760783 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:32.760794 | orchestrator | 2025-08-29 17:15:32.760806 | orchestrator | TASK [osism.services.docker : Install apt-transport-https package] ************* 2025-08-29 17:15:32.760819 | orchestrator | Friday 29 August 2025 17:14:32 +0000 (0:00:00.603) 0:05:23.122 ********* 2025-08-29 17:15:32.760830 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.760841 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.760852 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.760863 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.760873 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.760884 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.760895 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.760906 | orchestrator | 2025-08-29 17:15:32.760917 | orchestrator | TASK [osism.services.docker : Add repository gpg key] ************************** 2025-08-29 17:15:32.760928 | orchestrator | Friday 29 August 2025 17:14:39 +0000 (0:00:06.313) 0:05:29.435 ********* 2025-08-29 17:15:32.760939 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.760950 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.760961 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.760972 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.760982 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.760993 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.761004 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.761015 | orchestrator | 2025-08-29 17:15:32.761025 | orchestrator | TASK [osism.services.docker : Add repository] ********************************** 2025-08-29 17:15:32.761071 | orchestrator | Friday 29 August 2025 17:14:40 +0000 (0:00:00.960) 0:05:30.396 ********* 2025-08-29 17:15:32.761084 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.761130 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.761142 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.761186 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.761197 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.761208 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.761300 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.761648 | orchestrator | 2025-08-29 17:15:32.761694 | orchestrator | TASK [osism.services.docker : Update package cache] **************************** 2025-08-29 17:15:32.761706 | orchestrator | Friday 29 August 2025 17:14:47 +0000 (0:00:07.558) 0:05:37.954 ********* 2025-08-29 17:15:32.761717 | orchestrator | changed: [testbed-manager] 2025-08-29 17:15:32.761728 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.761738 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.761749 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.761760 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.761771 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.761781 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.761792 | orchestrator | 2025-08-29 17:15:32.761803 | orchestrator | TASK [osism.services.docker : Pin docker package version] ********************** 2025-08-29 17:15:32.761814 | orchestrator | Friday 29 August 2025 17:14:50 +0000 (0:00:03.157) 0:05:41.112 ********* 2025-08-29 17:15:32.761825 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.761862 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.761873 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.761884 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.761895 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.761905 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.761916 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.761926 | orchestrator | 2025-08-29 17:15:32.761937 | orchestrator | TASK [osism.services.docker : Pin docker-cli package version] ****************** 2025-08-29 17:15:32.761970 | orchestrator | Friday 29 August 2025 17:14:52 +0000 (0:00:01.450) 0:05:42.562 ********* 2025-08-29 17:15:32.761982 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.761992 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.762003 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.762014 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.762070 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.762081 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.762091 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.762102 | orchestrator | 2025-08-29 17:15:32.762113 | orchestrator | TASK [osism.services.docker : Unlock containerd package] *********************** 2025-08-29 17:15:32.762123 | orchestrator | Friday 29 August 2025 17:14:53 +0000 (0:00:01.316) 0:05:43.879 ********* 2025-08-29 17:15:32.762134 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:15:32.762145 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:15:32.762156 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:32.762167 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:32.762177 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:32.762188 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:32.762199 | orchestrator | changed: [testbed-manager] 2025-08-29 17:15:32.762209 | orchestrator | 2025-08-29 17:15:32.762237 | orchestrator | TASK [osism.services.docker : Install containerd package] ********************** 2025-08-29 17:15:32.762248 | orchestrator | Friday 29 August 2025 17:14:54 +0000 (0:00:00.583) 0:05:44.463 ********* 2025-08-29 17:15:32.762259 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.762270 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.762280 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.762291 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.762301 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.762311 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.762322 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.762333 | orchestrator | 2025-08-29 17:15:32.762343 | orchestrator | TASK [osism.services.docker : Lock containerd package] ************************* 2025-08-29 17:15:32.762354 | orchestrator | Friday 29 August 2025 17:15:04 +0000 (0:00:10.010) 0:05:54.474 ********* 2025-08-29 17:15:32.762365 | orchestrator | changed: [testbed-manager] 2025-08-29 17:15:32.762395 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.762407 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.762418 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.762428 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.762439 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.762449 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.762460 | orchestrator | 2025-08-29 17:15:32.762471 | orchestrator | TASK [osism.services.docker : Install docker-cli package] ********************** 2025-08-29 17:15:32.762481 | orchestrator | Friday 29 August 2025 17:15:05 +0000 (0:00:00.844) 0:05:55.318 ********* 2025-08-29 17:15:32.762492 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.762503 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.762514 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.762524 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.762535 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.762546 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.762556 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.762567 | orchestrator | 2025-08-29 17:15:32.762578 | orchestrator | TASK [osism.services.docker : Install docker package] ************************** 2025-08-29 17:15:32.762598 | orchestrator | Friday 29 August 2025 17:15:14 +0000 (0:00:09.563) 0:06:04.882 ********* 2025-08-29 17:15:32.762609 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.762620 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.763025 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.763103 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.763123 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.763197 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.763215 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.763309 | orchestrator | 2025-08-29 17:15:32.763331 | orchestrator | TASK [osism.services.docker : Unblock installation of python docker packages] *** 2025-08-29 17:15:32.763352 | orchestrator | Friday 29 August 2025 17:15:26 +0000 (0:00:11.730) 0:06:16.612 ********* 2025-08-29 17:15:32.763365 | orchestrator | ok: [testbed-manager] => (item=python3-docker) 2025-08-29 17:15:32.763376 | orchestrator | ok: [testbed-node-0] => (item=python3-docker) 2025-08-29 17:15:32.763387 | orchestrator | ok: [testbed-node-2] => (item=python3-docker) 2025-08-29 17:15:32.763398 | orchestrator | ok: [testbed-manager] => (item=python-docker) 2025-08-29 17:15:32.763408 | orchestrator | ok: [testbed-node-3] => (item=python3-docker) 2025-08-29 17:15:32.763419 | orchestrator | ok: [testbed-node-1] => (item=python3-docker) 2025-08-29 17:15:32.763429 | orchestrator | ok: [testbed-node-4] => (item=python3-docker) 2025-08-29 17:15:32.763439 | orchestrator | ok: [testbed-node-0] => (item=python-docker) 2025-08-29 17:15:32.763450 | orchestrator | ok: [testbed-node-5] => (item=python3-docker) 2025-08-29 17:15:32.763461 | orchestrator | ok: [testbed-node-2] => (item=python-docker) 2025-08-29 17:15:32.763471 | orchestrator | ok: [testbed-node-3] => (item=python-docker) 2025-08-29 17:15:32.763482 | orchestrator | ok: [testbed-node-1] => (item=python-docker) 2025-08-29 17:15:32.763492 | orchestrator | ok: [testbed-node-4] => (item=python-docker) 2025-08-29 17:15:32.763503 | orchestrator | ok: [testbed-node-5] => (item=python-docker) 2025-08-29 17:15:32.763513 | orchestrator | 2025-08-29 17:15:32.763524 | orchestrator | TASK [osism.services.docker : Install python3 docker package] ****************** 2025-08-29 17:15:32.763534 | orchestrator | Friday 29 August 2025 17:15:27 +0000 (0:00:01.229) 0:06:17.841 ********* 2025-08-29 17:15:32.763593 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:15:32.763914 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:15:32.763928 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:15:32.764005 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:32.764054 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:32.764143 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:32.764155 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:32.764166 | orchestrator | 2025-08-29 17:15:32.764176 | orchestrator | TASK [osism.services.docker : Install python3 docker package from Debian Sid] *** 2025-08-29 17:15:32.764240 | orchestrator | Friday 29 August 2025 17:15:28 +0000 (0:00:00.504) 0:06:18.346 ********* 2025-08-29 17:15:32.764254 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:32.764266 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:32.764277 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:32.764298 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:32.764309 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:32.764320 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:32.764331 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:32.764341 | orchestrator | 2025-08-29 17:15:32.764353 | orchestrator | TASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] *** 2025-08-29 17:15:32.764365 | orchestrator | Friday 29 August 2025 17:15:32 +0000 (0:00:03.925) 0:06:22.271 ********* 2025-08-29 17:15:32.764376 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:15:32.764387 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:15:32.764398 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:15:32.764408 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:32.764419 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:32.764442 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:32.764452 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:32.764463 | orchestrator | 2025-08-29 17:15:32.764475 | orchestrator | TASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] *** 2025-08-29 17:15:32.764486 | orchestrator | Friday 29 August 2025 17:15:32 +0000 (0:00:00.464) 0:06:22.735 ********* 2025-08-29 17:15:32.764496 | orchestrator | skipping: [testbed-manager] => (item=python3-docker)  2025-08-29 17:15:32.764507 | orchestrator | skipping: [testbed-manager] => (item=python-docker)  2025-08-29 17:15:32.764518 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:15:32.764528 | orchestrator | skipping: [testbed-node-0] => (item=python3-docker)  2025-08-29 17:15:32.764539 | orchestrator | skipping: [testbed-node-0] => (item=python-docker)  2025-08-29 17:15:32.764550 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:15:32.764560 | orchestrator | skipping: [testbed-node-1] => (item=python3-docker)  2025-08-29 17:15:32.764571 | orchestrator | skipping: [testbed-node-1] => (item=python-docker)  2025-08-29 17:15:32.764620 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:15:32.764667 | orchestrator | skipping: [testbed-node-2] => (item=python3-docker)  2025-08-29 17:15:32.764696 | orchestrator | skipping: [testbed-node-2] => (item=python-docker)  2025-08-29 17:15:51.972109 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:51.972189 | orchestrator | skipping: [testbed-node-3] => (item=python3-docker)  2025-08-29 17:15:51.972196 | orchestrator | skipping: [testbed-node-3] => (item=python-docker)  2025-08-29 17:15:51.972201 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:51.972206 | orchestrator | skipping: [testbed-node-4] => (item=python3-docker)  2025-08-29 17:15:51.972210 | orchestrator | skipping: [testbed-node-4] => (item=python-docker)  2025-08-29 17:15:51.972215 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:51.972219 | orchestrator | skipping: [testbed-node-5] => (item=python3-docker)  2025-08-29 17:15:51.972223 | orchestrator | skipping: [testbed-node-5] => (item=python-docker)  2025-08-29 17:15:51.972227 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:51.972283 | orchestrator | 2025-08-29 17:15:51.972290 | orchestrator | TASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] *** 2025-08-29 17:15:51.972295 | orchestrator | Friday 29 August 2025 17:15:32 +0000 (0:00:00.495) 0:06:23.231 ********* 2025-08-29 17:15:51.972299 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:15:51.972303 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:15:51.972307 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:15:51.972311 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:51.972315 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:51.972319 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:51.972323 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:51.972327 | orchestrator | 2025-08-29 17:15:51.972331 | orchestrator | TASK [osism.services.docker : Install docker packages (install python bindings from pip)] *** 2025-08-29 17:15:51.972336 | orchestrator | Friday 29 August 2025 17:15:33 +0000 (0:00:00.472) 0:06:23.704 ********* 2025-08-29 17:15:51.972340 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:15:51.972343 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:15:51.972347 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:15:51.972351 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:51.972355 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:51.972358 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:51.972362 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:51.972366 | orchestrator | 2025-08-29 17:15:51.972370 | orchestrator | TASK [osism.services.docker : Install packages required by docker login] ******* 2025-08-29 17:15:51.972374 | orchestrator | Friday 29 August 2025 17:15:33 +0000 (0:00:00.455) 0:06:24.159 ********* 2025-08-29 17:15:51.972377 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:15:51.972381 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:15:51.972385 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:15:51.972405 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:15:51.972409 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:15:51.972412 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:15:51.972416 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:15:51.972420 | orchestrator | 2025-08-29 17:15:51.972424 | orchestrator | TASK [osism.services.docker : Ensure that some packages are not installed] ***** 2025-08-29 17:15:51.972428 | orchestrator | Friday 29 August 2025 17:15:34 +0000 (0:00:00.673) 0:06:24.832 ********* 2025-08-29 17:15:51.972432 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972436 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:15:51.972440 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:15:51.972443 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:15:51.972447 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:15:51.972451 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:15:51.972454 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:15:51.972458 | orchestrator | 2025-08-29 17:15:51.972462 | orchestrator | TASK [osism.services.docker : Include config tasks] **************************** 2025-08-29 17:15:51.972466 | orchestrator | Friday 29 August 2025 17:15:36 +0000 (0:00:01.915) 0:06:26.747 ********* 2025-08-29 17:15:51.972471 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:15:51.972477 | orchestrator | 2025-08-29 17:15:51.972481 | orchestrator | TASK [osism.services.docker : Create plugins directory] ************************ 2025-08-29 17:15:51.972485 | orchestrator | Friday 29 August 2025 17:15:37 +0000 (0:00:00.803) 0:06:27.551 ********* 2025-08-29 17:15:51.972489 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972493 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:51.972497 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:51.972501 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:51.972505 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:51.972510 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:51.972516 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:51.972523 | orchestrator | 2025-08-29 17:15:51.972527 | orchestrator | TASK [osism.services.docker : Create systemd overlay directory] **************** 2025-08-29 17:15:51.972531 | orchestrator | Friday 29 August 2025 17:15:38 +0000 (0:00:00.839) 0:06:28.391 ********* 2025-08-29 17:15:51.972535 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972539 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:51.972543 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:51.972547 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:51.972551 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:51.972555 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:51.972559 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:51.972562 | orchestrator | 2025-08-29 17:15:51.972566 | orchestrator | TASK [osism.services.docker : Copy systemd overlay file] *********************** 2025-08-29 17:15:51.972570 | orchestrator | Friday 29 August 2025 17:15:39 +0000 (0:00:01.049) 0:06:29.441 ********* 2025-08-29 17:15:51.972574 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972578 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:51.972581 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:51.972585 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:51.972589 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:51.972592 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:51.972608 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:51.972612 | orchestrator | 2025-08-29 17:15:51.972615 | orchestrator | TASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] *** 2025-08-29 17:15:51.972619 | orchestrator | Friday 29 August 2025 17:15:40 +0000 (0:00:01.395) 0:06:30.837 ********* 2025-08-29 17:15:51.972633 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:15:51.972637 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:15:51.972640 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:15:51.972644 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:15:51.972652 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:15:51.972656 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:15:51.972659 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:15:51.972663 | orchestrator | 2025-08-29 17:15:51.972667 | orchestrator | TASK [osism.services.docker : Copy limits configuration file] ****************** 2025-08-29 17:15:51.972671 | orchestrator | Friday 29 August 2025 17:15:42 +0000 (0:00:01.438) 0:06:32.276 ********* 2025-08-29 17:15:51.972674 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972678 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:51.972682 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:51.972686 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:51.972690 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:51.972694 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:51.972698 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:51.972702 | orchestrator | 2025-08-29 17:15:51.972707 | orchestrator | TASK [osism.services.docker : Copy daemon.json configuration file] ************* 2025-08-29 17:15:51.972711 | orchestrator | Friday 29 August 2025 17:15:43 +0000 (0:00:01.332) 0:06:33.608 ********* 2025-08-29 17:15:51.972715 | orchestrator | changed: [testbed-manager] 2025-08-29 17:15:51.972719 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:15:51.972723 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:15:51.972728 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:15:51.972732 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:15:51.972736 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:15:51.972740 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:15:51.972744 | orchestrator | 2025-08-29 17:15:51.972748 | orchestrator | TASK [osism.services.docker : Include service tasks] *************************** 2025-08-29 17:15:51.972753 | orchestrator | Friday 29 August 2025 17:15:44 +0000 (0:00:01.351) 0:06:34.959 ********* 2025-08-29 17:15:51.972757 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:15:51.972761 | orchestrator | 2025-08-29 17:15:51.972766 | orchestrator | TASK [osism.services.docker : Reload systemd daemon] *************************** 2025-08-29 17:15:51.972770 | orchestrator | Friday 29 August 2025 17:15:45 +0000 (0:00:00.964) 0:06:35.924 ********* 2025-08-29 17:15:51.972774 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972778 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:15:51.972782 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:15:51.972786 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:15:51.972790 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:15:51.972794 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:15:51.972798 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:15:51.972803 | orchestrator | 2025-08-29 17:15:51.972807 | orchestrator | TASK [osism.services.docker : Manage service] ********************************** 2025-08-29 17:15:51.972811 | orchestrator | Friday 29 August 2025 17:15:47 +0000 (0:00:01.353) 0:06:37.278 ********* 2025-08-29 17:15:51.972815 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972819 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:15:51.972823 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:15:51.972827 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:15:51.972832 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:15:51.972836 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:15:51.972840 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:15:51.972844 | orchestrator | 2025-08-29 17:15:51.972848 | orchestrator | TASK [osism.services.docker : Manage docker socket service] ******************** 2025-08-29 17:15:51.972852 | orchestrator | Friday 29 August 2025 17:15:48 +0000 (0:00:01.116) 0:06:38.394 ********* 2025-08-29 17:15:51.972857 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972861 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:15:51.972865 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:15:51.972869 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:15:51.972873 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:15:51.972877 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:15:51.972884 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:15:51.972888 | orchestrator | 2025-08-29 17:15:51.972892 | orchestrator | TASK [osism.services.docker : Manage containerd service] *********************** 2025-08-29 17:15:51.972905 | orchestrator | Friday 29 August 2025 17:15:49 +0000 (0:00:01.458) 0:06:39.853 ********* 2025-08-29 17:15:51.972909 | orchestrator | ok: [testbed-manager] 2025-08-29 17:15:51.972913 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:15:51.972917 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:15:51.972921 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:15:51.972925 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:15:51.972930 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:15:51.972934 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:15:51.972938 | orchestrator | 2025-08-29 17:15:51.972942 | orchestrator | TASK [osism.services.docker : Include bootstrap tasks] ************************* 2025-08-29 17:15:51.972946 | orchestrator | Friday 29 August 2025 17:15:50 +0000 (0:00:01.217) 0:06:41.070 ********* 2025-08-29 17:15:51.972950 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:15:51.972955 | orchestrator | 2025-08-29 17:15:51.972959 | orchestrator | TASK [osism.services.docker : Flush handlers] ********************************** 2025-08-29 17:15:51.972963 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.856) 0:06:41.927 ********* 2025-08-29 17:15:51.972967 | orchestrator | 2025-08-29 17:15:51.972972 | orchestrator | TASK [osism.services.docker : Flush handlers] ********************************** 2025-08-29 17:15:51.972976 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.039) 0:06:41.966 ********* 2025-08-29 17:15:51.972980 | orchestrator | 2025-08-29 17:15:51.972984 | orchestrator | TASK [osism.services.docker : Flush handlers] ********************************** 2025-08-29 17:15:51.972988 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.038) 0:06:42.005 ********* 2025-08-29 17:15:51.972992 | orchestrator | 2025-08-29 17:15:51.972997 | orchestrator | TASK [osism.services.docker : Flush handlers] ********************************** 2025-08-29 17:15:51.973001 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.046) 0:06:42.051 ********* 2025-08-29 17:15:51.973005 | orchestrator | 2025-08-29 17:15:51.973012 | orchestrator | TASK [osism.services.docker : Flush handlers] ********************************** 2025-08-29 17:16:17.530345 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.038) 0:06:42.090 ********* 2025-08-29 17:16:17.530472 | orchestrator | 2025-08-29 17:16:17.530489 | orchestrator | TASK [osism.services.docker : Flush handlers] ********************************** 2025-08-29 17:16:17.530501 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.038) 0:06:42.129 ********* 2025-08-29 17:16:17.530512 | orchestrator | 2025-08-29 17:16:17.530524 | orchestrator | TASK [osism.services.docker : Flush handlers] ********************************** 2025-08-29 17:16:17.530535 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.045) 0:06:42.175 ********* 2025-08-29 17:16:17.530546 | orchestrator | 2025-08-29 17:16:17.530557 | orchestrator | RUNNING HANDLER [osism.commons.repository : Force update of package cache] ***** 2025-08-29 17:16:17.530568 | orchestrator | Friday 29 August 2025 17:15:51 +0000 (0:00:00.039) 0:06:42.214 ********* 2025-08-29 17:16:17.530579 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:17.530591 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:17.530603 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:17.530614 | orchestrator | 2025-08-29 17:16:17.530625 | orchestrator | RUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] ************* 2025-08-29 17:16:17.530635 | orchestrator | Friday 29 August 2025 17:15:53 +0000 (0:00:01.353) 0:06:43.568 ********* 2025-08-29 17:16:17.530646 | orchestrator | changed: [testbed-manager] 2025-08-29 17:16:17.530658 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:17.530669 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:17.530680 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:17.530690 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:17.530701 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:17.530746 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:17.530757 | orchestrator | 2025-08-29 17:16:17.530768 | orchestrator | RUNNING HANDLER [osism.services.smartd : Restart smartd service] *************** 2025-08-29 17:16:17.530779 | orchestrator | Friday 29 August 2025 17:15:54 +0000 (0:00:01.289) 0:06:44.858 ********* 2025-08-29 17:16:17.530790 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:17.530801 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:17.530811 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:17.530822 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:17.530833 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:17.530843 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:17.530854 | orchestrator | changed: [testbed-manager] 2025-08-29 17:16:17.530865 | orchestrator | 2025-08-29 17:16:17.530876 | orchestrator | RUNNING HANDLER [osism.services.docker : Restart docker service] *************** 2025-08-29 17:16:17.530886 | orchestrator | Friday 29 August 2025 17:15:56 +0000 (0:00:01.698) 0:06:46.556 ********* 2025-08-29 17:16:17.530897 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:17.530908 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:17.530918 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:17.530929 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:17.530939 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:17.530950 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:17.530961 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:17.530971 | orchestrator | 2025-08-29 17:16:17.530983 | orchestrator | RUNNING HANDLER [osism.services.docker : Wait after docker service restart] **** 2025-08-29 17:16:17.530994 | orchestrator | Friday 29 August 2025 17:15:58 +0000 (0:00:02.429) 0:06:48.986 ********* 2025-08-29 17:16:17.531005 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:17.531015 | orchestrator | 2025-08-29 17:16:17.531026 | orchestrator | TASK [osism.services.docker : Add user to docker group] ************************ 2025-08-29 17:16:17.531037 | orchestrator | Friday 29 August 2025 17:15:58 +0000 (0:00:00.113) 0:06:49.099 ********* 2025-08-29 17:16:17.531048 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:17.531059 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:17.531070 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:17.531080 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:17.531091 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:17.531101 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:17.531112 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:17.531123 | orchestrator | 2025-08-29 17:16:17.531134 | orchestrator | TASK [osism.services.docker : Log into private registry and force re-authorization] *** 2025-08-29 17:16:17.531161 | orchestrator | Friday 29 August 2025 17:15:59 +0000 (0:00:01.009) 0:06:50.108 ********* 2025-08-29 17:16:17.531172 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:17.531182 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:17.531193 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:16:17.531203 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:16:17.531214 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:16:17.531225 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:16:17.531235 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:16:17.531268 | orchestrator | 2025-08-29 17:16:17.531279 | orchestrator | TASK [osism.services.docker : Include facts tasks] ***************************** 2025-08-29 17:16:17.531290 | orchestrator | Friday 29 August 2025 17:16:00 +0000 (0:00:00.688) 0:06:50.797 ********* 2025-08-29 17:16:17.531302 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:16:17.531315 | orchestrator | 2025-08-29 17:16:17.531326 | orchestrator | TASK [osism.services.docker : Create facts directory] ************************** 2025-08-29 17:16:17.531337 | orchestrator | Friday 29 August 2025 17:16:01 +0000 (0:00:00.925) 0:06:51.723 ********* 2025-08-29 17:16:17.531348 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:17.531359 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:17.531378 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:17.531389 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:17.531400 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:17.531411 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:17.531422 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:17.531432 | orchestrator | 2025-08-29 17:16:17.531443 | orchestrator | TASK [osism.services.docker : Copy docker fact files] ************************** 2025-08-29 17:16:17.531454 | orchestrator | Friday 29 August 2025 17:16:02 +0000 (0:00:00.846) 0:06:52.569 ********* 2025-08-29 17:16:17.531465 | orchestrator | ok: [testbed-manager] => (item=docker_containers) 2025-08-29 17:16:17.531476 | orchestrator | changed: [testbed-node-0] => (item=docker_containers) 2025-08-29 17:16:17.531505 | orchestrator | changed: [testbed-node-1] => (item=docker_containers) 2025-08-29 17:16:17.531518 | orchestrator | changed: [testbed-node-2] => (item=docker_containers) 2025-08-29 17:16:17.531529 | orchestrator | changed: [testbed-node-3] => (item=docker_containers) 2025-08-29 17:16:17.531539 | orchestrator | changed: [testbed-node-4] => (item=docker_containers) 2025-08-29 17:16:17.531550 | orchestrator | changed: [testbed-node-5] => (item=docker_containers) 2025-08-29 17:16:17.531561 | orchestrator | ok: [testbed-manager] => (item=docker_images) 2025-08-29 17:16:17.531572 | orchestrator | changed: [testbed-node-0] => (item=docker_images) 2025-08-29 17:16:17.531583 | orchestrator | changed: [testbed-node-1] => (item=docker_images) 2025-08-29 17:16:17.531593 | orchestrator | changed: [testbed-node-2] => (item=docker_images) 2025-08-29 17:16:17.531604 | orchestrator | changed: [testbed-node-3] => (item=docker_images) 2025-08-29 17:16:17.531615 | orchestrator | changed: [testbed-node-4] => (item=docker_images) 2025-08-29 17:16:17.531626 | orchestrator | changed: [testbed-node-5] => (item=docker_images) 2025-08-29 17:16:17.531636 | orchestrator | 2025-08-29 17:16:17.531647 | orchestrator | TASK [osism.commons.docker_compose : This install type is not supported] ******* 2025-08-29 17:16:17.531658 | orchestrator | Friday 29 August 2025 17:16:05 +0000 (0:00:02.784) 0:06:55.353 ********* 2025-08-29 17:16:17.531669 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:17.531680 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:17.531691 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:16:17.531702 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:16:17.531713 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:16:17.531723 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:16:17.531734 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:16:17.531745 | orchestrator | 2025-08-29 17:16:17.531756 | orchestrator | TASK [osism.commons.docker_compose : Include distribution specific install tasks] *** 2025-08-29 17:16:17.531767 | orchestrator | Friday 29 August 2025 17:16:05 +0000 (0:00:00.489) 0:06:55.842 ********* 2025-08-29 17:16:17.531780 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:16:17.531793 | orchestrator | 2025-08-29 17:16:17.531804 | orchestrator | TASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] *** 2025-08-29 17:16:17.531814 | orchestrator | Friday 29 August 2025 17:16:06 +0000 (0:00:00.845) 0:06:56.688 ********* 2025-08-29 17:16:17.531825 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:17.531836 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:17.531847 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:17.531858 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:17.531868 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:17.531879 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:17.531890 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:17.531901 | orchestrator | 2025-08-29 17:16:17.531911 | orchestrator | TASK [osism.commons.docker_compose : Get checksum of docker-compose file] ****** 2025-08-29 17:16:17.531923 | orchestrator | Friday 29 August 2025 17:16:07 +0000 (0:00:01.057) 0:06:57.746 ********* 2025-08-29 17:16:17.531933 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:17.531951 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:17.531962 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:17.531972 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:17.531983 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:17.531994 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:17.532004 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:17.532015 | orchestrator | 2025-08-29 17:16:17.532026 | orchestrator | TASK [osism.commons.docker_compose : Remove docker-compose binary] ************* 2025-08-29 17:16:17.532037 | orchestrator | Friday 29 August 2025 17:16:08 +0000 (0:00:00.799) 0:06:58.545 ********* 2025-08-29 17:16:17.532048 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:17.532059 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:17.532069 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:16:17.532080 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:16:17.532096 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:16:17.532107 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:16:17.532118 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:16:17.532129 | orchestrator | 2025-08-29 17:16:17.532140 | orchestrator | TASK [osism.commons.docker_compose : Uninstall docker-compose package] ********* 2025-08-29 17:16:17.532151 | orchestrator | Friday 29 August 2025 17:16:08 +0000 (0:00:00.472) 0:06:59.018 ********* 2025-08-29 17:16:17.532162 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:17.532172 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:17.532183 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:17.532194 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:17.532205 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:17.532215 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:17.532226 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:17.532237 | orchestrator | 2025-08-29 17:16:17.532265 | orchestrator | TASK [osism.commons.docker_compose : Copy docker-compose script] *************** 2025-08-29 17:16:17.532277 | orchestrator | Friday 29 August 2025 17:16:10 +0000 (0:00:01.451) 0:07:00.470 ********* 2025-08-29 17:16:17.532287 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:17.532298 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:17.532309 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:16:17.532320 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:16:17.532331 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:16:17.532342 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:16:17.532352 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:16:17.532363 | orchestrator | 2025-08-29 17:16:17.532374 | orchestrator | TASK [osism.commons.docker_compose : Install docker-compose-plugin package] **** 2025-08-29 17:16:17.532385 | orchestrator | Friday 29 August 2025 17:16:10 +0000 (0:00:00.483) 0:07:00.954 ********* 2025-08-29 17:16:17.532395 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:17.532406 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:17.532417 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:17.532427 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:17.532438 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:17.532449 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:17.532460 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:17.532470 | orchestrator | 2025-08-29 17:16:17.532488 | orchestrator | TASK [osism.commons.docker_compose : Copy osism.target systemd file] *********** 2025-08-29 17:16:48.383014 | orchestrator | Friday 29 August 2025 17:16:17 +0000 (0:00:06.819) 0:07:07.773 ********* 2025-08-29 17:16:48.383135 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.383152 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:48.383165 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:48.383177 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:48.383188 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:48.383199 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:48.383211 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:48.383222 | orchestrator | 2025-08-29 17:16:48.383234 | orchestrator | TASK [osism.commons.docker_compose : Enable osism.target] ********************** 2025-08-29 17:16:48.383246 | orchestrator | Friday 29 August 2025 17:16:18 +0000 (0:00:01.214) 0:07:08.987 ********* 2025-08-29 17:16:48.383307 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.383319 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:48.383330 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:48.383342 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:48.383352 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:48.383363 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:48.383373 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:48.383384 | orchestrator | 2025-08-29 17:16:48.383395 | orchestrator | TASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] **** 2025-08-29 17:16:48.383406 | orchestrator | Friday 29 August 2025 17:16:20 +0000 (0:00:01.578) 0:07:10.566 ********* 2025-08-29 17:16:48.383417 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.383428 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:48.383438 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:48.383449 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:48.383459 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:48.383470 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:48.383481 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:48.383491 | orchestrator | 2025-08-29 17:16:48.383502 | orchestrator | TASK [osism.commons.facts : Create custom facts directory] ********************* 2025-08-29 17:16:48.383513 | orchestrator | Friday 29 August 2025 17:16:21 +0000 (0:00:01.585) 0:07:12.151 ********* 2025-08-29 17:16:48.383523 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.383534 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.383546 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.383558 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.383570 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.383582 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.383593 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.383606 | orchestrator | 2025-08-29 17:16:48.383618 | orchestrator | TASK [osism.commons.facts : Copy fact files] *********************************** 2025-08-29 17:16:48.383630 | orchestrator | Friday 29 August 2025 17:16:22 +0000 (0:00:01.006) 0:07:13.158 ********* 2025-08-29 17:16:48.383642 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:48.383654 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:48.383666 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:16:48.383678 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:16:48.383690 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:16:48.383702 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:16:48.383714 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:16:48.383726 | orchestrator | 2025-08-29 17:16:48.383738 | orchestrator | TASK [osism.services.chrony : Check minimum and maximum number of servers] ***** 2025-08-29 17:16:48.383750 | orchestrator | Friday 29 August 2025 17:16:23 +0000 (0:00:00.647) 0:07:13.805 ********* 2025-08-29 17:16:48.383762 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:48.383774 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:48.383786 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:16:48.383798 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:16:48.383810 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:16:48.383821 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:16:48.383833 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:16:48.383845 | orchestrator | 2025-08-29 17:16:48.383857 | orchestrator | TASK [osism.services.chrony : Gather variables for each operating system] ****** 2025-08-29 17:16:48.383869 | orchestrator | Friday 29 August 2025 17:16:23 +0000 (0:00:00.429) 0:07:14.234 ********* 2025-08-29 17:16:48.383881 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.383892 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.383918 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.383930 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.383941 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.383952 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.383962 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.383973 | orchestrator | 2025-08-29 17:16:48.383984 | orchestrator | TASK [osism.services.chrony : Set chrony_conf_file variable to default value] *** 2025-08-29 17:16:48.384002 | orchestrator | Friday 29 August 2025 17:16:24 +0000 (0:00:00.552) 0:07:14.787 ********* 2025-08-29 17:16:48.384013 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.384024 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.384034 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.384045 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.384056 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.384066 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.384077 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.384087 | orchestrator | 2025-08-29 17:16:48.384098 | orchestrator | TASK [osism.services.chrony : Set chrony_key_file variable to default value] *** 2025-08-29 17:16:48.384109 | orchestrator | Friday 29 August 2025 17:16:24 +0000 (0:00:00.406) 0:07:15.193 ********* 2025-08-29 17:16:48.384120 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.384130 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.384141 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.384151 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.384162 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.384172 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.384183 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.384193 | orchestrator | 2025-08-29 17:16:48.384204 | orchestrator | TASK [osism.services.chrony : Populate service facts] ************************** 2025-08-29 17:16:48.384215 | orchestrator | Friday 29 August 2025 17:16:25 +0000 (0:00:00.425) 0:07:15.619 ********* 2025-08-29 17:16:48.384226 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.384236 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.384247 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.384258 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.384286 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.384297 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.384307 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.384318 | orchestrator | 2025-08-29 17:16:48.384329 | orchestrator | TASK [osism.services.chrony : Manage timesyncd service] ************************ 2025-08-29 17:16:48.384355 | orchestrator | Friday 29 August 2025 17:16:30 +0000 (0:00:05.436) 0:07:21.055 ********* 2025-08-29 17:16:48.384366 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:16:48.384377 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:16:48.384388 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:16:48.384399 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:16:48.384410 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:16:48.384421 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:16:48.384431 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:16:48.384457 | orchestrator | 2025-08-29 17:16:48.384468 | orchestrator | TASK [osism.services.chrony : Include distribution specific install tasks] ***** 2025-08-29 17:16:48.384479 | orchestrator | Friday 29 August 2025 17:16:31 +0000 (0:00:00.535) 0:07:21.590 ********* 2025-08-29 17:16:48.384504 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:16:48.384517 | orchestrator | 2025-08-29 17:16:48.384528 | orchestrator | TASK [osism.services.chrony : Install package] ********************************* 2025-08-29 17:16:48.384539 | orchestrator | Friday 29 August 2025 17:16:32 +0000 (0:00:01.022) 0:07:22.613 ********* 2025-08-29 17:16:48.384550 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.384561 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.384571 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.384582 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.384593 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.384604 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.384615 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.384625 | orchestrator | 2025-08-29 17:16:48.384637 | orchestrator | TASK [osism.services.chrony : Manage chrony service] *************************** 2025-08-29 17:16:48.384648 | orchestrator | Friday 29 August 2025 17:16:34 +0000 (0:00:01.908) 0:07:24.522 ********* 2025-08-29 17:16:48.384666 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.384677 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.384687 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.384698 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.384708 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.384719 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.384730 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.384740 | orchestrator | 2025-08-29 17:16:48.384751 | orchestrator | TASK [osism.services.chrony : Check if configuration file exists] ************** 2025-08-29 17:16:48.384762 | orchestrator | Friday 29 August 2025 17:16:35 +0000 (0:00:01.724) 0:07:26.247 ********* 2025-08-29 17:16:48.384773 | orchestrator | ok: [testbed-manager] 2025-08-29 17:16:48.384783 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:16:48.384794 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:16:48.384805 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:16:48.384815 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:16:48.384826 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:16:48.384837 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:16:48.384847 | orchestrator | 2025-08-29 17:16:48.384858 | orchestrator | TASK [osism.services.chrony : Copy configuration file] ************************* 2025-08-29 17:16:48.384869 | orchestrator | Friday 29 August 2025 17:16:37 +0000 (0:00:01.027) 0:07:27.274 ********* 2025-08-29 17:16:48.384880 | orchestrator | changed: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2) 2025-08-29 17:16:48.384894 | orchestrator | changed: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2) 2025-08-29 17:16:48.384905 | orchestrator | changed: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2) 2025-08-29 17:16:48.384916 | orchestrator | changed: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2) 2025-08-29 17:16:48.384927 | orchestrator | changed: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2) 2025-08-29 17:16:48.384938 | orchestrator | changed: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2) 2025-08-29 17:16:48.384949 | orchestrator | changed: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2) 2025-08-29 17:16:48.384960 | orchestrator | 2025-08-29 17:16:48.384971 | orchestrator | TASK [osism.services.lldpd : Include distribution specific install tasks] ****** 2025-08-29 17:16:48.384982 | orchestrator | Friday 29 August 2025 17:16:38 +0000 (0:00:01.747) 0:07:29.021 ********* 2025-08-29 17:16:48.384993 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:16:48.385004 | orchestrator | 2025-08-29 17:16:48.385015 | orchestrator | TASK [osism.services.lldpd : Install lldpd package] **************************** 2025-08-29 17:16:48.385026 | orchestrator | Friday 29 August 2025 17:16:39 +0000 (0:00:00.761) 0:07:29.783 ********* 2025-08-29 17:16:48.385037 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:16:48.385048 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:16:48.385059 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:16:48.385069 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:16:48.385080 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:16:48.385090 | orchestrator | changed: [testbed-manager] 2025-08-29 17:16:48.385101 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:16:48.385112 | orchestrator | 2025-08-29 17:16:48.385122 | orchestrator | TASK [osism.services.lldpd : Manage lldpd service] ***************************** 2025-08-29 17:16:48.385140 | orchestrator | Friday 29 August 2025 17:16:48 +0000 (0:00:08.844) 0:07:38.628 ********* 2025-08-29 17:17:04.410727 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:04.410845 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:04.410861 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:04.410873 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:04.410884 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:04.410896 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:04.410907 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:04.410919 | orchestrator | 2025-08-29 17:17:04.410932 | orchestrator | RUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] ********* 2025-08-29 17:17:04.410945 | orchestrator | Friday 29 August 2025 17:16:50 +0000 (0:00:01.814) 0:07:40.442 ********* 2025-08-29 17:17:04.410957 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:04.410968 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:04.411028 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:04.411041 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:04.411059 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:04.411077 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:04.411093 | orchestrator | 2025-08-29 17:17:04.411111 | orchestrator | RUNNING HANDLER [osism.services.chrony : Restart chrony service] *************** 2025-08-29 17:17:04.411130 | orchestrator | Friday 29 August 2025 17:16:51 +0000 (0:00:01.348) 0:07:41.791 ********* 2025-08-29 17:17:04.411150 | orchestrator | changed: [testbed-manager] 2025-08-29 17:17:04.411172 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:17:04.411192 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:17:04.411204 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:17:04.411215 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:17:04.411229 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:17:04.411241 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:17:04.411254 | orchestrator | 2025-08-29 17:17:04.411267 | orchestrator | PLAY [Apply bootstrap role part 2] ********************************************* 2025-08-29 17:17:04.411313 | orchestrator | 2025-08-29 17:17:04.411327 | orchestrator | TASK [Include hardening role] ************************************************** 2025-08-29 17:17:04.411340 | orchestrator | Friday 29 August 2025 17:16:53 +0000 (0:00:01.485) 0:07:43.277 ********* 2025-08-29 17:17:04.411353 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:17:04.411365 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:17:04.411378 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:17:04.411390 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:17:04.411403 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:17:04.411415 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:17:04.411428 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:17:04.411441 | orchestrator | 2025-08-29 17:17:04.411454 | orchestrator | PLAY [Apply bootstrap roles part 3] ******************************************** 2025-08-29 17:17:04.411466 | orchestrator | 2025-08-29 17:17:04.411478 | orchestrator | TASK [osism.services.journald : Copy configuration file] *********************** 2025-08-29 17:17:04.411491 | orchestrator | Friday 29 August 2025 17:16:53 +0000 (0:00:00.513) 0:07:43.790 ********* 2025-08-29 17:17:04.411503 | orchestrator | changed: [testbed-manager] 2025-08-29 17:17:04.411515 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:17:04.411527 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:17:04.411540 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:17:04.411552 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:17:04.411564 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:17:04.411576 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:17:04.411588 | orchestrator | 2025-08-29 17:17:04.411601 | orchestrator | TASK [osism.services.journald : Manage journald service] *********************** 2025-08-29 17:17:04.411612 | orchestrator | Friday 29 August 2025 17:16:54 +0000 (0:00:01.307) 0:07:45.098 ********* 2025-08-29 17:17:04.411623 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:04.411634 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:04.411645 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:04.411656 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:04.411667 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:04.411701 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:04.411712 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:04.411723 | orchestrator | 2025-08-29 17:17:04.411734 | orchestrator | TASK [Include auditd role] ***************************************************** 2025-08-29 17:17:04.411745 | orchestrator | Friday 29 August 2025 17:16:56 +0000 (0:00:01.389) 0:07:46.488 ********* 2025-08-29 17:17:04.411755 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:17:04.411766 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:17:04.411783 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:17:04.411794 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:17:04.411805 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:17:04.411816 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:17:04.411827 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:17:04.411837 | orchestrator | 2025-08-29 17:17:04.411848 | orchestrator | RUNNING HANDLER [osism.services.journald : Restart journald service] *********** 2025-08-29 17:17:04.411859 | orchestrator | Friday 29 August 2025 17:16:57 +0000 (0:00:00.989) 0:07:47.478 ********* 2025-08-29 17:17:04.411870 | orchestrator | changed: [testbed-manager] 2025-08-29 17:17:04.411881 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:17:04.411891 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:17:04.411902 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:17:04.411913 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:17:04.411924 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:17:04.411934 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:17:04.411945 | orchestrator | 2025-08-29 17:17:04.411956 | orchestrator | PLAY [Set state bootstrap] ***************************************************** 2025-08-29 17:17:04.411967 | orchestrator | 2025-08-29 17:17:04.411978 | orchestrator | TASK [Set osism.bootstrap.status fact] ***************************************** 2025-08-29 17:17:04.411989 | orchestrator | Friday 29 August 2025 17:16:58 +0000 (0:00:01.196) 0:07:48.675 ********* 2025-08-29 17:17:04.412000 | orchestrator | included: osism.commons.state for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:17:04.412013 | orchestrator | 2025-08-29 17:17:04.412023 | orchestrator | TASK [osism.commons.state : Create custom facts directory] ********************* 2025-08-29 17:17:04.412034 | orchestrator | Friday 29 August 2025 17:16:59 +0000 (0:00:00.959) 0:07:49.634 ********* 2025-08-29 17:17:04.412045 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:04.412056 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:04.412067 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:04.412078 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:04.412089 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:04.412100 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:04.412119 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:04.412136 | orchestrator | 2025-08-29 17:17:04.412178 | orchestrator | TASK [osism.commons.state : Write state into file] ***************************** 2025-08-29 17:17:04.412197 | orchestrator | Friday 29 August 2025 17:17:00 +0000 (0:00:00.814) 0:07:50.449 ********* 2025-08-29 17:17:04.412212 | orchestrator | changed: [testbed-manager] 2025-08-29 17:17:04.412223 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:17:04.412234 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:17:04.412245 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:17:04.412256 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:17:04.412266 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:17:04.412351 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:17:04.412363 | orchestrator | 2025-08-29 17:17:04.412374 | orchestrator | TASK [Set osism.bootstrap.timestamp fact] ************************************** 2025-08-29 17:17:04.412385 | orchestrator | Friday 29 August 2025 17:17:01 +0000 (0:00:01.130) 0:07:51.580 ********* 2025-08-29 17:17:04.412396 | orchestrator | included: osism.commons.state for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:17:04.412407 | orchestrator | 2025-08-29 17:17:04.412418 | orchestrator | TASK [osism.commons.state : Create custom facts directory] ********************* 2025-08-29 17:17:04.412439 | orchestrator | Friday 29 August 2025 17:17:02 +0000 (0:00:01.151) 0:07:52.731 ********* 2025-08-29 17:17:04.412450 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:04.412461 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:04.412471 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:04.412482 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:04.412493 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:04.412503 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:04.412514 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:04.412524 | orchestrator | 2025-08-29 17:17:04.412535 | orchestrator | TASK [osism.commons.state : Write state into file] ***************************** 2025-08-29 17:17:04.412547 | orchestrator | Friday 29 August 2025 17:17:03 +0000 (0:00:00.832) 0:07:53.564 ********* 2025-08-29 17:17:04.412557 | orchestrator | changed: [testbed-manager] 2025-08-29 17:17:04.412568 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:17:04.412579 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:17:04.412590 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:17:04.412600 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:17:04.412611 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:17:04.412621 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:17:04.412632 | orchestrator | 2025-08-29 17:17:04.412643 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:17:04.412655 | orchestrator | testbed-manager : ok=162  changed=38  unreachable=0 failed=0 skipped=41  rescued=0 ignored=0 2025-08-29 17:17:04.412667 | orchestrator | testbed-node-0 : ok=170  changed=66  unreachable=0 failed=0 skipped=37  rescued=0 ignored=0 2025-08-29 17:17:04.412678 | orchestrator | testbed-node-1 : ok=170  changed=66  unreachable=0 failed=0 skipped=36  rescued=0 ignored=0 2025-08-29 17:17:04.412689 | orchestrator | testbed-node-2 : ok=170  changed=66  unreachable=0 failed=0 skipped=36  rescued=0 ignored=0 2025-08-29 17:17:04.412700 | orchestrator | testbed-node-3 : ok=169  changed=63  unreachable=0 failed=0 skipped=36  rescued=0 ignored=0 2025-08-29 17:17:04.412711 | orchestrator | testbed-node-4 : ok=169  changed=63  unreachable=0 failed=0 skipped=36  rescued=0 ignored=0 2025-08-29 17:17:04.412722 | orchestrator | testbed-node-5 : ok=169  changed=63  unreachable=0 failed=0 skipped=36  rescued=0 ignored=0 2025-08-29 17:17:04.412733 | orchestrator | 2025-08-29 17:17:04.412743 | orchestrator | 2025-08-29 17:17:04.412754 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:17:04.412765 | orchestrator | Friday 29 August 2025 17:17:04 +0000 (0:00:01.083) 0:07:54.647 ********* 2025-08-29 17:17:04.412776 | orchestrator | =============================================================================== 2025-08-29 17:17:04.412787 | orchestrator | osism.commons.packages : Install required packages --------------------- 77.45s 2025-08-29 17:17:04.412797 | orchestrator | osism.commons.packages : Download required packages -------------------- 38.33s 2025-08-29 17:17:04.412808 | orchestrator | osism.commons.cleanup : Cleanup installed packages --------------------- 35.90s 2025-08-29 17:17:04.412819 | orchestrator | osism.commons.repository : Update package cache ------------------------ 17.80s 2025-08-29 17:17:04.412829 | orchestrator | osism.services.docker : Install docker package ------------------------- 11.73s 2025-08-29 17:17:04.412840 | orchestrator | osism.commons.packages : Remove dependencies that are no longer required -- 10.86s 2025-08-29 17:17:04.412852 | orchestrator | osism.commons.systohc : Install util-linux-extra package --------------- 10.84s 2025-08-29 17:17:04.412862 | orchestrator | osism.services.docker : Install containerd package --------------------- 10.01s 2025-08-29 17:17:04.412873 | orchestrator | osism.services.docker : Install docker-cli package ---------------------- 9.56s 2025-08-29 17:17:04.412890 | orchestrator | osism.services.smartd : Install smartmontools package ------------------- 9.03s 2025-08-29 17:17:04.412901 | orchestrator | osism.services.lldpd : Install lldpd package ---------------------------- 8.84s 2025-08-29 17:17:04.412912 | orchestrator | osism.commons.cleanup : Remove cloudinit package ------------------------ 8.32s 2025-08-29 17:17:04.412922 | orchestrator | osism.services.rng : Install rng package -------------------------------- 8.25s 2025-08-29 17:17:04.412933 | orchestrator | osism.commons.cleanup : Uninstall unattended-upgrades package ----------- 7.74s 2025-08-29 17:17:04.412952 | orchestrator | osism.services.docker : Add repository ---------------------------------- 7.56s 2025-08-29 17:17:04.828642 | orchestrator | osism.commons.docker_compose : Install docker-compose-plugin package ---- 6.82s 2025-08-29 17:17:04.828759 | orchestrator | osism.services.docker : Install apt-transport-https package ------------- 6.31s 2025-08-29 17:17:04.828775 | orchestrator | osism.commons.cleanup : Remove dependencies that are no longer required --- 5.86s 2025-08-29 17:17:04.828786 | orchestrator | osism.commons.sysctl : Set sysctl parameters on rabbitmq ---------------- 5.67s 2025-08-29 17:17:04.828798 | orchestrator | osism.commons.cleanup : Populate service facts -------------------------- 5.64s 2025-08-29 17:17:05.082463 | orchestrator | + [[ -e /etc/redhat-release ]] 2025-08-29 17:17:05.082561 | orchestrator | + osism apply network 2025-08-29 17:17:17.405218 | orchestrator | 2025-08-29 17:17:17 | INFO  | Task 66543e91-abb2-4225-bdfc-2287c34c6e4d (network) was prepared for execution. 2025-08-29 17:17:17.405373 | orchestrator | 2025-08-29 17:17:17 | INFO  | It takes a moment until task 66543e91-abb2-4225-bdfc-2287c34c6e4d (network) has been started and output is visible here. 2025-08-29 17:17:45.273185 | orchestrator | 2025-08-29 17:17:45.273346 | orchestrator | PLAY [Apply role network] ****************************************************** 2025-08-29 17:17:45.273367 | orchestrator | 2025-08-29 17:17:45.273379 | orchestrator | TASK [osism.commons.network : Gather variables for each operating system] ****** 2025-08-29 17:17:45.273392 | orchestrator | Friday 29 August 2025 17:17:21 +0000 (0:00:00.270) 0:00:00.270 ********* 2025-08-29 17:17:45.273403 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.273416 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:45.273427 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:45.273438 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:45.273448 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:45.273459 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:45.273470 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:45.273481 | orchestrator | 2025-08-29 17:17:45.273492 | orchestrator | TASK [osism.commons.network : Include type specific tasks] ********************* 2025-08-29 17:17:45.273503 | orchestrator | Friday 29 August 2025 17:17:22 +0000 (0:00:00.660) 0:00:00.931 ********* 2025-08-29 17:17:45.273516 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:17:45.273529 | orchestrator | 2025-08-29 17:17:45.273541 | orchestrator | TASK [osism.commons.network : Install required packages] *********************** 2025-08-29 17:17:45.273552 | orchestrator | Friday 29 August 2025 17:17:23 +0000 (0:00:01.151) 0:00:02.083 ********* 2025-08-29 17:17:45.273563 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.273574 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:45.273585 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:45.273596 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:45.273606 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:45.273617 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:45.273628 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:45.273639 | orchestrator | 2025-08-29 17:17:45.273650 | orchestrator | TASK [osism.commons.network : Remove ifupdown package] ************************* 2025-08-29 17:17:45.273661 | orchestrator | Friday 29 August 2025 17:17:25 +0000 (0:00:02.038) 0:00:04.122 ********* 2025-08-29 17:17:45.273672 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.273683 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:45.273694 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:45.273729 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:45.273743 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:45.273755 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:45.273767 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:45.273779 | orchestrator | 2025-08-29 17:17:45.273791 | orchestrator | TASK [osism.commons.network : Create required directories] ********************* 2025-08-29 17:17:45.273804 | orchestrator | Friday 29 August 2025 17:17:27 +0000 (0:00:01.847) 0:00:05.970 ********* 2025-08-29 17:17:45.273831 | orchestrator | ok: [testbed-manager] => (item=/etc/netplan) 2025-08-29 17:17:45.273844 | orchestrator | ok: [testbed-node-0] => (item=/etc/netplan) 2025-08-29 17:17:45.273856 | orchestrator | ok: [testbed-node-1] => (item=/etc/netplan) 2025-08-29 17:17:45.273869 | orchestrator | ok: [testbed-node-2] => (item=/etc/netplan) 2025-08-29 17:17:45.273881 | orchestrator | ok: [testbed-node-3] => (item=/etc/netplan) 2025-08-29 17:17:45.273894 | orchestrator | ok: [testbed-node-4] => (item=/etc/netplan) 2025-08-29 17:17:45.273906 | orchestrator | ok: [testbed-node-5] => (item=/etc/netplan) 2025-08-29 17:17:45.273918 | orchestrator | 2025-08-29 17:17:45.273930 | orchestrator | TASK [osism.commons.network : Prepare netplan configuration template] ********** 2025-08-29 17:17:45.273943 | orchestrator | Friday 29 August 2025 17:17:28 +0000 (0:00:00.945) 0:00:06.915 ********* 2025-08-29 17:17:45.273956 | orchestrator | ok: [testbed-node-4 -> localhost] 2025-08-29 17:17:45.273969 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:17:45.273981 | orchestrator | ok: [testbed-node-2 -> localhost] 2025-08-29 17:17:45.273993 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:17:45.274005 | orchestrator | ok: [testbed-node-1 -> localhost] 2025-08-29 17:17:45.274070 | orchestrator | ok: [testbed-node-5 -> localhost] 2025-08-29 17:17:45.274084 | orchestrator | ok: [testbed-node-3 -> localhost] 2025-08-29 17:17:45.274096 | orchestrator | 2025-08-29 17:17:45.274107 | orchestrator | TASK [osism.commons.network : Copy netplan configuration] ********************** 2025-08-29 17:17:45.274118 | orchestrator | Friday 29 August 2025 17:17:31 +0000 (0:00:03.190) 0:00:10.105 ********* 2025-08-29 17:17:45.274129 | orchestrator | changed: [testbed-manager] 2025-08-29 17:17:45.274140 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:17:45.274151 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:17:45.274162 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:17:45.274172 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:17:45.274183 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:17:45.274194 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:17:45.274204 | orchestrator | 2025-08-29 17:17:45.274215 | orchestrator | TASK [osism.commons.network : Remove netplan configuration template] *********** 2025-08-29 17:17:45.274226 | orchestrator | Friday 29 August 2025 17:17:32 +0000 (0:00:01.569) 0:00:11.675 ********* 2025-08-29 17:17:45.274237 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:17:45.274247 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:17:45.274258 | orchestrator | ok: [testbed-node-1 -> localhost] 2025-08-29 17:17:45.274269 | orchestrator | ok: [testbed-node-2 -> localhost] 2025-08-29 17:17:45.274279 | orchestrator | ok: [testbed-node-4 -> localhost] 2025-08-29 17:17:45.274290 | orchestrator | ok: [testbed-node-3 -> localhost] 2025-08-29 17:17:45.274320 | orchestrator | ok: [testbed-node-5 -> localhost] 2025-08-29 17:17:45.274331 | orchestrator | 2025-08-29 17:17:45.274341 | orchestrator | TASK [osism.commons.network : Check if path for interface file exists] ********* 2025-08-29 17:17:45.274352 | orchestrator | Friday 29 August 2025 17:17:34 +0000 (0:00:01.843) 0:00:13.519 ********* 2025-08-29 17:17:45.274363 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.274374 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:45.274384 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:45.274395 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:45.274406 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:45.274416 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:45.274427 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:45.274438 | orchestrator | 2025-08-29 17:17:45.274458 | orchestrator | TASK [osism.commons.network : Copy interfaces file] **************************** 2025-08-29 17:17:45.274488 | orchestrator | Friday 29 August 2025 17:17:35 +0000 (0:00:01.066) 0:00:14.585 ********* 2025-08-29 17:17:45.274499 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:17:45.274510 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:17:45.274521 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:17:45.274532 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:17:45.274542 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:17:45.274553 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:17:45.274564 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:17:45.274574 | orchestrator | 2025-08-29 17:17:45.274585 | orchestrator | TASK [osism.commons.network : Install package networkd-dispatcher] ************* 2025-08-29 17:17:45.274596 | orchestrator | Friday 29 August 2025 17:17:36 +0000 (0:00:00.639) 0:00:15.224 ********* 2025-08-29 17:17:45.274607 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.274618 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:45.274628 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:45.274639 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:45.274650 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:45.274661 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:45.274671 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:45.274682 | orchestrator | 2025-08-29 17:17:45.274693 | orchestrator | TASK [osism.commons.network : Copy dispatcher scripts] ************************* 2025-08-29 17:17:45.274704 | orchestrator | Friday 29 August 2025 17:17:38 +0000 (0:00:02.159) 0:00:17.383 ********* 2025-08-29 17:17:45.274714 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:17:45.274725 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:17:45.274736 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:17:45.274747 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:17:45.274757 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:17:45.274768 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:17:45.274779 | orchestrator | changed: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'}) 2025-08-29 17:17:45.274792 | orchestrator | 2025-08-29 17:17:45.274802 | orchestrator | TASK [osism.commons.network : Manage service networkd-dispatcher] ************** 2025-08-29 17:17:45.274813 | orchestrator | Friday 29 August 2025 17:17:39 +0000 (0:00:00.865) 0:00:18.249 ********* 2025-08-29 17:17:45.274824 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.274834 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:17:45.274845 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:17:45.274856 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:17:45.274866 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:17:45.274877 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:17:45.274888 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:17:45.274899 | orchestrator | 2025-08-29 17:17:45.274910 | orchestrator | TASK [osism.commons.network : Include cleanup tasks] *************************** 2025-08-29 17:17:45.274920 | orchestrator | Friday 29 August 2025 17:17:41 +0000 (0:00:01.606) 0:00:19.855 ********* 2025-08-29 17:17:45.274937 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:17:45.274950 | orchestrator | 2025-08-29 17:17:45.274961 | orchestrator | TASK [osism.commons.network : List existing configuration files] *************** 2025-08-29 17:17:45.274972 | orchestrator | Friday 29 August 2025 17:17:42 +0000 (0:00:01.245) 0:00:21.101 ********* 2025-08-29 17:17:45.274983 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.274993 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:45.275004 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:45.275015 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:45.275026 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:45.275036 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:45.275047 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:45.275057 | orchestrator | 2025-08-29 17:17:45.275076 | orchestrator | TASK [osism.commons.network : Set network_configured_files fact] *************** 2025-08-29 17:17:45.275088 | orchestrator | Friday 29 August 2025 17:17:43 +0000 (0:00:00.930) 0:00:22.032 ********* 2025-08-29 17:17:45.275098 | orchestrator | ok: [testbed-manager] 2025-08-29 17:17:45.275109 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:17:45.275120 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:17:45.275130 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:17:45.275141 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:17:45.275152 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:17:45.275162 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:17:45.275173 | orchestrator | 2025-08-29 17:17:45.275184 | orchestrator | TASK [osism.commons.network : Remove unused configuration files] *************** 2025-08-29 17:17:45.275195 | orchestrator | Friday 29 August 2025 17:17:44 +0000 (0:00:00.789) 0:00:22.821 ********* 2025-08-29 17:17:45.275205 | orchestrator | skipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml)  2025-08-29 17:17:45.275216 | orchestrator | skipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml)  2025-08-29 17:17:45.275227 | orchestrator | skipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml)  2025-08-29 17:17:45.275238 | orchestrator | skipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml)  2025-08-29 17:17:45.275249 | orchestrator | changed: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml) 2025-08-29 17:17:45.275260 | orchestrator | skipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml)  2025-08-29 17:17:45.275270 | orchestrator | changed: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml) 2025-08-29 17:17:45.275281 | orchestrator | skipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml)  2025-08-29 17:17:45.275291 | orchestrator | changed: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml) 2025-08-29 17:17:45.275336 | orchestrator | skipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml)  2025-08-29 17:17:45.275347 | orchestrator | changed: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml) 2025-08-29 17:17:45.275357 | orchestrator | changed: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml) 2025-08-29 17:17:45.275368 | orchestrator | changed: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml) 2025-08-29 17:17:45.275379 | orchestrator | changed: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml) 2025-08-29 17:17:45.275390 | orchestrator | 2025-08-29 17:17:45.275408 | orchestrator | TASK [osism.commons.network : Include dummy interfaces] ************************ 2025-08-29 17:18:00.453764 | orchestrator | Friday 29 August 2025 17:17:45 +0000 (0:00:01.164) 0:00:23.986 ********* 2025-08-29 17:18:00.453878 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:18:00.453894 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:18:00.453906 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:18:00.453918 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:18:00.453929 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:18:00.453940 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:18:00.453951 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:18:00.453962 | orchestrator | 2025-08-29 17:18:00.453974 | orchestrator | TASK [osism.commons.network : Include vxlan interfaces] ************************ 2025-08-29 17:18:00.453986 | orchestrator | Friday 29 August 2025 17:17:45 +0000 (0:00:00.645) 0:00:24.632 ********* 2025-08-29 17:18:00.453999 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-manager, testbed-node-1, testbed-node-0, testbed-node-2, testbed-node-3, testbed-node-5, testbed-node-4 2025-08-29 17:18:00.454012 | orchestrator | 2025-08-29 17:18:00.454086 | orchestrator | TASK [osism.commons.network : Create systemd networkd netdev files] ************ 2025-08-29 17:18:00.454097 | orchestrator | Friday 29 August 2025 17:17:50 +0000 (0:00:04.283) 0:00:28.916 ********* 2025-08-29 17:18:00.454110 | orchestrator | changed: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454148 | orchestrator | changed: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454161 | orchestrator | changed: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454186 | orchestrator | changed: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454198 | orchestrator | changed: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454209 | orchestrator | changed: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454220 | orchestrator | changed: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454231 | orchestrator | changed: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454242 | orchestrator | changed: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454261 | orchestrator | changed: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454272 | orchestrator | changed: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454342 | orchestrator | changed: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454357 | orchestrator | changed: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454369 | orchestrator | changed: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454382 | orchestrator | 2025-08-29 17:18:00.454395 | orchestrator | TASK [osism.commons.network : Create systemd networkd network files] *********** 2025-08-29 17:18:00.454417 | orchestrator | Friday 29 August 2025 17:17:55 +0000 (0:00:05.082) 0:00:33.998 ********* 2025-08-29 17:18:00.454429 | orchestrator | changed: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454442 | orchestrator | changed: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454455 | orchestrator | changed: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454468 | orchestrator | changed: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454481 | orchestrator | changed: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454493 | orchestrator | changed: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454506 | orchestrator | changed: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454519 | orchestrator | changed: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454539 | orchestrator | changed: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}}) 2025-08-29 17:18:00.454552 | orchestrator | changed: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454565 | orchestrator | changed: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454577 | orchestrator | changed: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:00.454603 | orchestrator | changed: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:06.416605 | orchestrator | changed: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}}) 2025-08-29 17:18:06.416720 | orchestrator | 2025-08-29 17:18:06.416730 | orchestrator | TASK [osism.commons.network : Include networkd cleanup tasks] ****************** 2025-08-29 17:18:06.416739 | orchestrator | Friday 29 August 2025 17:18:00 +0000 (0:00:05.165) 0:00:39.163 ********* 2025-08-29 17:18:06.416747 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:18:06.416754 | orchestrator | 2025-08-29 17:18:06.416761 | orchestrator | TASK [osism.commons.network : List existing configuration files] *************** 2025-08-29 17:18:06.416767 | orchestrator | Friday 29 August 2025 17:18:01 +0000 (0:00:01.272) 0:00:40.436 ********* 2025-08-29 17:18:06.416774 | orchestrator | ok: [testbed-manager] 2025-08-29 17:18:06.416781 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:18:06.416787 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:18:06.416793 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:18:06.416799 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:18:06.416805 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:18:06.416811 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:18:06.416817 | orchestrator | 2025-08-29 17:18:06.416824 | orchestrator | TASK [osism.commons.network : Remove unused configuration files] *************** 2025-08-29 17:18:06.416830 | orchestrator | Friday 29 August 2025 17:18:02 +0000 (0:00:01.182) 0:00:41.619 ********* 2025-08-29 17:18:06.416836 | orchestrator | skipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network)  2025-08-29 17:18:06.416843 | orchestrator | skipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network)  2025-08-29 17:18:06.416849 | orchestrator | skipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev)  2025-08-29 17:18:06.416856 | orchestrator | skipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev)  2025-08-29 17:18:06.416862 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:18:06.416880 | orchestrator | skipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network)  2025-08-29 17:18:06.416887 | orchestrator | skipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network)  2025-08-29 17:18:06.416893 | orchestrator | skipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev)  2025-08-29 17:18:06.416899 | orchestrator | skipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev)  2025-08-29 17:18:06.416905 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:18:06.416911 | orchestrator | skipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network)  2025-08-29 17:18:06.416918 | orchestrator | skipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network)  2025-08-29 17:18:06.416924 | orchestrator | skipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev)  2025-08-29 17:18:06.416930 | orchestrator | skipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev)  2025-08-29 17:18:06.416936 | orchestrator | skipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network)  2025-08-29 17:18:06.416942 | orchestrator | skipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network)  2025-08-29 17:18:06.416948 | orchestrator | skipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev)  2025-08-29 17:18:06.416955 | orchestrator | skipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev)  2025-08-29 17:18:06.416961 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:18:06.416967 | orchestrator | skipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network)  2025-08-29 17:18:06.416973 | orchestrator | skipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network)  2025-08-29 17:18:06.416979 | orchestrator | skipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev)  2025-08-29 17:18:06.416985 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:18:06.416997 | orchestrator | skipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev)  2025-08-29 17:18:06.417003 | orchestrator | skipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network)  2025-08-29 17:18:06.417009 | orchestrator | skipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network)  2025-08-29 17:18:06.417016 | orchestrator | skipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev)  2025-08-29 17:18:06.417022 | orchestrator | skipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev)  2025-08-29 17:18:06.417028 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:18:06.417034 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:18:06.417040 | orchestrator | skipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network)  2025-08-29 17:18:06.417046 | orchestrator | skipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network)  2025-08-29 17:18:06.417052 | orchestrator | skipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev)  2025-08-29 17:18:06.417058 | orchestrator | skipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev)  2025-08-29 17:18:06.417064 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:18:06.417070 | orchestrator | 2025-08-29 17:18:06.417077 | orchestrator | RUNNING HANDLER [osism.commons.network : Reload systemd-networkd] ************** 2025-08-29 17:18:06.417095 | orchestrator | Friday 29 August 2025 17:18:04 +0000 (0:00:01.926) 0:00:43.545 ********* 2025-08-29 17:18:06.417101 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:18:06.417107 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:18:06.417114 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:18:06.417120 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:18:06.417126 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:18:06.417132 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:18:06.417138 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:18:06.417144 | orchestrator | 2025-08-29 17:18:06.417150 | orchestrator | RUNNING HANDLER [osism.commons.network : Netplan configuration changed] ******** 2025-08-29 17:18:06.417156 | orchestrator | Friday 29 August 2025 17:18:05 +0000 (0:00:00.597) 0:00:44.142 ********* 2025-08-29 17:18:06.417163 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:18:06.417170 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:18:06.417177 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:18:06.417184 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:18:06.417191 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:18:06.417198 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:18:06.417205 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:18:06.417211 | orchestrator | 2025-08-29 17:18:06.417218 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:18:06.417226 | orchestrator | testbed-manager : ok=21  changed=5  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:18:06.417235 | orchestrator | testbed-node-0 : ok=20  changed=5  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:18:06.417242 | orchestrator | testbed-node-1 : ok=20  changed=5  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:18:06.417249 | orchestrator | testbed-node-2 : ok=20  changed=5  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:18:06.417256 | orchestrator | testbed-node-3 : ok=20  changed=5  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:18:06.417266 | orchestrator | testbed-node-4 : ok=20  changed=5  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:18:06.417273 | orchestrator | testbed-node-5 : ok=20  changed=5  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:18:06.417285 | orchestrator | 2025-08-29 17:18:06.417293 | orchestrator | 2025-08-29 17:18:06.417300 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:18:06.417325 | orchestrator | Friday 29 August 2025 17:18:06 +0000 (0:00:00.642) 0:00:44.785 ********* 2025-08-29 17:18:06.417331 | orchestrator | =============================================================================== 2025-08-29 17:18:06.417337 | orchestrator | osism.commons.network : Create systemd networkd network files ----------- 5.17s 2025-08-29 17:18:06.417343 | orchestrator | osism.commons.network : Create systemd networkd netdev files ------------ 5.08s 2025-08-29 17:18:06.417349 | orchestrator | osism.commons.network : Include vxlan interfaces ------------------------ 4.28s 2025-08-29 17:18:06.417355 | orchestrator | osism.commons.network : Prepare netplan configuration template ---------- 3.19s 2025-08-29 17:18:06.417361 | orchestrator | osism.commons.network : Install package networkd-dispatcher ------------- 2.16s 2025-08-29 17:18:06.417367 | orchestrator | osism.commons.network : Install required packages ----------------------- 2.04s 2025-08-29 17:18:06.417374 | orchestrator | osism.commons.network : Remove unused configuration files --------------- 1.93s 2025-08-29 17:18:06.417380 | orchestrator | osism.commons.network : Remove ifupdown package ------------------------- 1.85s 2025-08-29 17:18:06.417386 | orchestrator | osism.commons.network : Remove netplan configuration template ----------- 1.84s 2025-08-29 17:18:06.417392 | orchestrator | osism.commons.network : Manage service networkd-dispatcher -------------- 1.61s 2025-08-29 17:18:06.417398 | orchestrator | osism.commons.network : Copy netplan configuration ---------------------- 1.57s 2025-08-29 17:18:06.417404 | orchestrator | osism.commons.network : Include networkd cleanup tasks ------------------ 1.27s 2025-08-29 17:18:06.417410 | orchestrator | osism.commons.network : Include cleanup tasks --------------------------- 1.25s 2025-08-29 17:18:06.417416 | orchestrator | osism.commons.network : List existing configuration files --------------- 1.18s 2025-08-29 17:18:06.417422 | orchestrator | osism.commons.network : Remove unused configuration files --------------- 1.16s 2025-08-29 17:18:06.417428 | orchestrator | osism.commons.network : Include type specific tasks --------------------- 1.15s 2025-08-29 17:18:06.417434 | orchestrator | osism.commons.network : Check if path for interface file exists --------- 1.07s 2025-08-29 17:18:06.417441 | orchestrator | osism.commons.network : Create required directories --------------------- 0.95s 2025-08-29 17:18:06.417447 | orchestrator | osism.commons.network : List existing configuration files --------------- 0.93s 2025-08-29 17:18:06.417453 | orchestrator | osism.commons.network : Copy dispatcher scripts ------------------------- 0.87s 2025-08-29 17:18:06.699842 | orchestrator | + osism apply wireguard 2025-08-29 17:18:18.621597 | orchestrator | 2025-08-29 17:18:18 | INFO  | Task 349d0965-05a7-4db1-a6d6-71b950fe35a8 (wireguard) was prepared for execution. 2025-08-29 17:18:18.621715 | orchestrator | 2025-08-29 17:18:18 | INFO  | It takes a moment until task 349d0965-05a7-4db1-a6d6-71b950fe35a8 (wireguard) has been started and output is visible here. 2025-08-29 17:18:38.284217 | orchestrator | 2025-08-29 17:18:38.284434 | orchestrator | PLAY [Apply role wireguard] **************************************************** 2025-08-29 17:18:38.284466 | orchestrator | 2025-08-29 17:18:38.284486 | orchestrator | TASK [osism.services.wireguard : Install iptables package] ********************* 2025-08-29 17:18:38.284504 | orchestrator | Friday 29 August 2025 17:18:22 +0000 (0:00:00.234) 0:00:00.234 ********* 2025-08-29 17:18:38.284522 | orchestrator | ok: [testbed-manager] 2025-08-29 17:18:38.284544 | orchestrator | 2025-08-29 17:18:38.284563 | orchestrator | TASK [osism.services.wireguard : Install wireguard package] ******************** 2025-08-29 17:18:38.284580 | orchestrator | Friday 29 August 2025 17:18:24 +0000 (0:00:01.552) 0:00:01.786 ********* 2025-08-29 17:18:38.284599 | orchestrator | changed: [testbed-manager] 2025-08-29 17:18:38.284619 | orchestrator | 2025-08-29 17:18:38.284638 | orchestrator | TASK [osism.services.wireguard : Create public and private key - server] ******* 2025-08-29 17:18:38.284656 | orchestrator | Friday 29 August 2025 17:18:30 +0000 (0:00:06.563) 0:00:08.350 ********* 2025-08-29 17:18:38.284709 | orchestrator | changed: [testbed-manager] 2025-08-29 17:18:38.284729 | orchestrator | 2025-08-29 17:18:38.284742 | orchestrator | TASK [osism.services.wireguard : Create preshared key] ************************* 2025-08-29 17:18:38.284755 | orchestrator | Friday 29 August 2025 17:18:31 +0000 (0:00:00.538) 0:00:08.888 ********* 2025-08-29 17:18:38.284767 | orchestrator | changed: [testbed-manager] 2025-08-29 17:18:38.284779 | orchestrator | 2025-08-29 17:18:38.284792 | orchestrator | TASK [osism.services.wireguard : Get preshared key] **************************** 2025-08-29 17:18:38.284804 | orchestrator | Friday 29 August 2025 17:18:31 +0000 (0:00:00.422) 0:00:09.310 ********* 2025-08-29 17:18:38.284815 | orchestrator | ok: [testbed-manager] 2025-08-29 17:18:38.284828 | orchestrator | 2025-08-29 17:18:38.284839 | orchestrator | TASK [osism.services.wireguard : Get public key - server] ********************** 2025-08-29 17:18:38.284852 | orchestrator | Friday 29 August 2025 17:18:32 +0000 (0:00:00.546) 0:00:09.857 ********* 2025-08-29 17:18:38.284863 | orchestrator | ok: [testbed-manager] 2025-08-29 17:18:38.284875 | orchestrator | 2025-08-29 17:18:38.284887 | orchestrator | TASK [osism.services.wireguard : Get private key - server] ********************* 2025-08-29 17:18:38.284899 | orchestrator | Friday 29 August 2025 17:18:32 +0000 (0:00:00.505) 0:00:10.363 ********* 2025-08-29 17:18:38.284911 | orchestrator | ok: [testbed-manager] 2025-08-29 17:18:38.284923 | orchestrator | 2025-08-29 17:18:38.284939 | orchestrator | TASK [osism.services.wireguard : Copy wg0.conf configuration file] ************* 2025-08-29 17:18:38.284975 | orchestrator | Friday 29 August 2025 17:18:33 +0000 (0:00:00.405) 0:00:10.768 ********* 2025-08-29 17:18:38.284996 | orchestrator | changed: [testbed-manager] 2025-08-29 17:18:38.285016 | orchestrator | 2025-08-29 17:18:38.285035 | orchestrator | TASK [osism.services.wireguard : Copy client configuration files] ************** 2025-08-29 17:18:38.285057 | orchestrator | Friday 29 August 2025 17:18:34 +0000 (0:00:01.270) 0:00:12.038 ********* 2025-08-29 17:18:38.285077 | orchestrator | changed: [testbed-manager] => (item=None) 2025-08-29 17:18:38.285092 | orchestrator | changed: [testbed-manager] 2025-08-29 17:18:38.285103 | orchestrator | 2025-08-29 17:18:38.285113 | orchestrator | TASK [osism.services.wireguard : Manage wg-quick@wg0.service service] ********** 2025-08-29 17:18:38.285124 | orchestrator | Friday 29 August 2025 17:18:35 +0000 (0:00:00.924) 0:00:12.963 ********* 2025-08-29 17:18:38.285135 | orchestrator | changed: [testbed-manager] 2025-08-29 17:18:38.285152 | orchestrator | 2025-08-29 17:18:38.285169 | orchestrator | RUNNING HANDLER [osism.services.wireguard : Restart wg0 service] *************** 2025-08-29 17:18:38.285187 | orchestrator | Friday 29 August 2025 17:18:37 +0000 (0:00:01.617) 0:00:14.580 ********* 2025-08-29 17:18:38.285204 | orchestrator | changed: [testbed-manager] 2025-08-29 17:18:38.285221 | orchestrator | 2025-08-29 17:18:38.285238 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:18:38.285258 | orchestrator | testbed-manager : ok=11  changed=7  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:18:38.285278 | orchestrator | 2025-08-29 17:18:38.285297 | orchestrator | 2025-08-29 17:18:38.285309 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:18:38.285354 | orchestrator | Friday 29 August 2025 17:18:37 +0000 (0:00:00.935) 0:00:15.516 ********* 2025-08-29 17:18:38.285366 | orchestrator | =============================================================================== 2025-08-29 17:18:38.285377 | orchestrator | osism.services.wireguard : Install wireguard package -------------------- 6.56s 2025-08-29 17:18:38.285388 | orchestrator | osism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.62s 2025-08-29 17:18:38.285398 | orchestrator | osism.services.wireguard : Install iptables package --------------------- 1.55s 2025-08-29 17:18:38.285409 | orchestrator | osism.services.wireguard : Copy wg0.conf configuration file ------------- 1.27s 2025-08-29 17:18:38.285420 | orchestrator | osism.services.wireguard : Restart wg0 service -------------------------- 0.94s 2025-08-29 17:18:38.285430 | orchestrator | osism.services.wireguard : Copy client configuration files -------------- 0.92s 2025-08-29 17:18:38.285441 | orchestrator | osism.services.wireguard : Get preshared key ---------------------------- 0.55s 2025-08-29 17:18:38.285464 | orchestrator | osism.services.wireguard : Create public and private key - server ------- 0.54s 2025-08-29 17:18:38.285474 | orchestrator | osism.services.wireguard : Get public key - server ---------------------- 0.51s 2025-08-29 17:18:38.285485 | orchestrator | osism.services.wireguard : Create preshared key ------------------------- 0.42s 2025-08-29 17:18:38.285496 | orchestrator | osism.services.wireguard : Get private key - server --------------------- 0.41s 2025-08-29 17:18:38.588925 | orchestrator | + sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh 2025-08-29 17:18:38.630310 | orchestrator | % Total % Received % Xferd Average Speed Time Time Time Current 2025-08-29 17:18:38.630435 | orchestrator | Dload Upload Total Spent Left Speed 2025-08-29 17:18:38.707124 | orchestrator | 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 15 100 15 0 0 200 0 --:--:-- --:--:-- --:--:-- 202 2025-08-29 17:18:38.725075 | orchestrator | + osism apply --environment custom workarounds 2025-08-29 17:18:40.471567 | orchestrator | 2025-08-29 17:18:40 | INFO  | Trying to run play workarounds in environment custom 2025-08-29 17:18:50.612894 | orchestrator | 2025-08-29 17:18:50 | INFO  | Task 2210944a-071b-488e-9a13-17a23c8697ae (workarounds) was prepared for execution. 2025-08-29 17:18:50.613005 | orchestrator | 2025-08-29 17:18:50 | INFO  | It takes a moment until task 2210944a-071b-488e-9a13-17a23c8697ae (workarounds) has been started and output is visible here. 2025-08-29 17:19:15.776519 | orchestrator | 2025-08-29 17:19:15.776634 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:19:15.776652 | orchestrator | 2025-08-29 17:19:15.776664 | orchestrator | TASK [Group hosts based on virtualization_role] ******************************** 2025-08-29 17:19:15.776676 | orchestrator | Friday 29 August 2025 17:18:54 +0000 (0:00:00.156) 0:00:00.156 ********* 2025-08-29 17:19:15.776687 | orchestrator | changed: [testbed-manager] => (item=virtualization_role_guest) 2025-08-29 17:19:15.776698 | orchestrator | changed: [testbed-node-3] => (item=virtualization_role_guest) 2025-08-29 17:19:15.776709 | orchestrator | changed: [testbed-node-4] => (item=virtualization_role_guest) 2025-08-29 17:19:15.776720 | orchestrator | changed: [testbed-node-5] => (item=virtualization_role_guest) 2025-08-29 17:19:15.776731 | orchestrator | changed: [testbed-node-0] => (item=virtualization_role_guest) 2025-08-29 17:19:15.776742 | orchestrator | changed: [testbed-node-1] => (item=virtualization_role_guest) 2025-08-29 17:19:15.776753 | orchestrator | changed: [testbed-node-2] => (item=virtualization_role_guest) 2025-08-29 17:19:15.776763 | orchestrator | 2025-08-29 17:19:15.776774 | orchestrator | PLAY [Apply netplan configuration on the manager node] ************************* 2025-08-29 17:19:15.776785 | orchestrator | 2025-08-29 17:19:15.776796 | orchestrator | TASK [Apply netplan configuration] ********************************************* 2025-08-29 17:19:15.776807 | orchestrator | Friday 29 August 2025 17:18:55 +0000 (0:00:00.717) 0:00:00.874 ********* 2025-08-29 17:19:15.776818 | orchestrator | ok: [testbed-manager] 2025-08-29 17:19:15.776830 | orchestrator | 2025-08-29 17:19:15.776842 | orchestrator | PLAY [Apply netplan configuration on all other nodes] ************************** 2025-08-29 17:19:15.776852 | orchestrator | 2025-08-29 17:19:15.776872 | orchestrator | TASK [Apply netplan configuration] ********************************************* 2025-08-29 17:19:15.776883 | orchestrator | Friday 29 August 2025 17:18:57 +0000 (0:00:02.281) 0:00:03.155 ********* 2025-08-29 17:19:15.776895 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:19:15.776906 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:19:15.776917 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:19:15.776928 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:19:15.776939 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:19:15.776950 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:19:15.776961 | orchestrator | 2025-08-29 17:19:15.776972 | orchestrator | PLAY [Add custom CA certificates to non-manager nodes] ************************* 2025-08-29 17:19:15.776983 | orchestrator | 2025-08-29 17:19:15.777012 | orchestrator | TASK [Copy custom CA certificates] ********************************************* 2025-08-29 17:19:15.777024 | orchestrator | Friday 29 August 2025 17:18:59 +0000 (0:00:01.804) 0:00:04.959 ********* 2025-08-29 17:19:15.777036 | orchestrator | changed: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt) 2025-08-29 17:19:15.777047 | orchestrator | changed: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt) 2025-08-29 17:19:15.777058 | orchestrator | changed: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt) 2025-08-29 17:19:15.777071 | orchestrator | changed: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt) 2025-08-29 17:19:15.777082 | orchestrator | changed: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt) 2025-08-29 17:19:15.777095 | orchestrator | changed: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt) 2025-08-29 17:19:15.777106 | orchestrator | 2025-08-29 17:19:15.777119 | orchestrator | TASK [Run update-ca-certificates] ********************************************** 2025-08-29 17:19:15.777131 | orchestrator | Friday 29 August 2025 17:19:00 +0000 (0:00:01.445) 0:00:06.405 ********* 2025-08-29 17:19:15.777143 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:19:15.777155 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:19:15.777167 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:19:15.777179 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:19:15.777191 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:19:15.777203 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:19:15.777214 | orchestrator | 2025-08-29 17:19:15.777227 | orchestrator | TASK [Run update-ca-trust] ***************************************************** 2025-08-29 17:19:15.777239 | orchestrator | Friday 29 August 2025 17:19:04 +0000 (0:00:03.344) 0:00:09.749 ********* 2025-08-29 17:19:15.777251 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:19:15.777263 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:19:15.777275 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:19:15.777288 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:19:15.777300 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:19:15.777312 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:19:15.777366 | orchestrator | 2025-08-29 17:19:15.777378 | orchestrator | PLAY [Add a workaround service] ************************************************ 2025-08-29 17:19:15.777391 | orchestrator | 2025-08-29 17:19:15.777403 | orchestrator | TASK [Copy workarounds.sh scripts] ********************************************* 2025-08-29 17:19:15.777416 | orchestrator | Friday 29 August 2025 17:19:04 +0000 (0:00:00.754) 0:00:10.504 ********* 2025-08-29 17:19:15.777427 | orchestrator | changed: [testbed-manager] 2025-08-29 17:19:15.777438 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:19:15.777449 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:19:15.777460 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:19:15.777471 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:19:15.777482 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:19:15.777492 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:19:15.777503 | orchestrator | 2025-08-29 17:19:15.777514 | orchestrator | TASK [Copy workarounds systemd unit file] ************************************** 2025-08-29 17:19:15.777525 | orchestrator | Friday 29 August 2025 17:19:06 +0000 (0:00:01.777) 0:00:12.281 ********* 2025-08-29 17:19:15.777535 | orchestrator | changed: [testbed-manager] 2025-08-29 17:19:15.777546 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:19:15.777557 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:19:15.777568 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:19:15.777578 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:19:15.777589 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:19:15.777617 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:19:15.777628 | orchestrator | 2025-08-29 17:19:15.777640 | orchestrator | TASK [Reload systemd daemon] *************************************************** 2025-08-29 17:19:15.777650 | orchestrator | Friday 29 August 2025 17:19:08 +0000 (0:00:01.947) 0:00:14.229 ********* 2025-08-29 17:19:15.777667 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:19:15.777678 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:19:15.777689 | orchestrator | ok: [testbed-manager] 2025-08-29 17:19:15.777699 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:19:15.777710 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:19:15.777721 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:19:15.777731 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:19:15.777742 | orchestrator | 2025-08-29 17:19:15.777753 | orchestrator | TASK [Enable workarounds.service (Debian)] ************************************* 2025-08-29 17:19:15.777764 | orchestrator | Friday 29 August 2025 17:19:10 +0000 (0:00:01.663) 0:00:15.893 ********* 2025-08-29 17:19:15.777775 | orchestrator | changed: [testbed-manager] 2025-08-29 17:19:15.777786 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:19:15.777797 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:19:15.777807 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:19:15.777818 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:19:15.777829 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:19:15.777839 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:19:15.777850 | orchestrator | 2025-08-29 17:19:15.777861 | orchestrator | TASK [Enable and start workarounds.service (RedHat)] *************************** 2025-08-29 17:19:15.777871 | orchestrator | Friday 29 August 2025 17:19:12 +0000 (0:00:01.839) 0:00:17.732 ********* 2025-08-29 17:19:15.777882 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:19:15.777893 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:19:15.777903 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:19:15.777914 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:19:15.777929 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:19:15.777940 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:19:15.777951 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:19:15.777962 | orchestrator | 2025-08-29 17:19:15.777973 | orchestrator | PLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ****************** 2025-08-29 17:19:15.777984 | orchestrator | 2025-08-29 17:19:15.777994 | orchestrator | TASK [Install python3-docker] ************************************************** 2025-08-29 17:19:15.778006 | orchestrator | Friday 29 August 2025 17:19:12 +0000 (0:00:00.652) 0:00:18.384 ********* 2025-08-29 17:19:15.778085 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:19:15.778100 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:19:15.778111 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:19:15.778121 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:19:15.778132 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:19:15.778143 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:19:15.778154 | orchestrator | ok: [testbed-manager] 2025-08-29 17:19:15.778165 | orchestrator | 2025-08-29 17:19:15.778176 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:19:15.778188 | orchestrator | testbed-manager : ok=7  changed=4  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:19:15.778200 | orchestrator | testbed-node-0 : ok=9  changed=6  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:15.778211 | orchestrator | testbed-node-1 : ok=9  changed=6  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:15.778222 | orchestrator | testbed-node-2 : ok=9  changed=6  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:15.778233 | orchestrator | testbed-node-3 : ok=9  changed=6  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:15.778244 | orchestrator | testbed-node-4 : ok=9  changed=6  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:15.778255 | orchestrator | testbed-node-5 : ok=9  changed=6  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:15.778272 | orchestrator | 2025-08-29 17:19:15.778284 | orchestrator | 2025-08-29 17:19:15.778294 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:19:15.778305 | orchestrator | Friday 29 August 2025 17:19:15 +0000 (0:00:02.942) 0:00:21.327 ********* 2025-08-29 17:19:15.778333 | orchestrator | =============================================================================== 2025-08-29 17:19:15.778344 | orchestrator | Run update-ca-certificates ---------------------------------------------- 3.34s 2025-08-29 17:19:15.778355 | orchestrator | Install python3-docker -------------------------------------------------- 2.94s 2025-08-29 17:19:15.778366 | orchestrator | Apply netplan configuration --------------------------------------------- 2.28s 2025-08-29 17:19:15.778377 | orchestrator | Copy workarounds systemd unit file -------------------------------------- 1.95s 2025-08-29 17:19:15.778387 | orchestrator | Enable workarounds.service (Debian) ------------------------------------- 1.84s 2025-08-29 17:19:15.778398 | orchestrator | Apply netplan configuration --------------------------------------------- 1.80s 2025-08-29 17:19:15.778409 | orchestrator | Copy workarounds.sh scripts --------------------------------------------- 1.78s 2025-08-29 17:19:15.778420 | orchestrator | Reload systemd daemon --------------------------------------------------- 1.66s 2025-08-29 17:19:15.778431 | orchestrator | Copy custom CA certificates --------------------------------------------- 1.45s 2025-08-29 17:19:15.778442 | orchestrator | Run update-ca-trust ----------------------------------------------------- 0.75s 2025-08-29 17:19:15.778453 | orchestrator | Group hosts based on virtualization_role -------------------------------- 0.72s 2025-08-29 17:19:15.778472 | orchestrator | Enable and start workarounds.service (RedHat) --------------------------- 0.65s 2025-08-29 17:19:16.491183 | orchestrator | + osism apply reboot -l testbed-nodes -e ireallymeanit=yes 2025-08-29 17:19:28.519352 | orchestrator | 2025-08-29 17:19:28 | INFO  | Task 16987030-0d49-4283-b490-5ad8fe01e1c8 (reboot) was prepared for execution. 2025-08-29 17:19:28.519453 | orchestrator | 2025-08-29 17:19:28 | INFO  | It takes a moment until task 16987030-0d49-4283-b490-5ad8fe01e1c8 (reboot) has been started and output is visible here. 2025-08-29 17:19:38.518892 | orchestrator | 2025-08-29 17:19:38.519011 | orchestrator | PLAY [Reboot systems] ********************************************************** 2025-08-29 17:19:38.519028 | orchestrator | 2025-08-29 17:19:38.519040 | orchestrator | TASK [Exit playbook, if user did not mean to reboot systems] ******************* 2025-08-29 17:19:38.519051 | orchestrator | Friday 29 August 2025 17:19:32 +0000 (0:00:00.233) 0:00:00.233 ********* 2025-08-29 17:19:38.519063 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:19:38.519074 | orchestrator | 2025-08-29 17:19:38.519085 | orchestrator | TASK [Reboot system - do not wait for the reboot to complete] ****************** 2025-08-29 17:19:38.519096 | orchestrator | Friday 29 August 2025 17:19:32 +0000 (0:00:00.098) 0:00:00.331 ********* 2025-08-29 17:19:38.519107 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:19:38.519118 | orchestrator | 2025-08-29 17:19:38.519129 | orchestrator | TASK [Reboot system - wait for the reboot to complete] ************************* 2025-08-29 17:19:38.519140 | orchestrator | Friday 29 August 2025 17:19:33 +0000 (0:00:00.956) 0:00:01.288 ********* 2025-08-29 17:19:38.519150 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:19:38.519161 | orchestrator | 2025-08-29 17:19:38.519173 | orchestrator | PLAY [Reboot systems] ********************************************************** 2025-08-29 17:19:38.519183 | orchestrator | 2025-08-29 17:19:38.519194 | orchestrator | TASK [Exit playbook, if user did not mean to reboot systems] ******************* 2025-08-29 17:19:38.519205 | orchestrator | Friday 29 August 2025 17:19:33 +0000 (0:00:00.100) 0:00:01.389 ********* 2025-08-29 17:19:38.519216 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:19:38.519227 | orchestrator | 2025-08-29 17:19:38.519237 | orchestrator | TASK [Reboot system - do not wait for the reboot to complete] ****************** 2025-08-29 17:19:38.519248 | orchestrator | Friday 29 August 2025 17:19:34 +0000 (0:00:00.095) 0:00:01.484 ********* 2025-08-29 17:19:38.519283 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:19:38.519295 | orchestrator | 2025-08-29 17:19:38.519306 | orchestrator | TASK [Reboot system - wait for the reboot to complete] ************************* 2025-08-29 17:19:38.519375 | orchestrator | Friday 29 August 2025 17:19:34 +0000 (0:00:00.646) 0:00:02.130 ********* 2025-08-29 17:19:38.519386 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:19:38.519397 | orchestrator | 2025-08-29 17:19:38.519408 | orchestrator | PLAY [Reboot systems] ********************************************************** 2025-08-29 17:19:38.519419 | orchestrator | 2025-08-29 17:19:38.519430 | orchestrator | TASK [Exit playbook, if user did not mean to reboot systems] ******************* 2025-08-29 17:19:38.519442 | orchestrator | Friday 29 August 2025 17:19:34 +0000 (0:00:00.108) 0:00:02.239 ********* 2025-08-29 17:19:38.519454 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:19:38.519466 | orchestrator | 2025-08-29 17:19:38.519478 | orchestrator | TASK [Reboot system - do not wait for the reboot to complete] ****************** 2025-08-29 17:19:38.519490 | orchestrator | Friday 29 August 2025 17:19:35 +0000 (0:00:00.209) 0:00:02.449 ********* 2025-08-29 17:19:38.519502 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:19:38.519514 | orchestrator | 2025-08-29 17:19:38.519525 | orchestrator | TASK [Reboot system - wait for the reboot to complete] ************************* 2025-08-29 17:19:38.519537 | orchestrator | Friday 29 August 2025 17:19:35 +0000 (0:00:00.602) 0:00:03.051 ********* 2025-08-29 17:19:38.519549 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:19:38.519561 | orchestrator | 2025-08-29 17:19:38.519578 | orchestrator | PLAY [Reboot systems] ********************************************************** 2025-08-29 17:19:38.519591 | orchestrator | 2025-08-29 17:19:38.519603 | orchestrator | TASK [Exit playbook, if user did not mean to reboot systems] ******************* 2025-08-29 17:19:38.519615 | orchestrator | Friday 29 August 2025 17:19:35 +0000 (0:00:00.110) 0:00:03.162 ********* 2025-08-29 17:19:38.519626 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:19:38.519638 | orchestrator | 2025-08-29 17:19:38.519649 | orchestrator | TASK [Reboot system - do not wait for the reboot to complete] ****************** 2025-08-29 17:19:38.519659 | orchestrator | Friday 29 August 2025 17:19:35 +0000 (0:00:00.098) 0:00:03.260 ********* 2025-08-29 17:19:38.519670 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:19:38.519681 | orchestrator | 2025-08-29 17:19:38.519691 | orchestrator | TASK [Reboot system - wait for the reboot to complete] ************************* 2025-08-29 17:19:38.519702 | orchestrator | Friday 29 August 2025 17:19:36 +0000 (0:00:00.634) 0:00:03.895 ********* 2025-08-29 17:19:38.519713 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:19:38.519723 | orchestrator | 2025-08-29 17:19:38.519734 | orchestrator | PLAY [Reboot systems] ********************************************************** 2025-08-29 17:19:38.519745 | orchestrator | 2025-08-29 17:19:38.519755 | orchestrator | TASK [Exit playbook, if user did not mean to reboot systems] ******************* 2025-08-29 17:19:38.519766 | orchestrator | Friday 29 August 2025 17:19:36 +0000 (0:00:00.116) 0:00:04.011 ********* 2025-08-29 17:19:38.519777 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:19:38.519788 | orchestrator | 2025-08-29 17:19:38.519798 | orchestrator | TASK [Reboot system - do not wait for the reboot to complete] ****************** 2025-08-29 17:19:38.519809 | orchestrator | Friday 29 August 2025 17:19:36 +0000 (0:00:00.103) 0:00:04.115 ********* 2025-08-29 17:19:38.519819 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:19:38.519830 | orchestrator | 2025-08-29 17:19:38.519841 | orchestrator | TASK [Reboot system - wait for the reboot to complete] ************************* 2025-08-29 17:19:38.519851 | orchestrator | Friday 29 August 2025 17:19:37 +0000 (0:00:00.635) 0:00:04.751 ********* 2025-08-29 17:19:38.519862 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:19:38.519872 | orchestrator | 2025-08-29 17:19:38.519883 | orchestrator | PLAY [Reboot systems] ********************************************************** 2025-08-29 17:19:38.519894 | orchestrator | 2025-08-29 17:19:38.519905 | orchestrator | TASK [Exit playbook, if user did not mean to reboot systems] ******************* 2025-08-29 17:19:38.519916 | orchestrator | Friday 29 August 2025 17:19:37 +0000 (0:00:00.108) 0:00:04.860 ********* 2025-08-29 17:19:38.519926 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:19:38.519946 | orchestrator | 2025-08-29 17:19:38.519957 | orchestrator | TASK [Reboot system - do not wait for the reboot to complete] ****************** 2025-08-29 17:19:38.519968 | orchestrator | Friday 29 August 2025 17:19:37 +0000 (0:00:00.094) 0:00:04.954 ********* 2025-08-29 17:19:38.519978 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:19:38.519989 | orchestrator | 2025-08-29 17:19:38.519999 | orchestrator | TASK [Reboot system - wait for the reboot to complete] ************************* 2025-08-29 17:19:38.520010 | orchestrator | Friday 29 August 2025 17:19:38 +0000 (0:00:00.631) 0:00:05.585 ********* 2025-08-29 17:19:38.520038 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:19:38.520049 | orchestrator | 2025-08-29 17:19:38.520060 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:19:38.520072 | orchestrator | testbed-node-0 : ok=1  changed=1  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:38.520101 | orchestrator | testbed-node-1 : ok=1  changed=1  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:38.520112 | orchestrator | testbed-node-2 : ok=1  changed=1  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:38.520136 | orchestrator | testbed-node-3 : ok=1  changed=1  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:38.520147 | orchestrator | testbed-node-4 : ok=1  changed=1  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:38.520158 | orchestrator | testbed-node-5 : ok=1  changed=1  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:19:38.520169 | orchestrator | 2025-08-29 17:19:38.520180 | orchestrator | 2025-08-29 17:19:38.520190 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:19:38.520201 | orchestrator | Friday 29 August 2025 17:19:38 +0000 (0:00:00.036) 0:00:05.622 ********* 2025-08-29 17:19:38.520212 | orchestrator | =============================================================================== 2025-08-29 17:19:38.520223 | orchestrator | Reboot system - do not wait for the reboot to complete ------------------ 4.11s 2025-08-29 17:19:38.520234 | orchestrator | Exit playbook, if user did not mean to reboot systems ------------------- 0.70s 2025-08-29 17:19:38.520244 | orchestrator | Reboot system - wait for the reboot to complete ------------------------- 0.58s 2025-08-29 17:19:38.821995 | orchestrator | + osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes 2025-08-29 17:19:50.741101 | orchestrator | 2025-08-29 17:19:50 | INFO  | Task 3572e037-163c-4731-936e-65a2afe05b15 (wait-for-connection) was prepared for execution. 2025-08-29 17:19:50.741211 | orchestrator | 2025-08-29 17:19:50 | INFO  | It takes a moment until task 3572e037-163c-4731-936e-65a2afe05b15 (wait-for-connection) has been started and output is visible here. 2025-08-29 17:20:06.294139 | orchestrator | 2025-08-29 17:20:06.294241 | orchestrator | PLAY [Wait until remote systems are reachable] ********************************* 2025-08-29 17:20:06.294257 | orchestrator | 2025-08-29 17:20:06.294269 | orchestrator | TASK [Wait until remote system is reachable] *********************************** 2025-08-29 17:20:06.294280 | orchestrator | Friday 29 August 2025 17:19:54 +0000 (0:00:00.186) 0:00:00.186 ********* 2025-08-29 17:20:06.294291 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:20:06.294303 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:20:06.294314 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:20:06.294364 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:20:06.294376 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:20:06.294386 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:20:06.294397 | orchestrator | 2025-08-29 17:20:06.294408 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:20:06.294420 | orchestrator | testbed-node-0 : ok=1  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:20:06.294455 | orchestrator | testbed-node-1 : ok=1  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:20:06.294467 | orchestrator | testbed-node-2 : ok=1  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:20:06.294478 | orchestrator | testbed-node-3 : ok=1  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:20:06.294489 | orchestrator | testbed-node-4 : ok=1  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:20:06.294500 | orchestrator | testbed-node-5 : ok=1  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:20:06.294511 | orchestrator | 2025-08-29 17:20:06.294522 | orchestrator | 2025-08-29 17:20:06.294532 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:20:06.294543 | orchestrator | Friday 29 August 2025 17:20:06 +0000 (0:00:11.473) 0:00:11.659 ********* 2025-08-29 17:20:06.294554 | orchestrator | =============================================================================== 2025-08-29 17:20:06.294565 | orchestrator | Wait until remote system is reachable ---------------------------------- 11.47s 2025-08-29 17:20:06.502542 | orchestrator | + osism apply hddtemp 2025-08-29 17:20:18.252171 | orchestrator | 2025-08-29 17:20:18 | INFO  | Task 3b396da4-a1c6-4468-b533-473e19f255ba (hddtemp) was prepared for execution. 2025-08-29 17:20:18.252284 | orchestrator | 2025-08-29 17:20:18 | INFO  | It takes a moment until task 3b396da4-a1c6-4468-b533-473e19f255ba (hddtemp) has been started and output is visible here. 2025-08-29 17:20:49.013475 | orchestrator | 2025-08-29 17:20:49.013559 | orchestrator | PLAY [Apply role hddtemp] ****************************************************** 2025-08-29 17:20:49.013569 | orchestrator | 2025-08-29 17:20:49.013577 | orchestrator | TASK [osism.services.hddtemp : Gather variables for each operating system] ***** 2025-08-29 17:20:49.013586 | orchestrator | Friday 29 August 2025 17:20:22 +0000 (0:00:00.274) 0:00:00.274 ********* 2025-08-29 17:20:49.013593 | orchestrator | ok: [testbed-manager] 2025-08-29 17:20:49.013603 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:20:49.013607 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:20:49.013611 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:20:49.013615 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:20:49.013619 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:20:49.013623 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:20:49.013627 | orchestrator | 2025-08-29 17:20:49.013631 | orchestrator | TASK [osism.services.hddtemp : Include distribution specific install tasks] **** 2025-08-29 17:20:49.013635 | orchestrator | Friday 29 August 2025 17:20:23 +0000 (0:00:00.696) 0:00:00.970 ********* 2025-08-29 17:20:49.013653 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:20:49.013659 | orchestrator | 2025-08-29 17:20:49.013663 | orchestrator | TASK [osism.services.hddtemp : Remove hddtemp package] ************************* 2025-08-29 17:20:49.013667 | orchestrator | Friday 29 August 2025 17:20:24 +0000 (0:00:01.203) 0:00:02.174 ********* 2025-08-29 17:20:49.013671 | orchestrator | ok: [testbed-manager] 2025-08-29 17:20:49.013675 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:20:49.013679 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:20:49.013682 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:20:49.013686 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:20:49.013690 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:20:49.013694 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:20:49.013697 | orchestrator | 2025-08-29 17:20:49.013701 | orchestrator | TASK [osism.services.hddtemp : Enable Kernel Module drivetemp] ***************** 2025-08-29 17:20:49.013705 | orchestrator | Friday 29 August 2025 17:20:26 +0000 (0:00:02.122) 0:00:04.297 ********* 2025-08-29 17:20:49.013724 | orchestrator | changed: [testbed-manager] 2025-08-29 17:20:49.013729 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:20:49.013733 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:20:49.013737 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:20:49.013740 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:20:49.013744 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:20:49.013748 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:20:49.013751 | orchestrator | 2025-08-29 17:20:49.013755 | orchestrator | TASK [osism.services.hddtemp : Check if drivetemp module is available] ********* 2025-08-29 17:20:49.013759 | orchestrator | Friday 29 August 2025 17:20:27 +0000 (0:00:01.214) 0:00:05.511 ********* 2025-08-29 17:20:49.013763 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:20:49.013767 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:20:49.013770 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:20:49.013774 | orchestrator | ok: [testbed-manager] 2025-08-29 17:20:49.013778 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:20:49.013781 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:20:49.013785 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:20:49.013789 | orchestrator | 2025-08-29 17:20:49.013793 | orchestrator | TASK [osism.services.hddtemp : Load Kernel Module drivetemp] ******************* 2025-08-29 17:20:49.013796 | orchestrator | Friday 29 August 2025 17:20:29 +0000 (0:00:01.983) 0:00:07.495 ********* 2025-08-29 17:20:49.013800 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:20:49.013804 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:20:49.013808 | orchestrator | changed: [testbed-manager] 2025-08-29 17:20:49.013811 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:20:49.013815 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:20:49.013819 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:20:49.013823 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:20:49.013826 | orchestrator | 2025-08-29 17:20:49.013830 | orchestrator | TASK [osism.services.hddtemp : Install lm-sensors] ***************************** 2025-08-29 17:20:49.013834 | orchestrator | Friday 29 August 2025 17:20:30 +0000 (0:00:00.806) 0:00:08.301 ********* 2025-08-29 17:20:49.013838 | orchestrator | changed: [testbed-manager] 2025-08-29 17:20:49.013841 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:20:49.013845 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:20:49.013849 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:20:49.013853 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:20:49.013856 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:20:49.013860 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:20:49.013864 | orchestrator | 2025-08-29 17:20:49.013867 | orchestrator | TASK [osism.services.hddtemp : Include distribution specific service tasks] **** 2025-08-29 17:20:49.013871 | orchestrator | Friday 29 August 2025 17:20:44 +0000 (0:00:14.058) 0:00:22.360 ********* 2025-08-29 17:20:49.013875 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:20:49.013879 | orchestrator | 2025-08-29 17:20:49.013883 | orchestrator | TASK [osism.services.hddtemp : Manage lm-sensors service] ********************** 2025-08-29 17:20:49.013887 | orchestrator | Friday 29 August 2025 17:20:45 +0000 (0:00:01.461) 0:00:23.821 ********* 2025-08-29 17:20:49.013891 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:20:49.013894 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:20:49.013898 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:20:49.013902 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:20:49.013905 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:20:49.013909 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:20:49.013913 | orchestrator | changed: [testbed-manager] 2025-08-29 17:20:49.013916 | orchestrator | 2025-08-29 17:20:49.013921 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:20:49.013925 | orchestrator | testbed-manager : ok=9  changed=4  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:20:49.013945 | orchestrator | testbed-node-0 : ok=8  changed=3  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:20:49.013950 | orchestrator | testbed-node-1 : ok=8  changed=3  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:20:49.013954 | orchestrator | testbed-node-2 : ok=8  changed=3  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:20:49.013958 | orchestrator | testbed-node-3 : ok=8  changed=3  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:20:49.013962 | orchestrator | testbed-node-4 : ok=8  changed=3  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:20:49.013968 | orchestrator | testbed-node-5 : ok=8  changed=3  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:20:49.013972 | orchestrator | 2025-08-29 17:20:49.013976 | orchestrator | 2025-08-29 17:20:49.013980 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:20:49.013984 | orchestrator | Friday 29 August 2025 17:20:48 +0000 (0:00:02.614) 0:00:26.436 ********* 2025-08-29 17:20:49.013988 | orchestrator | =============================================================================== 2025-08-29 17:20:49.013991 | orchestrator | osism.services.hddtemp : Install lm-sensors ---------------------------- 14.06s 2025-08-29 17:20:49.013995 | orchestrator | osism.services.hddtemp : Manage lm-sensors service ---------------------- 2.61s 2025-08-29 17:20:49.013999 | orchestrator | osism.services.hddtemp : Remove hddtemp package ------------------------- 2.12s 2025-08-29 17:20:49.014003 | orchestrator | osism.services.hddtemp : Check if drivetemp module is available --------- 1.98s 2025-08-29 17:20:49.014007 | orchestrator | osism.services.hddtemp : Include distribution specific service tasks ---- 1.46s 2025-08-29 17:20:49.014010 | orchestrator | osism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.21s 2025-08-29 17:20:49.014050 | orchestrator | osism.services.hddtemp : Include distribution specific install tasks ---- 1.20s 2025-08-29 17:20:49.014056 | orchestrator | osism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.81s 2025-08-29 17:20:49.014061 | orchestrator | osism.services.hddtemp : Gather variables for each operating system ----- 0.70s 2025-08-29 17:20:49.333299 | orchestrator | ++ semver 9.2.0 7.1.1 2025-08-29 17:20:49.378812 | orchestrator | + [[ 1 -ge 0 ]] 2025-08-29 17:20:49.378880 | orchestrator | + sudo systemctl restart manager.service 2025-08-29 17:21:03.257129 | orchestrator | + [[ ceph-ansible == \c\e\p\h\-\a\n\s\i\b\l\e ]] 2025-08-29 17:21:03.257240 | orchestrator | + wait_for_container_healthy 60 ceph-ansible 2025-08-29 17:21:03.257257 | orchestrator | + local max_attempts=60 2025-08-29 17:21:03.257270 | orchestrator | + local name=ceph-ansible 2025-08-29 17:21:03.257282 | orchestrator | + local attempt_num=1 2025-08-29 17:21:03.257293 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:03.293096 | orchestrator | + [[ unhealthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:03.293178 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:03.293194 | orchestrator | + sleep 5 2025-08-29 17:21:08.299212 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:08.335231 | orchestrator | + [[ unhealthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:08.335354 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:08.335374 | orchestrator | + sleep 5 2025-08-29 17:21:13.340544 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:13.380523 | orchestrator | + [[ unhealthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:13.380595 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:13.380608 | orchestrator | + sleep 5 2025-08-29 17:21:18.385058 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:18.428349 | orchestrator | + [[ unhealthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:18.428418 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:18.428431 | orchestrator | + sleep 5 2025-08-29 17:21:23.435578 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:23.473573 | orchestrator | + [[ unhealthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:23.473652 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:23.473666 | orchestrator | + sleep 5 2025-08-29 17:21:28.479328 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:28.530740 | orchestrator | + [[ unhealthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:28.530810 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:28.530823 | orchestrator | + sleep 5 2025-08-29 17:21:33.536158 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:33.580066 | orchestrator | + [[ unhealthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:33.580165 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:33.580179 | orchestrator | + sleep 5 2025-08-29 17:21:38.586392 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:38.620798 | orchestrator | + [[ starting == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:38.620888 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:38.620902 | orchestrator | + sleep 5 2025-08-29 17:21:43.623668 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:43.655424 | orchestrator | + [[ starting == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:43.655495 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:43.655509 | orchestrator | + sleep 5 2025-08-29 17:21:48.658792 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:48.699354 | orchestrator | + [[ starting == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:48.699443 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:48.699457 | orchestrator | + sleep 5 2025-08-29 17:21:53.704232 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:53.746359 | orchestrator | + [[ starting == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:53.746424 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:53.746435 | orchestrator | + sleep 5 2025-08-29 17:21:58.751738 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:21:58.790663 | orchestrator | + [[ starting == \h\e\a\l\t\h\y ]] 2025-08-29 17:21:58.790739 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:21:58.790755 | orchestrator | + sleep 5 2025-08-29 17:22:03.794654 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:22:03.834670 | orchestrator | + [[ starting == \h\e\a\l\t\h\y ]] 2025-08-29 17:22:03.834737 | orchestrator | + (( attempt_num++ == max_attempts )) 2025-08-29 17:22:03.834751 | orchestrator | + sleep 5 2025-08-29 17:22:08.840162 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible 2025-08-29 17:22:08.881437 | orchestrator | + [[ healthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:22:08.881543 | orchestrator | + wait_for_container_healthy 60 kolla-ansible 2025-08-29 17:22:08.881559 | orchestrator | + local max_attempts=60 2025-08-29 17:22:08.881571 | orchestrator | + local name=kolla-ansible 2025-08-29 17:22:08.881583 | orchestrator | + local attempt_num=1 2025-08-29 17:22:08.881594 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible 2025-08-29 17:22:08.909674 | orchestrator | + [[ healthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:22:08.909714 | orchestrator | + wait_for_container_healthy 60 osism-ansible 2025-08-29 17:22:08.909727 | orchestrator | + local max_attempts=60 2025-08-29 17:22:08.909739 | orchestrator | + local name=osism-ansible 2025-08-29 17:22:08.909750 | orchestrator | + local attempt_num=1 2025-08-29 17:22:08.910431 | orchestrator | ++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible 2025-08-29 17:22:08.949639 | orchestrator | + [[ healthy == \h\e\a\l\t\h\y ]] 2025-08-29 17:22:08.949673 | orchestrator | + [[ true == \t\r\u\e ]] 2025-08-29 17:22:08.949684 | orchestrator | + sh -c /opt/configuration/scripts/disable-ara.sh 2025-08-29 17:22:09.119845 | orchestrator | ARA in ceph-ansible already disabled. 2025-08-29 17:22:09.286672 | orchestrator | ARA in kolla-ansible already disabled. 2025-08-29 17:22:09.433041 | orchestrator | ARA in osism-ansible already disabled. 2025-08-29 17:22:09.581702 | orchestrator | ARA in osism-kubernetes already disabled. 2025-08-29 17:22:09.582056 | orchestrator | + osism apply gather-facts 2025-08-29 17:22:21.649775 | orchestrator | 2025-08-29 17:22:21 | INFO  | Task 9c2fa442-a32b-4cd7-8150-8148c6fbeae9 (gather-facts) was prepared for execution. 2025-08-29 17:22:21.649889 | orchestrator | 2025-08-29 17:22:21 | INFO  | It takes a moment until task 9c2fa442-a32b-4cd7-8150-8148c6fbeae9 (gather-facts) has been started and output is visible here. 2025-08-29 17:22:34.496973 | orchestrator | 2025-08-29 17:22:34.497093 | orchestrator | PLAY [Gather facts for all hosts] ********************************************** 2025-08-29 17:22:34.497118 | orchestrator | 2025-08-29 17:22:34.497138 | orchestrator | TASK [Gathers facts about hosts] *********************************************** 2025-08-29 17:22:34.497156 | orchestrator | Friday 29 August 2025 17:22:25 +0000 (0:00:00.227) 0:00:00.227 ********* 2025-08-29 17:22:34.497184 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:22:34.497205 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:22:34.497223 | orchestrator | ok: [testbed-manager] 2025-08-29 17:22:34.497243 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:22:34.497256 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:22:34.497267 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:22:34.497325 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:22:34.497338 | orchestrator | 2025-08-29 17:22:34.497349 | orchestrator | PLAY [Gather facts for all hosts if using --limit] ***************************** 2025-08-29 17:22:34.497361 | orchestrator | 2025-08-29 17:22:34.497372 | orchestrator | TASK [Gather facts for all hosts] ********************************************** 2025-08-29 17:22:34.497383 | orchestrator | Friday 29 August 2025 17:22:33 +0000 (0:00:08.341) 0:00:08.569 ********* 2025-08-29 17:22:34.497423 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:22:34.497436 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:22:34.497446 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:22:34.497458 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:22:34.497469 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:22:34.497480 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:22:34.497490 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:22:34.497501 | orchestrator | 2025-08-29 17:22:34.497512 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:22:34.497523 | orchestrator | testbed-manager : ok=1  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:22:34.497536 | orchestrator | testbed-node-0 : ok=1  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:22:34.497547 | orchestrator | testbed-node-1 : ok=1  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:22:34.497558 | orchestrator | testbed-node-2 : ok=1  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:22:34.497569 | orchestrator | testbed-node-3 : ok=1  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:22:34.497580 | orchestrator | testbed-node-4 : ok=1  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:22:34.497591 | orchestrator | testbed-node-5 : ok=1  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:22:34.497602 | orchestrator | 2025-08-29 17:22:34.497613 | orchestrator | 2025-08-29 17:22:34.497624 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:22:34.497635 | orchestrator | Friday 29 August 2025 17:22:34 +0000 (0:00:00.515) 0:00:09.084 ********* 2025-08-29 17:22:34.497646 | orchestrator | =============================================================================== 2025-08-29 17:22:34.497657 | orchestrator | Gathers facts about hosts ----------------------------------------------- 8.34s 2025-08-29 17:22:34.497668 | orchestrator | Gather facts for all hosts ---------------------------------------------- 0.52s 2025-08-29 17:22:34.763676 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper 2025-08-29 17:22:34.774870 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes 2025-08-29 17:22:34.784964 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi 2025-08-29 17:22:34.794415 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible 2025-08-29 17:22:34.807649 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook 2025-08-29 17:22:34.817175 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure 2025-08-29 17:22:34.830476 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack 2025-08-29 17:22:34.840039 | orchestrator | + sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring 2025-08-29 17:22:34.853463 | orchestrator | + sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes 2025-08-29 17:22:34.863005 | orchestrator | + sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi 2025-08-29 17:22:34.872059 | orchestrator | + sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible 2025-08-29 17:22:34.885409 | orchestrator | + sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook 2025-08-29 17:22:34.894252 | orchestrator | + sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure 2025-08-29 17:22:34.902606 | orchestrator | + sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack 2025-08-29 17:22:34.911254 | orchestrator | + sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring 2025-08-29 17:22:34.921024 | orchestrator | + sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack 2025-08-29 17:22:34.934922 | orchestrator | + sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia 2025-08-29 17:22:34.945784 | orchestrator | + sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi 2025-08-29 17:22:34.960892 | orchestrator | + sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry 2025-08-29 17:22:34.975246 | orchestrator | + sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images 2025-08-29 17:22:34.983467 | orchestrator | + [[ false == \t\r\u\e ]] 2025-08-29 17:22:35.305454 | orchestrator | ok: Runtime: 0:22:35.136315 2025-08-29 17:22:35.421884 | 2025-08-29 17:22:35.422042 | TASK [Deploy services] 2025-08-29 17:22:35.954048 | orchestrator | skipping: Conditional result was False 2025-08-29 17:22:35.966674 | 2025-08-29 17:22:35.966824 | TASK [Deploy in a nutshell] 2025-08-29 17:22:36.654508 | orchestrator | + set -e 2025-08-29 17:22:36.654640 | orchestrator | + source /opt/configuration/scripts/include.sh 2025-08-29 17:22:36.654656 | orchestrator | ++ export INTERACTIVE=false 2025-08-29 17:22:36.654669 | orchestrator | ++ INTERACTIVE=false 2025-08-29 17:22:36.654677 | orchestrator | ++ export OSISM_APPLY_RETRY=1 2025-08-29 17:22:36.654682 | orchestrator | ++ OSISM_APPLY_RETRY=1 2025-08-29 17:22:36.654696 | orchestrator | + source /opt/manager-vars.sh 2025-08-29 17:22:36.654719 | orchestrator | ++ export NUMBER_OF_NODES=6 2025-08-29 17:22:36.654730 | orchestrator | ++ NUMBER_OF_NODES=6 2025-08-29 17:22:36.654735 | orchestrator | ++ export CEPH_VERSION=reef 2025-08-29 17:22:36.654741 | orchestrator | ++ CEPH_VERSION=reef 2025-08-29 17:22:36.654746 | orchestrator | ++ export CONFIGURATION_VERSION=main 2025-08-29 17:22:36.654756 | orchestrator | ++ CONFIGURATION_VERSION=main 2025-08-29 17:22:36.654762 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 17:22:36.654772 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 17:22:36.654776 | orchestrator | ++ export OPENSTACK_VERSION=2024.2 2025-08-29 17:22:36.654782 | orchestrator | ++ OPENSTACK_VERSION=2024.2 2025-08-29 17:22:36.654789 | orchestrator | ++ export ARA=false 2025-08-29 17:22:36.654798 | orchestrator | ++ ARA=false 2025-08-29 17:22:36.654805 | orchestrator | ++ export DEPLOY_MODE=manager 2025-08-29 17:22:36.654812 | orchestrator | ++ DEPLOY_MODE=manager 2025-08-29 17:22:36.654816 | orchestrator | ++ export TEMPEST=false 2025-08-29 17:22:36.654819 | orchestrator | ++ TEMPEST=false 2025-08-29 17:22:36.654914 | orchestrator | ++ export IS_ZUUL=true 2025-08-29 17:22:36.654923 | orchestrator | ++ IS_ZUUL=true 2025-08-29 17:22:36.654927 | orchestrator | ++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 17:22:36.654934 | orchestrator | ++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 17:22:36.654943 | orchestrator | ++ export EXTERNAL_API=false 2025-08-29 17:22:36.654949 | orchestrator | ++ EXTERNAL_API=false 2025-08-29 17:22:36.654956 | orchestrator | ++ export IMAGE_USER=ubuntu 2025-08-29 17:22:36.654960 | orchestrator | ++ IMAGE_USER=ubuntu 2025-08-29 17:22:36.654964 | orchestrator | ++ export IMAGE_NODE_USER=ubuntu 2025-08-29 17:22:36.654968 | orchestrator | ++ IMAGE_NODE_USER=ubuntu 2025-08-29 17:22:36.655048 | orchestrator | ++ export CEPH_STACK=ceph-ansible 2025-08-29 17:22:36.655055 | orchestrator | ++ CEPH_STACK=ceph-ansible 2025-08-29 17:22:36.655110 | orchestrator | 2025-08-29 17:22:36.655120 | orchestrator | # PULL IMAGES 2025-08-29 17:22:36.655126 | orchestrator | 2025-08-29 17:22:36.655132 | orchestrator | + echo 2025-08-29 17:22:36.655136 | orchestrator | + echo '# PULL IMAGES' 2025-08-29 17:22:36.655140 | orchestrator | + echo 2025-08-29 17:22:36.656910 | orchestrator | ++ semver 9.2.0 7.0.0 2025-08-29 17:22:36.719555 | orchestrator | + [[ 1 -ge 0 ]] 2025-08-29 17:22:36.719612 | orchestrator | + osism apply --no-wait -r 2 -e custom pull-images 2025-08-29 17:22:38.273588 | orchestrator | 2025-08-29 17:22:38 | INFO  | Trying to run play pull-images in environment custom 2025-08-29 17:22:48.483872 | orchestrator | 2025-08-29 17:22:48 | INFO  | Task 41b512ce-efba-4b78-8c6d-a850f9e7b0ba (pull-images) was prepared for execution. 2025-08-29 17:22:48.483965 | orchestrator | 2025-08-29 17:22:48 | INFO  | Task 41b512ce-efba-4b78-8c6d-a850f9e7b0ba is running in background. No more output. Check ARA for logs. 2025-08-29 17:22:50.571673 | orchestrator | 2025-08-29 17:22:50 | INFO  | Trying to run play wipe-partitions in environment custom 2025-08-29 17:23:00.785610 | orchestrator | 2025-08-29 17:23:00 | INFO  | Task 75f6d8bf-c64e-490a-a8fb-2acd2d1bcfd5 (wipe-partitions) was prepared for execution. 2025-08-29 17:23:00.785711 | orchestrator | 2025-08-29 17:23:00 | INFO  | It takes a moment until task 75f6d8bf-c64e-490a-a8fb-2acd2d1bcfd5 (wipe-partitions) has been started and output is visible here. 2025-08-29 17:23:12.312007 | orchestrator | 2025-08-29 17:23:12.312115 | orchestrator | PLAY [Wipe partitions] ********************************************************* 2025-08-29 17:23:12.312131 | orchestrator | 2025-08-29 17:23:12.312143 | orchestrator | TASK [Find all logical devices owned by UID 167] ******************************* 2025-08-29 17:23:12.312163 | orchestrator | Friday 29 August 2025 17:23:04 +0000 (0:00:00.116) 0:00:00.116 ********* 2025-08-29 17:23:12.312176 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:23:12.312188 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:23:12.312199 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:23:12.312210 | orchestrator | 2025-08-29 17:23:12.312221 | orchestrator | TASK [Remove all rook related logical devices] ********************************* 2025-08-29 17:23:12.312257 | orchestrator | Friday 29 August 2025 17:23:04 +0000 (0:00:00.518) 0:00:00.634 ********* 2025-08-29 17:23:12.312269 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:12.312325 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:23:12.312337 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:23:12.312352 | orchestrator | 2025-08-29 17:23:12.312363 | orchestrator | TASK [Find all logical devices with prefix ceph] ******************************* 2025-08-29 17:23:12.312375 | orchestrator | Friday 29 August 2025 17:23:04 +0000 (0:00:00.211) 0:00:00.846 ********* 2025-08-29 17:23:12.312386 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:23:12.312397 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:23:12.312407 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:23:12.312418 | orchestrator | 2025-08-29 17:23:12.312429 | orchestrator | TASK [Remove all ceph related logical devices] ********************************* 2025-08-29 17:23:12.312440 | orchestrator | Friday 29 August 2025 17:23:05 +0000 (0:00:00.612) 0:00:01.459 ********* 2025-08-29 17:23:12.312451 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:12.312462 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:23:12.312472 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:23:12.312483 | orchestrator | 2025-08-29 17:23:12.312494 | orchestrator | TASK [Check device availability] *********************************************** 2025-08-29 17:23:12.312504 | orchestrator | Friday 29 August 2025 17:23:05 +0000 (0:00:00.213) 0:00:01.672 ********* 2025-08-29 17:23:12.312516 | orchestrator | changed: [testbed-node-3] => (item=/dev/sdb) 2025-08-29 17:23:12.312531 | orchestrator | changed: [testbed-node-5] => (item=/dev/sdb) 2025-08-29 17:23:12.312542 | orchestrator | changed: [testbed-node-4] => (item=/dev/sdb) 2025-08-29 17:23:12.312554 | orchestrator | changed: [testbed-node-3] => (item=/dev/sdc) 2025-08-29 17:23:12.312567 | orchestrator | changed: [testbed-node-5] => (item=/dev/sdc) 2025-08-29 17:23:12.312579 | orchestrator | changed: [testbed-node-4] => (item=/dev/sdc) 2025-08-29 17:23:12.312591 | orchestrator | changed: [testbed-node-3] => (item=/dev/sdd) 2025-08-29 17:23:12.312604 | orchestrator | changed: [testbed-node-5] => (item=/dev/sdd) 2025-08-29 17:23:12.312616 | orchestrator | changed: [testbed-node-4] => (item=/dev/sdd) 2025-08-29 17:23:12.312628 | orchestrator | 2025-08-29 17:23:12.312640 | orchestrator | TASK [Wipe partitions with wipefs] ********************************************* 2025-08-29 17:23:12.312652 | orchestrator | Friday 29 August 2025 17:23:06 +0000 (0:00:01.220) 0:00:02.893 ********* 2025-08-29 17:23:12.312665 | orchestrator | ok: [testbed-node-3] => (item=/dev/sdb) 2025-08-29 17:23:12.312677 | orchestrator | ok: [testbed-node-4] => (item=/dev/sdb) 2025-08-29 17:23:12.312689 | orchestrator | ok: [testbed-node-5] => (item=/dev/sdb) 2025-08-29 17:23:12.312702 | orchestrator | ok: [testbed-node-3] => (item=/dev/sdc) 2025-08-29 17:23:12.312720 | orchestrator | ok: [testbed-node-4] => (item=/dev/sdc) 2025-08-29 17:23:12.312739 | orchestrator | ok: [testbed-node-5] => (item=/dev/sdc) 2025-08-29 17:23:12.312760 | orchestrator | ok: [testbed-node-3] => (item=/dev/sdd) 2025-08-29 17:23:12.312778 | orchestrator | ok: [testbed-node-4] => (item=/dev/sdd) 2025-08-29 17:23:12.312797 | orchestrator | ok: [testbed-node-5] => (item=/dev/sdd) 2025-08-29 17:23:12.312815 | orchestrator | 2025-08-29 17:23:12.312835 | orchestrator | TASK [Overwrite first 32M with zeros] ****************************************** 2025-08-29 17:23:12.312856 | orchestrator | Friday 29 August 2025 17:23:08 +0000 (0:00:01.363) 0:00:04.256 ********* 2025-08-29 17:23:12.312876 | orchestrator | changed: [testbed-node-3] => (item=/dev/sdb) 2025-08-29 17:23:12.312893 | orchestrator | changed: [testbed-node-4] => (item=/dev/sdb) 2025-08-29 17:23:12.312907 | orchestrator | changed: [testbed-node-5] => (item=/dev/sdb) 2025-08-29 17:23:12.312919 | orchestrator | changed: [testbed-node-3] => (item=/dev/sdc) 2025-08-29 17:23:12.312930 | orchestrator | changed: [testbed-node-4] => (item=/dev/sdc) 2025-08-29 17:23:12.312949 | orchestrator | changed: [testbed-node-5] => (item=/dev/sdc) 2025-08-29 17:23:12.312960 | orchestrator | changed: [testbed-node-3] => (item=/dev/sdd) 2025-08-29 17:23:12.312971 | orchestrator | changed: [testbed-node-4] => (item=/dev/sdd) 2025-08-29 17:23:12.312992 | orchestrator | changed: [testbed-node-5] => (item=/dev/sdd) 2025-08-29 17:23:12.313003 | orchestrator | 2025-08-29 17:23:12.313013 | orchestrator | TASK [Reload udev rules] ******************************************************* 2025-08-29 17:23:12.313024 | orchestrator | Friday 29 August 2025 17:23:10 +0000 (0:00:02.424) 0:00:06.681 ********* 2025-08-29 17:23:12.313035 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:23:12.313045 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:23:12.313056 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:23:12.313066 | orchestrator | 2025-08-29 17:23:12.313077 | orchestrator | TASK [Request device events from the kernel] *********************************** 2025-08-29 17:23:12.313087 | orchestrator | Friday 29 August 2025 17:23:11 +0000 (0:00:00.622) 0:00:07.303 ********* 2025-08-29 17:23:12.313098 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:23:12.313109 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:23:12.313119 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:23:12.313130 | orchestrator | 2025-08-29 17:23:12.313140 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:23:12.313153 | orchestrator | testbed-node-3 : ok=7  changed=5  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:12.313166 | orchestrator | testbed-node-4 : ok=7  changed=5  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:12.313195 | orchestrator | testbed-node-5 : ok=7  changed=5  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:12.313207 | orchestrator | 2025-08-29 17:23:12.313218 | orchestrator | 2025-08-29 17:23:12.313229 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:23:12.313239 | orchestrator | Friday 29 August 2025 17:23:12 +0000 (0:00:00.631) 0:00:07.935 ********* 2025-08-29 17:23:12.313250 | orchestrator | =============================================================================== 2025-08-29 17:23:12.313261 | orchestrator | Overwrite first 32M with zeros ------------------------------------------ 2.42s 2025-08-29 17:23:12.313271 | orchestrator | Wipe partitions with wipefs --------------------------------------------- 1.36s 2025-08-29 17:23:12.313311 | orchestrator | Check device availability ----------------------------------------------- 1.22s 2025-08-29 17:23:12.313325 | orchestrator | Request device events from the kernel ----------------------------------- 0.63s 2025-08-29 17:23:12.313343 | orchestrator | Reload udev rules ------------------------------------------------------- 0.62s 2025-08-29 17:23:12.313361 | orchestrator | Find all logical devices with prefix ceph ------------------------------- 0.61s 2025-08-29 17:23:12.313379 | orchestrator | Find all logical devices owned by UID 167 ------------------------------- 0.52s 2025-08-29 17:23:12.313407 | orchestrator | Remove all ceph related logical devices --------------------------------- 0.21s 2025-08-29 17:23:12.313428 | orchestrator | Remove all rook related logical devices --------------------------------- 0.21s 2025-08-29 17:23:24.581383 | orchestrator | 2025-08-29 17:23:24 | INFO  | Task dce79cc0-1f69-45fe-84bc-290e949a2495 (facts) was prepared for execution. 2025-08-29 17:23:24.581485 | orchestrator | 2025-08-29 17:23:24 | INFO  | It takes a moment until task dce79cc0-1f69-45fe-84bc-290e949a2495 (facts) has been started and output is visible here. 2025-08-29 17:23:37.190821 | orchestrator | 2025-08-29 17:23:37.190915 | orchestrator | PLAY [Apply role facts] ******************************************************** 2025-08-29 17:23:37.190932 | orchestrator | 2025-08-29 17:23:37.190945 | orchestrator | TASK [osism.commons.facts : Create custom facts directory] ********************* 2025-08-29 17:23:37.190959 | orchestrator | Friday 29 August 2025 17:23:28 +0000 (0:00:00.267) 0:00:00.267 ********* 2025-08-29 17:23:37.190968 | orchestrator | ok: [testbed-manager] 2025-08-29 17:23:37.190976 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:23:37.190984 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:23:37.190992 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:23:37.191017 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:23:37.191025 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:23:37.191033 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:23:37.191041 | orchestrator | 2025-08-29 17:23:37.191051 | orchestrator | TASK [osism.commons.facts : Copy fact files] *********************************** 2025-08-29 17:23:37.191060 | orchestrator | Friday 29 August 2025 17:23:29 +0000 (0:00:01.129) 0:00:01.397 ********* 2025-08-29 17:23:37.191068 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:23:37.191076 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:23:37.191084 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:23:37.191091 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:23:37.191099 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:37.191107 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:23:37.191114 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:23:37.191122 | orchestrator | 2025-08-29 17:23:37.191130 | orchestrator | PLAY [Gather facts for all hosts] ********************************************** 2025-08-29 17:23:37.191138 | orchestrator | 2025-08-29 17:23:37.191146 | orchestrator | TASK [Gathers facts about hosts] *********************************************** 2025-08-29 17:23:37.191153 | orchestrator | Friday 29 August 2025 17:23:30 +0000 (0:00:01.158) 0:00:02.555 ********* 2025-08-29 17:23:37.191161 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:23:37.191169 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:23:37.191177 | orchestrator | ok: [testbed-manager] 2025-08-29 17:23:37.191185 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:23:37.191193 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:23:37.191201 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:23:37.191208 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:23:37.191216 | orchestrator | 2025-08-29 17:23:37.191224 | orchestrator | PLAY [Gather facts for all hosts if using --limit] ***************************** 2025-08-29 17:23:37.191232 | orchestrator | 2025-08-29 17:23:37.191239 | orchestrator | TASK [Gather facts for all hosts] ********************************************** 2025-08-29 17:23:37.191255 | orchestrator | Friday 29 August 2025 17:23:36 +0000 (0:00:05.461) 0:00:08.017 ********* 2025-08-29 17:23:37.191263 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:23:37.191271 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:23:37.191305 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:23:37.191314 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:23:37.191321 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:37.191329 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:23:37.191337 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:23:37.191345 | orchestrator | 2025-08-29 17:23:37.191353 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:23:37.191361 | orchestrator | testbed-manager : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:37.191370 | orchestrator | testbed-node-0 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:37.191379 | orchestrator | testbed-node-1 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:37.191388 | orchestrator | testbed-node-2 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:37.191397 | orchestrator | testbed-node-3 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:37.191406 | orchestrator | testbed-node-4 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:37.191415 | orchestrator | testbed-node-5 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:23:37.191423 | orchestrator | 2025-08-29 17:23:37.191432 | orchestrator | 2025-08-29 17:23:37.191441 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:23:37.191455 | orchestrator | Friday 29 August 2025 17:23:36 +0000 (0:00:00.483) 0:00:08.501 ********* 2025-08-29 17:23:37.191464 | orchestrator | =============================================================================== 2025-08-29 17:23:37.191473 | orchestrator | Gathers facts about hosts ----------------------------------------------- 5.46s 2025-08-29 17:23:37.191482 | orchestrator | osism.commons.facts : Copy fact files ----------------------------------- 1.16s 2025-08-29 17:23:37.191492 | orchestrator | osism.commons.facts : Create custom facts directory --------------------- 1.13s 2025-08-29 17:23:37.191500 | orchestrator | Gather facts for all hosts ---------------------------------------------- 0.48s 2025-08-29 17:23:39.266973 | orchestrator | 2025-08-29 17:23:39 | INFO  | Task c4c43eb3-731e-4050-b6df-7993126debd2 (ceph-configure-lvm-volumes) was prepared for execution. 2025-08-29 17:23:39.267092 | orchestrator | 2025-08-29 17:23:39 | INFO  | It takes a moment until task c4c43eb3-731e-4050-b6df-7993126debd2 (ceph-configure-lvm-volumes) has been started and output is visible here. 2025-08-29 17:23:50.766658 | orchestrator | 2025-08-29 17:23:50.766764 | orchestrator | PLAY [Ceph configure LVM] ****************************************************** 2025-08-29 17:23:50.766780 | orchestrator | 2025-08-29 17:23:50.766791 | orchestrator | TASK [Get extra vars for Ceph configuration] *********************************** 2025-08-29 17:23:50.766803 | orchestrator | Friday 29 August 2025 17:23:43 +0000 (0:00:00.321) 0:00:00.321 ********* 2025-08-29 17:23:50.766814 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 17:23:50.766824 | orchestrator | 2025-08-29 17:23:50.766834 | orchestrator | TASK [Get initial list of available block devices] ***************************** 2025-08-29 17:23:50.766844 | orchestrator | Friday 29 August 2025 17:23:43 +0000 (0:00:00.244) 0:00:00.566 ********* 2025-08-29 17:23:50.766854 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:23:50.766865 | orchestrator | 2025-08-29 17:23:50.766874 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.766884 | orchestrator | Friday 29 August 2025 17:23:43 +0000 (0:00:00.226) 0:00:00.792 ********* 2025-08-29 17:23:50.766894 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0) 2025-08-29 17:23:50.766904 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1) 2025-08-29 17:23:50.766913 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2) 2025-08-29 17:23:50.766923 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3) 2025-08-29 17:23:50.766933 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4) 2025-08-29 17:23:50.766943 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5) 2025-08-29 17:23:50.766952 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6) 2025-08-29 17:23:50.766962 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7) 2025-08-29 17:23:50.766972 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda) 2025-08-29 17:23:50.766981 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb) 2025-08-29 17:23:50.766997 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc) 2025-08-29 17:23:50.767025 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd) 2025-08-29 17:23:50.767041 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0) 2025-08-29 17:23:50.767058 | orchestrator | 2025-08-29 17:23:50.767076 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767094 | orchestrator | Friday 29 August 2025 17:23:44 +0000 (0:00:00.348) 0:00:01.140 ********* 2025-08-29 17:23:50.767110 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767126 | orchestrator | 2025-08-29 17:23:50.767159 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767170 | orchestrator | Friday 29 August 2025 17:23:44 +0000 (0:00:00.480) 0:00:01.621 ********* 2025-08-29 17:23:50.767181 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767192 | orchestrator | 2025-08-29 17:23:50.767203 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767214 | orchestrator | Friday 29 August 2025 17:23:44 +0000 (0:00:00.193) 0:00:01.814 ********* 2025-08-29 17:23:50.767226 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767236 | orchestrator | 2025-08-29 17:23:50.767247 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767258 | orchestrator | Friday 29 August 2025 17:23:45 +0000 (0:00:00.190) 0:00:02.005 ********* 2025-08-29 17:23:50.767270 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767307 | orchestrator | 2025-08-29 17:23:50.767324 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767336 | orchestrator | Friday 29 August 2025 17:23:45 +0000 (0:00:00.191) 0:00:02.196 ********* 2025-08-29 17:23:50.767347 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767358 | orchestrator | 2025-08-29 17:23:50.767369 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767381 | orchestrator | Friday 29 August 2025 17:23:45 +0000 (0:00:00.182) 0:00:02.378 ********* 2025-08-29 17:23:50.767392 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767403 | orchestrator | 2025-08-29 17:23:50.767415 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767426 | orchestrator | Friday 29 August 2025 17:23:45 +0000 (0:00:00.186) 0:00:02.565 ********* 2025-08-29 17:23:50.767437 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767448 | orchestrator | 2025-08-29 17:23:50.767459 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767471 | orchestrator | Friday 29 August 2025 17:23:45 +0000 (0:00:00.188) 0:00:02.753 ********* 2025-08-29 17:23:50.767482 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767493 | orchestrator | 2025-08-29 17:23:50.767504 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767515 | orchestrator | Friday 29 August 2025 17:23:45 +0000 (0:00:00.196) 0:00:02.950 ********* 2025-08-29 17:23:50.767526 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1) 2025-08-29 17:23:50.767536 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1) 2025-08-29 17:23:50.767546 | orchestrator | 2025-08-29 17:23:50.767556 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767566 | orchestrator | Friday 29 August 2025 17:23:46 +0000 (0:00:00.409) 0:00:03.359 ********* 2025-08-29 17:23:50.767594 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322) 2025-08-29 17:23:50.767605 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322) 2025-08-29 17:23:50.767614 | orchestrator | 2025-08-29 17:23:50.767624 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767634 | orchestrator | Friday 29 August 2025 17:23:46 +0000 (0:00:00.388) 0:00:03.747 ********* 2025-08-29 17:23:50.767643 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f) 2025-08-29 17:23:50.767653 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f) 2025-08-29 17:23:50.767663 | orchestrator | 2025-08-29 17:23:50.767672 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767682 | orchestrator | Friday 29 August 2025 17:23:47 +0000 (0:00:00.618) 0:00:04.365 ********* 2025-08-29 17:23:50.767691 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5) 2025-08-29 17:23:50.767709 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5) 2025-08-29 17:23:50.767719 | orchestrator | 2025-08-29 17:23:50.767729 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:50.767739 | orchestrator | Friday 29 August 2025 17:23:47 +0000 (0:00:00.615) 0:00:04.981 ********* 2025-08-29 17:23:50.767749 | orchestrator | ok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001) 2025-08-29 17:23:50.767758 | orchestrator | 2025-08-29 17:23:50.767768 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.767784 | orchestrator | Friday 29 August 2025 17:23:48 +0000 (0:00:00.741) 0:00:05.723 ********* 2025-08-29 17:23:50.767795 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0) 2025-08-29 17:23:50.767805 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1) 2025-08-29 17:23:50.767814 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2) 2025-08-29 17:23:50.767824 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3) 2025-08-29 17:23:50.767833 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4) 2025-08-29 17:23:50.767843 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5) 2025-08-29 17:23:50.767853 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6) 2025-08-29 17:23:50.767862 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7) 2025-08-29 17:23:50.767872 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda) 2025-08-29 17:23:50.767881 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb) 2025-08-29 17:23:50.767891 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc) 2025-08-29 17:23:50.767901 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd) 2025-08-29 17:23:50.767910 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0) 2025-08-29 17:23:50.767920 | orchestrator | 2025-08-29 17:23:50.767930 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.767940 | orchestrator | Friday 29 August 2025 17:23:49 +0000 (0:00:00.377) 0:00:06.101 ********* 2025-08-29 17:23:50.767949 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767959 | orchestrator | 2025-08-29 17:23:50.767968 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.767978 | orchestrator | Friday 29 August 2025 17:23:49 +0000 (0:00:00.206) 0:00:06.308 ********* 2025-08-29 17:23:50.767988 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.767997 | orchestrator | 2025-08-29 17:23:50.768007 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.768017 | orchestrator | Friday 29 August 2025 17:23:49 +0000 (0:00:00.193) 0:00:06.501 ********* 2025-08-29 17:23:50.768026 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.768036 | orchestrator | 2025-08-29 17:23:50.768052 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.768069 | orchestrator | Friday 29 August 2025 17:23:49 +0000 (0:00:00.238) 0:00:06.740 ********* 2025-08-29 17:23:50.768086 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.768104 | orchestrator | 2025-08-29 17:23:50.768120 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.768137 | orchestrator | Friday 29 August 2025 17:23:49 +0000 (0:00:00.196) 0:00:06.937 ********* 2025-08-29 17:23:50.768150 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.768160 | orchestrator | 2025-08-29 17:23:50.768169 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.768186 | orchestrator | Friday 29 August 2025 17:23:50 +0000 (0:00:00.203) 0:00:07.141 ********* 2025-08-29 17:23:50.768195 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.768205 | orchestrator | 2025-08-29 17:23:50.768214 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.768224 | orchestrator | Friday 29 August 2025 17:23:50 +0000 (0:00:00.220) 0:00:07.361 ********* 2025-08-29 17:23:50.768233 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:50.768243 | orchestrator | 2025-08-29 17:23:50.768253 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:50.768262 | orchestrator | Friday 29 August 2025 17:23:50 +0000 (0:00:00.192) 0:00:07.554 ********* 2025-08-29 17:23:50.768311 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.030090 | orchestrator | 2025-08-29 17:23:58.030968 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:58.031004 | orchestrator | Friday 29 August 2025 17:23:50 +0000 (0:00:00.202) 0:00:07.756 ********* 2025-08-29 17:23:58.031017 | orchestrator | ok: [testbed-node-3] => (item=sda1) 2025-08-29 17:23:58.031030 | orchestrator | ok: [testbed-node-3] => (item=sda14) 2025-08-29 17:23:58.031041 | orchestrator | ok: [testbed-node-3] => (item=sda15) 2025-08-29 17:23:58.031052 | orchestrator | ok: [testbed-node-3] => (item=sda16) 2025-08-29 17:23:58.031064 | orchestrator | 2025-08-29 17:23:58.031075 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:58.031086 | orchestrator | Friday 29 August 2025 17:23:51 +0000 (0:00:01.030) 0:00:08.787 ********* 2025-08-29 17:23:58.031097 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031108 | orchestrator | 2025-08-29 17:23:58.031120 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:58.031131 | orchestrator | Friday 29 August 2025 17:23:51 +0000 (0:00:00.185) 0:00:08.972 ********* 2025-08-29 17:23:58.031142 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031153 | orchestrator | 2025-08-29 17:23:58.031170 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:58.031188 | orchestrator | Friday 29 August 2025 17:23:52 +0000 (0:00:00.198) 0:00:09.171 ********* 2025-08-29 17:23:58.031199 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031210 | orchestrator | 2025-08-29 17:23:58.031221 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:23:58.031232 | orchestrator | Friday 29 August 2025 17:23:52 +0000 (0:00:00.213) 0:00:09.384 ********* 2025-08-29 17:23:58.031243 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031254 | orchestrator | 2025-08-29 17:23:58.031264 | orchestrator | TASK [Set UUIDs for OSD VGs/LVs] *********************************************** 2025-08-29 17:23:58.031275 | orchestrator | Friday 29 August 2025 17:23:52 +0000 (0:00:00.191) 0:00:09.575 ********* 2025-08-29 17:23:58.031350 | orchestrator | ok: [testbed-node-3] => (item={'key': 'sdb', 'value': None}) 2025-08-29 17:23:58.031363 | orchestrator | ok: [testbed-node-3] => (item={'key': 'sdc', 'value': None}) 2025-08-29 17:23:58.031374 | orchestrator | 2025-08-29 17:23:58.031385 | orchestrator | TASK [Generate WAL VG names] *************************************************** 2025-08-29 17:23:58.031396 | orchestrator | Friday 29 August 2025 17:23:52 +0000 (0:00:00.158) 0:00:09.734 ********* 2025-08-29 17:23:58.031426 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031438 | orchestrator | 2025-08-29 17:23:58.031449 | orchestrator | TASK [Generate DB VG names] **************************************************** 2025-08-29 17:23:58.031460 | orchestrator | Friday 29 August 2025 17:23:52 +0000 (0:00:00.131) 0:00:09.866 ********* 2025-08-29 17:23:58.031471 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031483 | orchestrator | 2025-08-29 17:23:58.031494 | orchestrator | TASK [Generate shared DB/WAL VG names] ***************************************** 2025-08-29 17:23:58.031505 | orchestrator | Friday 29 August 2025 17:23:53 +0000 (0:00:00.140) 0:00:10.006 ********* 2025-08-29 17:23:58.031516 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031527 | orchestrator | 2025-08-29 17:23:58.031570 | orchestrator | TASK [Define lvm_volumes structures] ******************************************* 2025-08-29 17:23:58.031581 | orchestrator | Friday 29 August 2025 17:23:53 +0000 (0:00:00.121) 0:00:10.127 ********* 2025-08-29 17:23:58.031593 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:23:58.031604 | orchestrator | 2025-08-29 17:23:58.031615 | orchestrator | TASK [Generate lvm_volumes structure (block only)] ***************************** 2025-08-29 17:23:58.031626 | orchestrator | Friday 29 August 2025 17:23:53 +0000 (0:00:00.119) 0:00:10.247 ********* 2025-08-29 17:23:58.031637 | orchestrator | ok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}}) 2025-08-29 17:23:58.031655 | orchestrator | ok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '67ff47d9-d75a-55af-b095-c4dbbf8f796a'}}) 2025-08-29 17:23:58.031673 | orchestrator | 2025-08-29 17:23:58.031692 | orchestrator | TASK [Generate lvm_volumes structure (block + db)] ***************************** 2025-08-29 17:23:58.031710 | orchestrator | Friday 29 August 2025 17:23:53 +0000 (0:00:00.149) 0:00:10.396 ********* 2025-08-29 17:23:58.031728 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}})  2025-08-29 17:23:58.031760 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '67ff47d9-d75a-55af-b095-c4dbbf8f796a'}})  2025-08-29 17:23:58.031784 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031803 | orchestrator | 2025-08-29 17:23:58.031821 | orchestrator | TASK [Generate lvm_volumes structure (block + wal)] **************************** 2025-08-29 17:23:58.031841 | orchestrator | Friday 29 August 2025 17:23:53 +0000 (0:00:00.149) 0:00:10.546 ********* 2025-08-29 17:23:58.031860 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}})  2025-08-29 17:23:58.031878 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '67ff47d9-d75a-55af-b095-c4dbbf8f796a'}})  2025-08-29 17:23:58.031894 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031905 | orchestrator | 2025-08-29 17:23:58.031916 | orchestrator | TASK [Generate lvm_volumes structure (block + db + wal)] *********************** 2025-08-29 17:23:58.031927 | orchestrator | Friday 29 August 2025 17:23:53 +0000 (0:00:00.144) 0:00:10.690 ********* 2025-08-29 17:23:58.031937 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}})  2025-08-29 17:23:58.031948 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '67ff47d9-d75a-55af-b095-c4dbbf8f796a'}})  2025-08-29 17:23:58.031959 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.031969 | orchestrator | 2025-08-29 17:23:58.032000 | orchestrator | TASK [Compile lvm_volumes] ***************************************************** 2025-08-29 17:23:58.032012 | orchestrator | Friday 29 August 2025 17:23:53 +0000 (0:00:00.308) 0:00:10.998 ********* 2025-08-29 17:23:58.032022 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:23:58.032033 | orchestrator | 2025-08-29 17:23:58.032052 | orchestrator | TASK [Set OSD devices config data] ********************************************* 2025-08-29 17:23:58.032121 | orchestrator | Friday 29 August 2025 17:23:54 +0000 (0:00:00.142) 0:00:11.141 ********* 2025-08-29 17:23:58.032133 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:23:58.032144 | orchestrator | 2025-08-29 17:23:58.032155 | orchestrator | TASK [Set DB devices config data] ********************************************** 2025-08-29 17:23:58.032166 | orchestrator | Friday 29 August 2025 17:23:54 +0000 (0:00:00.140) 0:00:11.281 ********* 2025-08-29 17:23:58.032177 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.032188 | orchestrator | 2025-08-29 17:23:58.032198 | orchestrator | TASK [Set WAL devices config data] ********************************************* 2025-08-29 17:23:58.032209 | orchestrator | Friday 29 August 2025 17:23:54 +0000 (0:00:00.127) 0:00:11.409 ********* 2025-08-29 17:23:58.032220 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.032231 | orchestrator | 2025-08-29 17:23:58.032241 | orchestrator | TASK [Set DB+WAL devices config data] ****************************************** 2025-08-29 17:23:58.032264 | orchestrator | Friday 29 August 2025 17:23:54 +0000 (0:00:00.132) 0:00:11.542 ********* 2025-08-29 17:23:58.032340 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.032352 | orchestrator | 2025-08-29 17:23:58.032363 | orchestrator | TASK [Print ceph_osd_devices] ************************************************** 2025-08-29 17:23:58.032374 | orchestrator | Friday 29 August 2025 17:23:54 +0000 (0:00:00.141) 0:00:11.683 ********* 2025-08-29 17:23:58.032392 | orchestrator | ok: [testbed-node-3] => { 2025-08-29 17:23:58.032411 | orchestrator |  "ceph_osd_devices": { 2025-08-29 17:23:58.032429 | orchestrator |  "sdb": { 2025-08-29 17:23:58.032447 | orchestrator |  "osd_lvm_uuid": "25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b" 2025-08-29 17:23:58.032465 | orchestrator |  }, 2025-08-29 17:23:58.032483 | orchestrator |  "sdc": { 2025-08-29 17:23:58.032501 | orchestrator |  "osd_lvm_uuid": "67ff47d9-d75a-55af-b095-c4dbbf8f796a" 2025-08-29 17:23:58.032521 | orchestrator |  } 2025-08-29 17:23:58.032539 | orchestrator |  } 2025-08-29 17:23:58.032555 | orchestrator | } 2025-08-29 17:23:58.032566 | orchestrator | 2025-08-29 17:23:58.032578 | orchestrator | TASK [Print WAL devices] ******************************************************* 2025-08-29 17:23:58.032589 | orchestrator | Friday 29 August 2025 17:23:54 +0000 (0:00:00.131) 0:00:11.815 ********* 2025-08-29 17:23:58.032600 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.032611 | orchestrator | 2025-08-29 17:23:58.032622 | orchestrator | TASK [Print DB devices] ******************************************************** 2025-08-29 17:23:58.032632 | orchestrator | Friday 29 August 2025 17:23:54 +0000 (0:00:00.123) 0:00:11.939 ********* 2025-08-29 17:23:58.032643 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.032654 | orchestrator | 2025-08-29 17:23:58.032665 | orchestrator | TASK [Print shared DB/WAL devices] ********************************************* 2025-08-29 17:23:58.032676 | orchestrator | Friday 29 August 2025 17:23:55 +0000 (0:00:00.116) 0:00:12.055 ********* 2025-08-29 17:23:58.032687 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:23:58.032697 | orchestrator | 2025-08-29 17:23:58.032708 | orchestrator | TASK [Print configuration data] ************************************************ 2025-08-29 17:23:58.032719 | orchestrator | Friday 29 August 2025 17:23:55 +0000 (0:00:00.135) 0:00:12.190 ********* 2025-08-29 17:23:58.032730 | orchestrator | changed: [testbed-node-3] => { 2025-08-29 17:23:58.032741 | orchestrator |  "_ceph_configure_lvm_config_data": { 2025-08-29 17:23:58.032752 | orchestrator |  "ceph_osd_devices": { 2025-08-29 17:23:58.032763 | orchestrator |  "sdb": { 2025-08-29 17:23:58.032774 | orchestrator |  "osd_lvm_uuid": "25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b" 2025-08-29 17:23:58.032785 | orchestrator |  }, 2025-08-29 17:23:58.032796 | orchestrator |  "sdc": { 2025-08-29 17:23:58.032807 | orchestrator |  "osd_lvm_uuid": "67ff47d9-d75a-55af-b095-c4dbbf8f796a" 2025-08-29 17:23:58.032818 | orchestrator |  } 2025-08-29 17:23:58.032829 | orchestrator |  }, 2025-08-29 17:23:58.032840 | orchestrator |  "lvm_volumes": [ 2025-08-29 17:23:58.032850 | orchestrator |  { 2025-08-29 17:23:58.032861 | orchestrator |  "data": "osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b", 2025-08-29 17:23:58.032874 | orchestrator |  "data_vg": "ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b" 2025-08-29 17:23:58.032893 | orchestrator |  }, 2025-08-29 17:23:58.032911 | orchestrator |  { 2025-08-29 17:23:58.032934 | orchestrator |  "data": "osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a", 2025-08-29 17:23:58.032958 | orchestrator |  "data_vg": "ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a" 2025-08-29 17:23:58.032976 | orchestrator |  } 2025-08-29 17:23:58.032995 | orchestrator |  ] 2025-08-29 17:23:58.033013 | orchestrator |  } 2025-08-29 17:23:58.033032 | orchestrator | } 2025-08-29 17:23:58.033044 | orchestrator | 2025-08-29 17:23:58.033062 | orchestrator | RUNNING HANDLER [Write configuration file] ************************************* 2025-08-29 17:23:58.033074 | orchestrator | Friday 29 August 2025 17:23:55 +0000 (0:00:00.188) 0:00:12.379 ********* 2025-08-29 17:23:58.033177 | orchestrator | changed: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 17:23:58.033191 | orchestrator | 2025-08-29 17:23:58.033202 | orchestrator | PLAY [Ceph configure LVM] ****************************************************** 2025-08-29 17:23:58.033214 | orchestrator | 2025-08-29 17:23:58.033225 | orchestrator | TASK [Get extra vars for Ceph configuration] *********************************** 2025-08-29 17:23:58.033236 | orchestrator | Friday 29 August 2025 17:23:57 +0000 (0:00:02.114) 0:00:14.493 ********* 2025-08-29 17:23:58.033247 | orchestrator | ok: [testbed-node-4 -> testbed-manager(192.168.16.5)] 2025-08-29 17:23:58.033258 | orchestrator | 2025-08-29 17:23:58.033269 | orchestrator | TASK [Get initial list of available block devices] ***************************** 2025-08-29 17:23:58.033299 | orchestrator | Friday 29 August 2025 17:23:57 +0000 (0:00:00.268) 0:00:14.762 ********* 2025-08-29 17:23:58.033311 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:23:58.033322 | orchestrator | 2025-08-29 17:23:58.033333 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:23:58.033356 | orchestrator | Friday 29 August 2025 17:23:58 +0000 (0:00:00.254) 0:00:15.017 ********* 2025-08-29 17:24:05.676744 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0) 2025-08-29 17:24:05.676849 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1) 2025-08-29 17:24:05.676864 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2) 2025-08-29 17:24:05.676875 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3) 2025-08-29 17:24:05.676887 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4) 2025-08-29 17:24:05.676898 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5) 2025-08-29 17:24:05.676908 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6) 2025-08-29 17:24:05.676919 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7) 2025-08-29 17:24:05.676930 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda) 2025-08-29 17:24:05.676941 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb) 2025-08-29 17:24:05.676952 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc) 2025-08-29 17:24:05.676981 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd) 2025-08-29 17:24:05.676992 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0) 2025-08-29 17:24:05.677014 | orchestrator | 2025-08-29 17:24:05.677031 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677043 | orchestrator | Friday 29 August 2025 17:23:58 +0000 (0:00:00.398) 0:00:15.416 ********* 2025-08-29 17:24:05.677055 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677067 | orchestrator | 2025-08-29 17:24:05.677078 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677089 | orchestrator | Friday 29 August 2025 17:23:58 +0000 (0:00:00.187) 0:00:15.603 ********* 2025-08-29 17:24:05.677099 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677110 | orchestrator | 2025-08-29 17:24:05.677120 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677131 | orchestrator | Friday 29 August 2025 17:23:58 +0000 (0:00:00.189) 0:00:15.792 ********* 2025-08-29 17:24:05.677142 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677153 | orchestrator | 2025-08-29 17:24:05.677163 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677174 | orchestrator | Friday 29 August 2025 17:23:58 +0000 (0:00:00.191) 0:00:15.984 ********* 2025-08-29 17:24:05.677185 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677196 | orchestrator | 2025-08-29 17:24:05.677230 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677242 | orchestrator | Friday 29 August 2025 17:23:59 +0000 (0:00:00.205) 0:00:16.190 ********* 2025-08-29 17:24:05.677253 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677263 | orchestrator | 2025-08-29 17:24:05.677274 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677306 | orchestrator | Friday 29 August 2025 17:23:59 +0000 (0:00:00.200) 0:00:16.391 ********* 2025-08-29 17:24:05.677319 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677331 | orchestrator | 2025-08-29 17:24:05.677344 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677373 | orchestrator | Friday 29 August 2025 17:23:59 +0000 (0:00:00.523) 0:00:16.914 ********* 2025-08-29 17:24:05.677386 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677398 | orchestrator | 2025-08-29 17:24:05.677411 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677422 | orchestrator | Friday 29 August 2025 17:24:00 +0000 (0:00:00.195) 0:00:17.109 ********* 2025-08-29 17:24:05.677432 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677443 | orchestrator | 2025-08-29 17:24:05.677454 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677464 | orchestrator | Friday 29 August 2025 17:24:00 +0000 (0:00:00.211) 0:00:17.320 ********* 2025-08-29 17:24:05.677475 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320) 2025-08-29 17:24:05.677487 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320) 2025-08-29 17:24:05.677498 | orchestrator | 2025-08-29 17:24:05.677508 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677519 | orchestrator | Friday 29 August 2025 17:24:00 +0000 (0:00:00.391) 0:00:17.712 ********* 2025-08-29 17:24:05.677530 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3) 2025-08-29 17:24:05.677540 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3) 2025-08-29 17:24:05.677551 | orchestrator | 2025-08-29 17:24:05.677561 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677572 | orchestrator | Friday 29 August 2025 17:24:01 +0000 (0:00:00.395) 0:00:18.108 ********* 2025-08-29 17:24:05.677582 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6) 2025-08-29 17:24:05.677593 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6) 2025-08-29 17:24:05.677603 | orchestrator | 2025-08-29 17:24:05.677614 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677625 | orchestrator | Friday 29 August 2025 17:24:01 +0000 (0:00:00.445) 0:00:18.554 ********* 2025-08-29 17:24:05.677651 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474) 2025-08-29 17:24:05.677663 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474) 2025-08-29 17:24:05.677674 | orchestrator | 2025-08-29 17:24:05.677685 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:05.677696 | orchestrator | Friday 29 August 2025 17:24:01 +0000 (0:00:00.410) 0:00:18.964 ********* 2025-08-29 17:24:05.677707 | orchestrator | ok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001) 2025-08-29 17:24:05.677717 | orchestrator | 2025-08-29 17:24:05.677728 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.677739 | orchestrator | Friday 29 August 2025 17:24:02 +0000 (0:00:00.321) 0:00:19.286 ********* 2025-08-29 17:24:05.677750 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0) 2025-08-29 17:24:05.677760 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1) 2025-08-29 17:24:05.677780 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2) 2025-08-29 17:24:05.677791 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3) 2025-08-29 17:24:05.677802 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4) 2025-08-29 17:24:05.677812 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5) 2025-08-29 17:24:05.677823 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6) 2025-08-29 17:24:05.677834 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7) 2025-08-29 17:24:05.677845 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda) 2025-08-29 17:24:05.677855 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb) 2025-08-29 17:24:05.677866 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc) 2025-08-29 17:24:05.677877 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd) 2025-08-29 17:24:05.677888 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0) 2025-08-29 17:24:05.677898 | orchestrator | 2025-08-29 17:24:05.677909 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.677920 | orchestrator | Friday 29 August 2025 17:24:02 +0000 (0:00:00.393) 0:00:19.679 ********* 2025-08-29 17:24:05.677931 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677942 | orchestrator | 2025-08-29 17:24:05.677953 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.677963 | orchestrator | Friday 29 August 2025 17:24:02 +0000 (0:00:00.203) 0:00:19.883 ********* 2025-08-29 17:24:05.677974 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.677985 | orchestrator | 2025-08-29 17:24:05.678009 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678086 | orchestrator | Friday 29 August 2025 17:24:03 +0000 (0:00:00.599) 0:00:20.483 ********* 2025-08-29 17:24:05.678098 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.678109 | orchestrator | 2025-08-29 17:24:05.678119 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678130 | orchestrator | Friday 29 August 2025 17:24:03 +0000 (0:00:00.196) 0:00:20.679 ********* 2025-08-29 17:24:05.678141 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.678152 | orchestrator | 2025-08-29 17:24:05.678162 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678173 | orchestrator | Friday 29 August 2025 17:24:03 +0000 (0:00:00.217) 0:00:20.897 ********* 2025-08-29 17:24:05.678184 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.678195 | orchestrator | 2025-08-29 17:24:05.678206 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678216 | orchestrator | Friday 29 August 2025 17:24:04 +0000 (0:00:00.194) 0:00:21.091 ********* 2025-08-29 17:24:05.678227 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.678238 | orchestrator | 2025-08-29 17:24:05.678248 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678259 | orchestrator | Friday 29 August 2025 17:24:04 +0000 (0:00:00.204) 0:00:21.295 ********* 2025-08-29 17:24:05.678270 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.678307 | orchestrator | 2025-08-29 17:24:05.678320 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678330 | orchestrator | Friday 29 August 2025 17:24:04 +0000 (0:00:00.199) 0:00:21.494 ********* 2025-08-29 17:24:05.678341 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.678352 | orchestrator | 2025-08-29 17:24:05.678362 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678380 | orchestrator | Friday 29 August 2025 17:24:04 +0000 (0:00:00.228) 0:00:21.723 ********* 2025-08-29 17:24:05.678391 | orchestrator | ok: [testbed-node-4] => (item=sda1) 2025-08-29 17:24:05.678403 | orchestrator | ok: [testbed-node-4] => (item=sda14) 2025-08-29 17:24:05.678414 | orchestrator | ok: [testbed-node-4] => (item=sda15) 2025-08-29 17:24:05.678424 | orchestrator | ok: [testbed-node-4] => (item=sda16) 2025-08-29 17:24:05.678435 | orchestrator | 2025-08-29 17:24:05.678446 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:05.678456 | orchestrator | Friday 29 August 2025 17:24:05 +0000 (0:00:00.699) 0:00:22.423 ********* 2025-08-29 17:24:05.678467 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:05.678478 | orchestrator | 2025-08-29 17:24:05.678497 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:11.633021 | orchestrator | Friday 29 August 2025 17:24:05 +0000 (0:00:00.242) 0:00:22.666 ********* 2025-08-29 17:24:11.633094 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633105 | orchestrator | 2025-08-29 17:24:11.633113 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:11.633121 | orchestrator | Friday 29 August 2025 17:24:05 +0000 (0:00:00.211) 0:00:22.877 ********* 2025-08-29 17:24:11.633128 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633135 | orchestrator | 2025-08-29 17:24:11.633143 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:11.633150 | orchestrator | Friday 29 August 2025 17:24:06 +0000 (0:00:00.205) 0:00:23.083 ********* 2025-08-29 17:24:11.633157 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633164 | orchestrator | 2025-08-29 17:24:11.633171 | orchestrator | TASK [Set UUIDs for OSD VGs/LVs] *********************************************** 2025-08-29 17:24:11.633179 | orchestrator | Friday 29 August 2025 17:24:06 +0000 (0:00:00.205) 0:00:23.288 ********* 2025-08-29 17:24:11.633186 | orchestrator | ok: [testbed-node-4] => (item={'key': 'sdb', 'value': None}) 2025-08-29 17:24:11.633193 | orchestrator | ok: [testbed-node-4] => (item={'key': 'sdc', 'value': None}) 2025-08-29 17:24:11.633200 | orchestrator | 2025-08-29 17:24:11.633207 | orchestrator | TASK [Generate WAL VG names] *************************************************** 2025-08-29 17:24:11.633214 | orchestrator | Friday 29 August 2025 17:24:06 +0000 (0:00:00.375) 0:00:23.664 ********* 2025-08-29 17:24:11.633222 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633229 | orchestrator | 2025-08-29 17:24:11.633236 | orchestrator | TASK [Generate DB VG names] **************************************************** 2025-08-29 17:24:11.633243 | orchestrator | Friday 29 August 2025 17:24:06 +0000 (0:00:00.138) 0:00:23.802 ********* 2025-08-29 17:24:11.633251 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633258 | orchestrator | 2025-08-29 17:24:11.633265 | orchestrator | TASK [Generate shared DB/WAL VG names] ***************************************** 2025-08-29 17:24:11.633272 | orchestrator | Friday 29 August 2025 17:24:06 +0000 (0:00:00.140) 0:00:23.942 ********* 2025-08-29 17:24:11.633279 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633336 | orchestrator | 2025-08-29 17:24:11.633346 | orchestrator | TASK [Define lvm_volumes structures] ******************************************* 2025-08-29 17:24:11.633353 | orchestrator | Friday 29 August 2025 17:24:07 +0000 (0:00:00.131) 0:00:24.074 ********* 2025-08-29 17:24:11.633360 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:24:11.633368 | orchestrator | 2025-08-29 17:24:11.633375 | orchestrator | TASK [Generate lvm_volumes structure (block only)] ***************************** 2025-08-29 17:24:11.633382 | orchestrator | Friday 29 August 2025 17:24:07 +0000 (0:00:00.127) 0:00:24.201 ********* 2025-08-29 17:24:11.633390 | orchestrator | ok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b89384ec-5219-5f2a-8735-84f78c8179d2'}}) 2025-08-29 17:24:11.633397 | orchestrator | ok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '860f9296-4089-56f8-8238-0f24b03dbca2'}}) 2025-08-29 17:24:11.633404 | orchestrator | 2025-08-29 17:24:11.633412 | orchestrator | TASK [Generate lvm_volumes structure (block + db)] ***************************** 2025-08-29 17:24:11.633435 | orchestrator | Friday 29 August 2025 17:24:07 +0000 (0:00:00.144) 0:00:24.345 ********* 2025-08-29 17:24:11.633443 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b89384ec-5219-5f2a-8735-84f78c8179d2'}})  2025-08-29 17:24:11.633452 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '860f9296-4089-56f8-8238-0f24b03dbca2'}})  2025-08-29 17:24:11.633459 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633466 | orchestrator | 2025-08-29 17:24:11.633484 | orchestrator | TASK [Generate lvm_volumes structure (block + wal)] **************************** 2025-08-29 17:24:11.633492 | orchestrator | Friday 29 August 2025 17:24:07 +0000 (0:00:00.139) 0:00:24.485 ********* 2025-08-29 17:24:11.633499 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b89384ec-5219-5f2a-8735-84f78c8179d2'}})  2025-08-29 17:24:11.633506 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '860f9296-4089-56f8-8238-0f24b03dbca2'}})  2025-08-29 17:24:11.633513 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633520 | orchestrator | 2025-08-29 17:24:11.633527 | orchestrator | TASK [Generate lvm_volumes structure (block + db + wal)] *********************** 2025-08-29 17:24:11.633534 | orchestrator | Friday 29 August 2025 17:24:07 +0000 (0:00:00.150) 0:00:24.635 ********* 2025-08-29 17:24:11.633541 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b89384ec-5219-5f2a-8735-84f78c8179d2'}})  2025-08-29 17:24:11.633548 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '860f9296-4089-56f8-8238-0f24b03dbca2'}})  2025-08-29 17:24:11.633555 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633562 | orchestrator | 2025-08-29 17:24:11.633569 | orchestrator | TASK [Compile lvm_volumes] ***************************************************** 2025-08-29 17:24:11.633577 | orchestrator | Friday 29 August 2025 17:24:07 +0000 (0:00:00.138) 0:00:24.774 ********* 2025-08-29 17:24:11.633584 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:24:11.633591 | orchestrator | 2025-08-29 17:24:11.633599 | orchestrator | TASK [Set OSD devices config data] ********************************************* 2025-08-29 17:24:11.633607 | orchestrator | Friday 29 August 2025 17:24:07 +0000 (0:00:00.135) 0:00:24.910 ********* 2025-08-29 17:24:11.633615 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:24:11.633623 | orchestrator | 2025-08-29 17:24:11.633631 | orchestrator | TASK [Set DB devices config data] ********************************************** 2025-08-29 17:24:11.633639 | orchestrator | Friday 29 August 2025 17:24:08 +0000 (0:00:00.131) 0:00:25.041 ********* 2025-08-29 17:24:11.633647 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633655 | orchestrator | 2025-08-29 17:24:11.633675 | orchestrator | TASK [Set WAL devices config data] ********************************************* 2025-08-29 17:24:11.633684 | orchestrator | Friday 29 August 2025 17:24:08 +0000 (0:00:00.126) 0:00:25.168 ********* 2025-08-29 17:24:11.633692 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633700 | orchestrator | 2025-08-29 17:24:11.633708 | orchestrator | TASK [Set DB+WAL devices config data] ****************************************** 2025-08-29 17:24:11.633716 | orchestrator | Friday 29 August 2025 17:24:08 +0000 (0:00:00.330) 0:00:25.498 ********* 2025-08-29 17:24:11.633723 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633731 | orchestrator | 2025-08-29 17:24:11.633739 | orchestrator | TASK [Print ceph_osd_devices] ************************************************** 2025-08-29 17:24:11.633747 | orchestrator | Friday 29 August 2025 17:24:08 +0000 (0:00:00.129) 0:00:25.628 ********* 2025-08-29 17:24:11.633755 | orchestrator | ok: [testbed-node-4] => { 2025-08-29 17:24:11.633763 | orchestrator |  "ceph_osd_devices": { 2025-08-29 17:24:11.633771 | orchestrator |  "sdb": { 2025-08-29 17:24:11.633779 | orchestrator |  "osd_lvm_uuid": "b89384ec-5219-5f2a-8735-84f78c8179d2" 2025-08-29 17:24:11.633787 | orchestrator |  }, 2025-08-29 17:24:11.633795 | orchestrator |  "sdc": { 2025-08-29 17:24:11.633803 | orchestrator |  "osd_lvm_uuid": "860f9296-4089-56f8-8238-0f24b03dbca2" 2025-08-29 17:24:11.633816 | orchestrator |  } 2025-08-29 17:24:11.633824 | orchestrator |  } 2025-08-29 17:24:11.633832 | orchestrator | } 2025-08-29 17:24:11.633840 | orchestrator | 2025-08-29 17:24:11.633848 | orchestrator | TASK [Print WAL devices] ******************************************************* 2025-08-29 17:24:11.633857 | orchestrator | Friday 29 August 2025 17:24:08 +0000 (0:00:00.149) 0:00:25.777 ********* 2025-08-29 17:24:11.633865 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633873 | orchestrator | 2025-08-29 17:24:11.633881 | orchestrator | TASK [Print DB devices] ******************************************************** 2025-08-29 17:24:11.633889 | orchestrator | Friday 29 August 2025 17:24:08 +0000 (0:00:00.127) 0:00:25.905 ********* 2025-08-29 17:24:11.633898 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633906 | orchestrator | 2025-08-29 17:24:11.633914 | orchestrator | TASK [Print shared DB/WAL devices] ********************************************* 2025-08-29 17:24:11.633922 | orchestrator | Friday 29 August 2025 17:24:09 +0000 (0:00:00.122) 0:00:26.028 ********* 2025-08-29 17:24:11.633930 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:24:11.633938 | orchestrator | 2025-08-29 17:24:11.633946 | orchestrator | TASK [Print configuration data] ************************************************ 2025-08-29 17:24:11.633954 | orchestrator | Friday 29 August 2025 17:24:09 +0000 (0:00:00.128) 0:00:26.157 ********* 2025-08-29 17:24:11.633961 | orchestrator | changed: [testbed-node-4] => { 2025-08-29 17:24:11.633968 | orchestrator |  "_ceph_configure_lvm_config_data": { 2025-08-29 17:24:11.633975 | orchestrator |  "ceph_osd_devices": { 2025-08-29 17:24:11.633982 | orchestrator |  "sdb": { 2025-08-29 17:24:11.633989 | orchestrator |  "osd_lvm_uuid": "b89384ec-5219-5f2a-8735-84f78c8179d2" 2025-08-29 17:24:11.633996 | orchestrator |  }, 2025-08-29 17:24:11.634003 | orchestrator |  "sdc": { 2025-08-29 17:24:11.634011 | orchestrator |  "osd_lvm_uuid": "860f9296-4089-56f8-8238-0f24b03dbca2" 2025-08-29 17:24:11.634058 | orchestrator |  } 2025-08-29 17:24:11.634066 | orchestrator |  }, 2025-08-29 17:24:11.634073 | orchestrator |  "lvm_volumes": [ 2025-08-29 17:24:11.634080 | orchestrator |  { 2025-08-29 17:24:11.634087 | orchestrator |  "data": "osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2", 2025-08-29 17:24:11.634094 | orchestrator |  "data_vg": "ceph-b89384ec-5219-5f2a-8735-84f78c8179d2" 2025-08-29 17:24:11.634101 | orchestrator |  }, 2025-08-29 17:24:11.634109 | orchestrator |  { 2025-08-29 17:24:11.634116 | orchestrator |  "data": "osd-block-860f9296-4089-56f8-8238-0f24b03dbca2", 2025-08-29 17:24:11.634123 | orchestrator |  "data_vg": "ceph-860f9296-4089-56f8-8238-0f24b03dbca2" 2025-08-29 17:24:11.634130 | orchestrator |  } 2025-08-29 17:24:11.634137 | orchestrator |  ] 2025-08-29 17:24:11.634144 | orchestrator |  } 2025-08-29 17:24:11.634151 | orchestrator | } 2025-08-29 17:24:11.634158 | orchestrator | 2025-08-29 17:24:11.634165 | orchestrator | RUNNING HANDLER [Write configuration file] ************************************* 2025-08-29 17:24:11.634172 | orchestrator | Friday 29 August 2025 17:24:09 +0000 (0:00:00.185) 0:00:26.343 ********* 2025-08-29 17:24:11.634179 | orchestrator | changed: [testbed-node-4 -> testbed-manager(192.168.16.5)] 2025-08-29 17:24:11.634186 | orchestrator | 2025-08-29 17:24:11.634193 | orchestrator | PLAY [Ceph configure LVM] ****************************************************** 2025-08-29 17:24:11.634200 | orchestrator | 2025-08-29 17:24:11.634207 | orchestrator | TASK [Get extra vars for Ceph configuration] *********************************** 2025-08-29 17:24:11.634215 | orchestrator | Friday 29 August 2025 17:24:10 +0000 (0:00:01.083) 0:00:27.426 ********* 2025-08-29 17:24:11.634222 | orchestrator | ok: [testbed-node-5 -> testbed-manager(192.168.16.5)] 2025-08-29 17:24:11.634229 | orchestrator | 2025-08-29 17:24:11.634236 | orchestrator | TASK [Get initial list of available block devices] ***************************** 2025-08-29 17:24:11.634243 | orchestrator | Friday 29 August 2025 17:24:10 +0000 (0:00:00.432) 0:00:27.858 ********* 2025-08-29 17:24:11.634250 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:24:11.634261 | orchestrator | 2025-08-29 17:24:11.634272 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:11.634280 | orchestrator | Friday 29 August 2025 17:24:11 +0000 (0:00:00.436) 0:00:28.295 ********* 2025-08-29 17:24:11.634303 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0) 2025-08-29 17:24:11.634310 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1) 2025-08-29 17:24:11.634317 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2) 2025-08-29 17:24:11.634324 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3) 2025-08-29 17:24:11.634331 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4) 2025-08-29 17:24:11.634338 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5) 2025-08-29 17:24:11.634350 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6) 2025-08-29 17:24:18.570150 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7) 2025-08-29 17:24:18.570238 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda) 2025-08-29 17:24:18.570253 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb) 2025-08-29 17:24:18.570264 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc) 2025-08-29 17:24:18.570275 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd) 2025-08-29 17:24:18.570322 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0) 2025-08-29 17:24:18.570334 | orchestrator | 2025-08-29 17:24:18.570347 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570359 | orchestrator | Friday 29 August 2025 17:24:11 +0000 (0:00:00.331) 0:00:28.627 ********* 2025-08-29 17:24:18.570370 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570381 | orchestrator | 2025-08-29 17:24:18.570392 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570404 | orchestrator | Friday 29 August 2025 17:24:11 +0000 (0:00:00.162) 0:00:28.789 ********* 2025-08-29 17:24:18.570414 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570425 | orchestrator | 2025-08-29 17:24:18.570436 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570447 | orchestrator | Friday 29 August 2025 17:24:11 +0000 (0:00:00.164) 0:00:28.954 ********* 2025-08-29 17:24:18.570458 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570469 | orchestrator | 2025-08-29 17:24:18.570479 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570490 | orchestrator | Friday 29 August 2025 17:24:12 +0000 (0:00:00.170) 0:00:29.124 ********* 2025-08-29 17:24:18.570501 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570512 | orchestrator | 2025-08-29 17:24:18.570522 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570533 | orchestrator | Friday 29 August 2025 17:24:12 +0000 (0:00:00.162) 0:00:29.287 ********* 2025-08-29 17:24:18.570544 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570555 | orchestrator | 2025-08-29 17:24:18.570565 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570576 | orchestrator | Friday 29 August 2025 17:24:12 +0000 (0:00:00.159) 0:00:29.447 ********* 2025-08-29 17:24:18.570587 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570597 | orchestrator | 2025-08-29 17:24:18.570608 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570619 | orchestrator | Friday 29 August 2025 17:24:12 +0000 (0:00:00.131) 0:00:29.579 ********* 2025-08-29 17:24:18.570630 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570643 | orchestrator | 2025-08-29 17:24:18.570677 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570690 | orchestrator | Friday 29 August 2025 17:24:12 +0000 (0:00:00.142) 0:00:29.722 ********* 2025-08-29 17:24:18.570702 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.570714 | orchestrator | 2025-08-29 17:24:18.570726 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570739 | orchestrator | Friday 29 August 2025 17:24:12 +0000 (0:00:00.145) 0:00:29.867 ********* 2025-08-29 17:24:18.570751 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f) 2025-08-29 17:24:18.570763 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f) 2025-08-29 17:24:18.570776 | orchestrator | 2025-08-29 17:24:18.570789 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570801 | orchestrator | Friday 29 August 2025 17:24:13 +0000 (0:00:00.490) 0:00:30.358 ********* 2025-08-29 17:24:18.570813 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6) 2025-08-29 17:24:18.570825 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6) 2025-08-29 17:24:18.570836 | orchestrator | 2025-08-29 17:24:18.570847 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570857 | orchestrator | Friday 29 August 2025 17:24:14 +0000 (0:00:00.676) 0:00:31.035 ********* 2025-08-29 17:24:18.570868 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6) 2025-08-29 17:24:18.570879 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6) 2025-08-29 17:24:18.570889 | orchestrator | 2025-08-29 17:24:18.570900 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570911 | orchestrator | Friday 29 August 2025 17:24:14 +0000 (0:00:00.387) 0:00:31.423 ********* 2025-08-29 17:24:18.570921 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290) 2025-08-29 17:24:18.570932 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290) 2025-08-29 17:24:18.570942 | orchestrator | 2025-08-29 17:24:18.570953 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:24:18.570964 | orchestrator | Friday 29 August 2025 17:24:14 +0000 (0:00:00.393) 0:00:31.817 ********* 2025-08-29 17:24:18.570974 | orchestrator | ok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001) 2025-08-29 17:24:18.570985 | orchestrator | 2025-08-29 17:24:18.570996 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571006 | orchestrator | Friday 29 August 2025 17:24:15 +0000 (0:00:00.251) 0:00:32.068 ********* 2025-08-29 17:24:18.571032 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0) 2025-08-29 17:24:18.571044 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1) 2025-08-29 17:24:18.571055 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2) 2025-08-29 17:24:18.571066 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3) 2025-08-29 17:24:18.571076 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4) 2025-08-29 17:24:18.571087 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5) 2025-08-29 17:24:18.571111 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6) 2025-08-29 17:24:18.571123 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7) 2025-08-29 17:24:18.571133 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda) 2025-08-29 17:24:18.571152 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb) 2025-08-29 17:24:18.571163 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc) 2025-08-29 17:24:18.571173 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd) 2025-08-29 17:24:18.571184 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0) 2025-08-29 17:24:18.571195 | orchestrator | 2025-08-29 17:24:18.571206 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571216 | orchestrator | Friday 29 August 2025 17:24:15 +0000 (0:00:00.347) 0:00:32.415 ********* 2025-08-29 17:24:18.571227 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571238 | orchestrator | 2025-08-29 17:24:18.571249 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571259 | orchestrator | Friday 29 August 2025 17:24:15 +0000 (0:00:00.183) 0:00:32.599 ********* 2025-08-29 17:24:18.571270 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571281 | orchestrator | 2025-08-29 17:24:18.571325 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571337 | orchestrator | Friday 29 August 2025 17:24:15 +0000 (0:00:00.153) 0:00:32.752 ********* 2025-08-29 17:24:18.571347 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571358 | orchestrator | 2025-08-29 17:24:18.571369 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571385 | orchestrator | Friday 29 August 2025 17:24:15 +0000 (0:00:00.190) 0:00:32.943 ********* 2025-08-29 17:24:18.571396 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571433 | orchestrator | 2025-08-29 17:24:18.571444 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571455 | orchestrator | Friday 29 August 2025 17:24:16 +0000 (0:00:00.198) 0:00:33.142 ********* 2025-08-29 17:24:18.571466 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571476 | orchestrator | 2025-08-29 17:24:18.571487 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571498 | orchestrator | Friday 29 August 2025 17:24:16 +0000 (0:00:00.206) 0:00:33.348 ********* 2025-08-29 17:24:18.571509 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571519 | orchestrator | 2025-08-29 17:24:18.571530 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571541 | orchestrator | Friday 29 August 2025 17:24:16 +0000 (0:00:00.529) 0:00:33.878 ********* 2025-08-29 17:24:18.571552 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571563 | orchestrator | 2025-08-29 17:24:18.571573 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571584 | orchestrator | Friday 29 August 2025 17:24:17 +0000 (0:00:00.189) 0:00:34.067 ********* 2025-08-29 17:24:18.571595 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571606 | orchestrator | 2025-08-29 17:24:18.571616 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571627 | orchestrator | Friday 29 August 2025 17:24:17 +0000 (0:00:00.202) 0:00:34.270 ********* 2025-08-29 17:24:18.571638 | orchestrator | ok: [testbed-node-5] => (item=sda1) 2025-08-29 17:24:18.571649 | orchestrator | ok: [testbed-node-5] => (item=sda14) 2025-08-29 17:24:18.571659 | orchestrator | ok: [testbed-node-5] => (item=sda15) 2025-08-29 17:24:18.571670 | orchestrator | ok: [testbed-node-5] => (item=sda16) 2025-08-29 17:24:18.571681 | orchestrator | 2025-08-29 17:24:18.571692 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571702 | orchestrator | Friday 29 August 2025 17:24:17 +0000 (0:00:00.572) 0:00:34.842 ********* 2025-08-29 17:24:18.571713 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571724 | orchestrator | 2025-08-29 17:24:18.571735 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571746 | orchestrator | Friday 29 August 2025 17:24:18 +0000 (0:00:00.162) 0:00:35.004 ********* 2025-08-29 17:24:18.571763 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571774 | orchestrator | 2025-08-29 17:24:18.571785 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571796 | orchestrator | Friday 29 August 2025 17:24:18 +0000 (0:00:00.196) 0:00:35.200 ********* 2025-08-29 17:24:18.571807 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571818 | orchestrator | 2025-08-29 17:24:18.571829 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:24:18.571839 | orchestrator | Friday 29 August 2025 17:24:18 +0000 (0:00:00.159) 0:00:35.360 ********* 2025-08-29 17:24:18.571850 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:18.571861 | orchestrator | 2025-08-29 17:24:18.571872 | orchestrator | TASK [Set UUIDs for OSD VGs/LVs] *********************************************** 2025-08-29 17:24:18.571889 | orchestrator | Friday 29 August 2025 17:24:18 +0000 (0:00:00.200) 0:00:35.560 ********* 2025-08-29 17:24:22.297074 | orchestrator | ok: [testbed-node-5] => (item={'key': 'sdb', 'value': None}) 2025-08-29 17:24:22.297145 | orchestrator | ok: [testbed-node-5] => (item={'key': 'sdc', 'value': None}) 2025-08-29 17:24:22.297155 | orchestrator | 2025-08-29 17:24:22.297164 | orchestrator | TASK [Generate WAL VG names] *************************************************** 2025-08-29 17:24:22.297171 | orchestrator | Friday 29 August 2025 17:24:18 +0000 (0:00:00.172) 0:00:35.732 ********* 2025-08-29 17:24:22.297179 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297187 | orchestrator | 2025-08-29 17:24:22.297194 | orchestrator | TASK [Generate DB VG names] **************************************************** 2025-08-29 17:24:22.297201 | orchestrator | Friday 29 August 2025 17:24:18 +0000 (0:00:00.140) 0:00:35.872 ********* 2025-08-29 17:24:22.297209 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297216 | orchestrator | 2025-08-29 17:24:22.297223 | orchestrator | TASK [Generate shared DB/WAL VG names] ***************************************** 2025-08-29 17:24:22.297230 | orchestrator | Friday 29 August 2025 17:24:19 +0000 (0:00:00.126) 0:00:35.999 ********* 2025-08-29 17:24:22.297237 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297244 | orchestrator | 2025-08-29 17:24:22.297251 | orchestrator | TASK [Define lvm_volumes structures] ******************************************* 2025-08-29 17:24:22.297258 | orchestrator | Friday 29 August 2025 17:24:19 +0000 (0:00:00.111) 0:00:36.111 ********* 2025-08-29 17:24:22.297265 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:24:22.297273 | orchestrator | 2025-08-29 17:24:22.297280 | orchestrator | TASK [Generate lvm_volumes structure (block only)] ***************************** 2025-08-29 17:24:22.297320 | orchestrator | Friday 29 August 2025 17:24:19 +0000 (0:00:00.232) 0:00:36.344 ********* 2025-08-29 17:24:22.297328 | orchestrator | ok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}}) 2025-08-29 17:24:22.297336 | orchestrator | ok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '21b183ef-757d-561c-bead-7bb3aee28288'}}) 2025-08-29 17:24:22.297343 | orchestrator | 2025-08-29 17:24:22.297350 | orchestrator | TASK [Generate lvm_volumes structure (block + db)] ***************************** 2025-08-29 17:24:22.297357 | orchestrator | Friday 29 August 2025 17:24:19 +0000 (0:00:00.147) 0:00:36.492 ********* 2025-08-29 17:24:22.297365 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}})  2025-08-29 17:24:22.297373 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '21b183ef-757d-561c-bead-7bb3aee28288'}})  2025-08-29 17:24:22.297380 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297388 | orchestrator | 2025-08-29 17:24:22.297395 | orchestrator | TASK [Generate lvm_volumes structure (block + wal)] **************************** 2025-08-29 17:24:22.297402 | orchestrator | Friday 29 August 2025 17:24:19 +0000 (0:00:00.146) 0:00:36.638 ********* 2025-08-29 17:24:22.297410 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}})  2025-08-29 17:24:22.297417 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '21b183ef-757d-561c-bead-7bb3aee28288'}})  2025-08-29 17:24:22.297442 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297450 | orchestrator | 2025-08-29 17:24:22.297457 | orchestrator | TASK [Generate lvm_volumes structure (block + db + wal)] *********************** 2025-08-29 17:24:22.297465 | orchestrator | Friday 29 August 2025 17:24:19 +0000 (0:00:00.147) 0:00:36.785 ********* 2025-08-29 17:24:22.297472 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}})  2025-08-29 17:24:22.297491 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '21b183ef-757d-561c-bead-7bb3aee28288'}})  2025-08-29 17:24:22.297498 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297506 | orchestrator | 2025-08-29 17:24:22.297513 | orchestrator | TASK [Compile lvm_volumes] ***************************************************** 2025-08-29 17:24:22.297520 | orchestrator | Friday 29 August 2025 17:24:19 +0000 (0:00:00.139) 0:00:36.925 ********* 2025-08-29 17:24:22.297527 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:24:22.297534 | orchestrator | 2025-08-29 17:24:22.297541 | orchestrator | TASK [Set OSD devices config data] ********************************************* 2025-08-29 17:24:22.297549 | orchestrator | Friday 29 August 2025 17:24:20 +0000 (0:00:00.130) 0:00:37.056 ********* 2025-08-29 17:24:22.297556 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:24:22.297563 | orchestrator | 2025-08-29 17:24:22.297570 | orchestrator | TASK [Set DB devices config data] ********************************************** 2025-08-29 17:24:22.297577 | orchestrator | Friday 29 August 2025 17:24:20 +0000 (0:00:00.140) 0:00:37.197 ********* 2025-08-29 17:24:22.297584 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297591 | orchestrator | 2025-08-29 17:24:22.297598 | orchestrator | TASK [Set WAL devices config data] ********************************************* 2025-08-29 17:24:22.297605 | orchestrator | Friday 29 August 2025 17:24:20 +0000 (0:00:00.128) 0:00:37.325 ********* 2025-08-29 17:24:22.297613 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297620 | orchestrator | 2025-08-29 17:24:22.297627 | orchestrator | TASK [Set DB+WAL devices config data] ****************************************** 2025-08-29 17:24:22.297634 | orchestrator | Friday 29 August 2025 17:24:20 +0000 (0:00:00.115) 0:00:37.440 ********* 2025-08-29 17:24:22.297641 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297648 | orchestrator | 2025-08-29 17:24:22.297655 | orchestrator | TASK [Print ceph_osd_devices] ************************************************** 2025-08-29 17:24:22.297662 | orchestrator | Friday 29 August 2025 17:24:20 +0000 (0:00:00.148) 0:00:37.588 ********* 2025-08-29 17:24:22.297669 | orchestrator | ok: [testbed-node-5] => { 2025-08-29 17:24:22.297677 | orchestrator |  "ceph_osd_devices": { 2025-08-29 17:24:22.297684 | orchestrator |  "sdb": { 2025-08-29 17:24:22.297691 | orchestrator |  "osd_lvm_uuid": "e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe" 2025-08-29 17:24:22.297709 | orchestrator |  }, 2025-08-29 17:24:22.297717 | orchestrator |  "sdc": { 2025-08-29 17:24:22.297724 | orchestrator |  "osd_lvm_uuid": "21b183ef-757d-561c-bead-7bb3aee28288" 2025-08-29 17:24:22.297731 | orchestrator |  } 2025-08-29 17:24:22.297739 | orchestrator |  } 2025-08-29 17:24:22.297746 | orchestrator | } 2025-08-29 17:24:22.297753 | orchestrator | 2025-08-29 17:24:22.297761 | orchestrator | TASK [Print WAL devices] ******************************************************* 2025-08-29 17:24:22.297768 | orchestrator | Friday 29 August 2025 17:24:20 +0000 (0:00:00.123) 0:00:37.712 ********* 2025-08-29 17:24:22.297775 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297782 | orchestrator | 2025-08-29 17:24:22.297789 | orchestrator | TASK [Print DB devices] ******************************************************** 2025-08-29 17:24:22.297796 | orchestrator | Friday 29 August 2025 17:24:20 +0000 (0:00:00.119) 0:00:37.831 ********* 2025-08-29 17:24:22.297803 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297810 | orchestrator | 2025-08-29 17:24:22.297817 | orchestrator | TASK [Print shared DB/WAL devices] ********************************************* 2025-08-29 17:24:22.297830 | orchestrator | Friday 29 August 2025 17:24:21 +0000 (0:00:00.287) 0:00:38.119 ********* 2025-08-29 17:24:22.297837 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:24:22.297844 | orchestrator | 2025-08-29 17:24:22.297852 | orchestrator | TASK [Print configuration data] ************************************************ 2025-08-29 17:24:22.297859 | orchestrator | Friday 29 August 2025 17:24:21 +0000 (0:00:00.109) 0:00:38.229 ********* 2025-08-29 17:24:22.297866 | orchestrator | changed: [testbed-node-5] => { 2025-08-29 17:24:22.297873 | orchestrator |  "_ceph_configure_lvm_config_data": { 2025-08-29 17:24:22.297880 | orchestrator |  "ceph_osd_devices": { 2025-08-29 17:24:22.297887 | orchestrator |  "sdb": { 2025-08-29 17:24:22.297894 | orchestrator |  "osd_lvm_uuid": "e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe" 2025-08-29 17:24:22.297901 | orchestrator |  }, 2025-08-29 17:24:22.297908 | orchestrator |  "sdc": { 2025-08-29 17:24:22.297915 | orchestrator |  "osd_lvm_uuid": "21b183ef-757d-561c-bead-7bb3aee28288" 2025-08-29 17:24:22.297923 | orchestrator |  } 2025-08-29 17:24:22.297930 | orchestrator |  }, 2025-08-29 17:24:22.297937 | orchestrator |  "lvm_volumes": [ 2025-08-29 17:24:22.297944 | orchestrator |  { 2025-08-29 17:24:22.297951 | orchestrator |  "data": "osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe", 2025-08-29 17:24:22.297958 | orchestrator |  "data_vg": "ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe" 2025-08-29 17:24:22.297965 | orchestrator |  }, 2025-08-29 17:24:22.297972 | orchestrator |  { 2025-08-29 17:24:22.297979 | orchestrator |  "data": "osd-block-21b183ef-757d-561c-bead-7bb3aee28288", 2025-08-29 17:24:22.297986 | orchestrator |  "data_vg": "ceph-21b183ef-757d-561c-bead-7bb3aee28288" 2025-08-29 17:24:22.297993 | orchestrator |  } 2025-08-29 17:24:22.298000 | orchestrator |  ] 2025-08-29 17:24:22.298007 | orchestrator |  } 2025-08-29 17:24:22.298053 | orchestrator | } 2025-08-29 17:24:22.298065 | orchestrator | 2025-08-29 17:24:22.298073 | orchestrator | RUNNING HANDLER [Write configuration file] ************************************* 2025-08-29 17:24:22.298080 | orchestrator | Friday 29 August 2025 17:24:21 +0000 (0:00:00.179) 0:00:38.408 ********* 2025-08-29 17:24:22.298087 | orchestrator | changed: [testbed-node-5 -> testbed-manager(192.168.16.5)] 2025-08-29 17:24:22.298094 | orchestrator | 2025-08-29 17:24:22.298102 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:24:22.298109 | orchestrator | testbed-node-3 : ok=42  changed=2  unreachable=0 failed=0 skipped=32  rescued=0 ignored=0 2025-08-29 17:24:22.298118 | orchestrator | testbed-node-4 : ok=42  changed=2  unreachable=0 failed=0 skipped=32  rescued=0 ignored=0 2025-08-29 17:24:22.298125 | orchestrator | testbed-node-5 : ok=42  changed=2  unreachable=0 failed=0 skipped=32  rescued=0 ignored=0 2025-08-29 17:24:22.298133 | orchestrator | 2025-08-29 17:24:22.298140 | orchestrator | 2025-08-29 17:24:22.298147 | orchestrator | 2025-08-29 17:24:22.298155 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:24:22.298162 | orchestrator | Friday 29 August 2025 17:24:22 +0000 (0:00:00.873) 0:00:39.282 ********* 2025-08-29 17:24:22.298169 | orchestrator | =============================================================================== 2025-08-29 17:24:22.298176 | orchestrator | Write configuration file ------------------------------------------------ 4.07s 2025-08-29 17:24:22.298184 | orchestrator | Add known partitions to the list of available block devices ------------- 1.12s 2025-08-29 17:24:22.298191 | orchestrator | Add known links to the list of available block devices ------------------ 1.08s 2025-08-29 17:24:22.298198 | orchestrator | Add known partitions to the list of available block devices ------------- 1.03s 2025-08-29 17:24:22.298205 | orchestrator | Get extra vars for Ceph configuration ----------------------------------- 0.95s 2025-08-29 17:24:22.298213 | orchestrator | Get initial list of available block devices ----------------------------- 0.92s 2025-08-29 17:24:22.298224 | orchestrator | Add known links to the list of available block devices ------------------ 0.74s 2025-08-29 17:24:22.298232 | orchestrator | Set UUIDs for OSD VGs/LVs ----------------------------------------------- 0.71s 2025-08-29 17:24:22.298239 | orchestrator | Add known partitions to the list of available block devices ------------- 0.70s 2025-08-29 17:24:22.298246 | orchestrator | Add known links to the list of available block devices ------------------ 0.68s 2025-08-29 17:24:22.298254 | orchestrator | Add known links to the list of available block devices ------------------ 0.62s 2025-08-29 17:24:22.298261 | orchestrator | Add known links to the list of available block devices ------------------ 0.62s 2025-08-29 17:24:22.298268 | orchestrator | Add known partitions to the list of available block devices ------------- 0.60s 2025-08-29 17:24:22.298276 | orchestrator | Generate lvm_volumes structure (block + db + wal) ----------------------- 0.59s 2025-08-29 17:24:22.298308 | orchestrator | Set WAL devices config data --------------------------------------------- 0.58s 2025-08-29 17:24:22.523757 | orchestrator | Add known partitions to the list of available block devices ------------- 0.57s 2025-08-29 17:24:22.523837 | orchestrator | Print configuration data ------------------------------------------------ 0.55s 2025-08-29 17:24:22.523850 | orchestrator | Add known partitions to the list of available block devices ------------- 0.53s 2025-08-29 17:24:22.523862 | orchestrator | Print DB devices -------------------------------------------------------- 0.53s 2025-08-29 17:24:22.523873 | orchestrator | Add known links to the list of available block devices ------------------ 0.52s 2025-08-29 17:24:44.664942 | orchestrator | 2025-08-29 17:24:44 | INFO  | Task 46e40d9e-6727-4f0b-be22-c34d4a3b38af (sync inventory) is running in background. Output coming soon. 2025-08-29 17:25:03.405202 | orchestrator | 2025-08-29 17:24:46 | INFO  | Starting group_vars file reorganization 2025-08-29 17:25:03.405315 | orchestrator | 2025-08-29 17:24:46 | INFO  | Moved 0 file(s) to their respective directories 2025-08-29 17:25:03.405332 | orchestrator | 2025-08-29 17:24:46 | INFO  | Group_vars file reorganization completed 2025-08-29 17:25:03.405344 | orchestrator | 2025-08-29 17:24:48 | INFO  | Starting variable preparation from inventory 2025-08-29 17:25:03.405355 | orchestrator | 2025-08-29 17:24:49 | INFO  | Writing 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts 2025-08-29 17:25:03.405366 | orchestrator | 2025-08-29 17:24:49 | INFO  | Writing 050-infrastructure-cephclient-mons.yml with cephclient_mons 2025-08-29 17:25:03.405377 | orchestrator | 2025-08-29 17:24:49 | INFO  | Writing 050-ceph-cluster-fsid.yml with ceph_cluster_fsid 2025-08-29 17:25:03.405405 | orchestrator | 2025-08-29 17:24:49 | INFO  | 3 file(s) written, 6 host(s) processed 2025-08-29 17:25:03.405417 | orchestrator | 2025-08-29 17:24:49 | INFO  | Variable preparation completed 2025-08-29 17:25:03.405428 | orchestrator | 2025-08-29 17:24:50 | INFO  | Starting inventory overwrite handling 2025-08-29 17:25:03.405439 | orchestrator | 2025-08-29 17:24:50 | INFO  | Handling group overwrites in 99-overwrite 2025-08-29 17:25:03.405450 | orchestrator | 2025-08-29 17:24:50 | INFO  | Removing group frr:children from 60-generic 2025-08-29 17:25:03.405466 | orchestrator | 2025-08-29 17:24:50 | INFO  | Removing group storage:children from 50-kolla 2025-08-29 17:25:03.405477 | orchestrator | 2025-08-29 17:24:50 | INFO  | Removing group netbird:children from 50-infrastruture 2025-08-29 17:25:03.405488 | orchestrator | 2025-08-29 17:24:50 | INFO  | Removing group ceph-mds from 50-ceph 2025-08-29 17:25:03.405499 | orchestrator | 2025-08-29 17:24:50 | INFO  | Removing group ceph-rgw from 50-ceph 2025-08-29 17:25:03.405510 | orchestrator | 2025-08-29 17:24:50 | INFO  | Handling group overwrites in 20-roles 2025-08-29 17:25:03.405521 | orchestrator | 2025-08-29 17:24:50 | INFO  | Removing group k3s_node from 50-infrastruture 2025-08-29 17:25:03.405551 | orchestrator | 2025-08-29 17:24:50 | INFO  | Removed 6 group(s) in total 2025-08-29 17:25:03.405563 | orchestrator | 2025-08-29 17:24:50 | INFO  | Inventory overwrite handling completed 2025-08-29 17:25:03.405574 | orchestrator | 2025-08-29 17:24:51 | INFO  | Starting merge of inventory files 2025-08-29 17:25:03.405585 | orchestrator | 2025-08-29 17:24:51 | INFO  | Inventory files merged successfully 2025-08-29 17:25:03.405596 | orchestrator | 2025-08-29 17:24:55 | INFO  | Generating ClusterShell configuration from Ansible inventory 2025-08-29 17:25:03.405607 | orchestrator | 2025-08-29 17:25:02 | INFO  | Successfully wrote ClusterShell configuration 2025-08-29 17:25:03.405618 | orchestrator | [master d944f7d] 2025-08-29-17-25 2025-08-29 17:25:03.405629 | orchestrator | 1 file changed, 30 insertions(+), 9 deletions(-) 2025-08-29 17:25:05.417169 | orchestrator | 2025-08-29 17:25:05 | INFO  | Task e9c580a3-b57b-498a-a9d8-511a0e9e8d32 (ceph-create-lvm-devices) was prepared for execution. 2025-08-29 17:25:05.417251 | orchestrator | 2025-08-29 17:25:05 | INFO  | It takes a moment until task e9c580a3-b57b-498a-a9d8-511a0e9e8d32 (ceph-create-lvm-devices) has been started and output is visible here. 2025-08-29 17:25:15.804222 | orchestrator | 2025-08-29 17:25:15.804348 | orchestrator | PLAY [Ceph create LVM devices] ************************************************* 2025-08-29 17:25:15.804367 | orchestrator | 2025-08-29 17:25:15.804380 | orchestrator | TASK [Get extra vars for Ceph configuration] *********************************** 2025-08-29 17:25:15.804392 | orchestrator | Friday 29 August 2025 17:25:09 +0000 (0:00:00.251) 0:00:00.251 ********* 2025-08-29 17:25:15.804403 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 17:25:15.804414 | orchestrator | 2025-08-29 17:25:15.804425 | orchestrator | TASK [Get initial list of available block devices] ***************************** 2025-08-29 17:25:15.804435 | orchestrator | Friday 29 August 2025 17:25:09 +0000 (0:00:00.184) 0:00:00.436 ********* 2025-08-29 17:25:15.804446 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:15.804458 | orchestrator | 2025-08-29 17:25:15.804469 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804479 | orchestrator | Friday 29 August 2025 17:25:09 +0000 (0:00:00.163) 0:00:00.599 ********* 2025-08-29 17:25:15.804490 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0) 2025-08-29 17:25:15.804501 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1) 2025-08-29 17:25:15.804513 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2) 2025-08-29 17:25:15.804524 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3) 2025-08-29 17:25:15.804535 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4) 2025-08-29 17:25:15.804545 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5) 2025-08-29 17:25:15.804556 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6) 2025-08-29 17:25:15.804567 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7) 2025-08-29 17:25:15.804578 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda) 2025-08-29 17:25:15.804588 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb) 2025-08-29 17:25:15.804599 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc) 2025-08-29 17:25:15.804609 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd) 2025-08-29 17:25:15.804620 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0) 2025-08-29 17:25:15.804631 | orchestrator | 2025-08-29 17:25:15.804641 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804671 | orchestrator | Friday 29 August 2025 17:25:09 +0000 (0:00:00.355) 0:00:00.955 ********* 2025-08-29 17:25:15.804683 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.804694 | orchestrator | 2025-08-29 17:25:15.804705 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804716 | orchestrator | Friday 29 August 2025 17:25:10 +0000 (0:00:00.358) 0:00:01.313 ********* 2025-08-29 17:25:15.804726 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.804737 | orchestrator | 2025-08-29 17:25:15.804748 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804759 | orchestrator | Friday 29 August 2025 17:25:10 +0000 (0:00:00.172) 0:00:01.486 ********* 2025-08-29 17:25:15.804770 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.804782 | orchestrator | 2025-08-29 17:25:15.804795 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804808 | orchestrator | Friday 29 August 2025 17:25:10 +0000 (0:00:00.156) 0:00:01.642 ********* 2025-08-29 17:25:15.804820 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.804833 | orchestrator | 2025-08-29 17:25:15.804845 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804857 | orchestrator | Friday 29 August 2025 17:25:10 +0000 (0:00:00.167) 0:00:01.810 ********* 2025-08-29 17:25:15.804870 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.804882 | orchestrator | 2025-08-29 17:25:15.804894 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804906 | orchestrator | Friday 29 August 2025 17:25:10 +0000 (0:00:00.182) 0:00:01.992 ********* 2025-08-29 17:25:15.804918 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.804930 | orchestrator | 2025-08-29 17:25:15.804943 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.804955 | orchestrator | Friday 29 August 2025 17:25:11 +0000 (0:00:00.191) 0:00:02.183 ********* 2025-08-29 17:25:15.804967 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.804979 | orchestrator | 2025-08-29 17:25:15.804991 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.805003 | orchestrator | Friday 29 August 2025 17:25:11 +0000 (0:00:00.140) 0:00:02.324 ********* 2025-08-29 17:25:15.805016 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805028 | orchestrator | 2025-08-29 17:25:15.805041 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.805053 | orchestrator | Friday 29 August 2025 17:25:11 +0000 (0:00:00.176) 0:00:02.501 ********* 2025-08-29 17:25:15.805065 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1) 2025-08-29 17:25:15.805078 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1) 2025-08-29 17:25:15.805089 | orchestrator | 2025-08-29 17:25:15.805101 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.805114 | orchestrator | Friday 29 August 2025 17:25:11 +0000 (0:00:00.451) 0:00:02.952 ********* 2025-08-29 17:25:15.805141 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322) 2025-08-29 17:25:15.805153 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322) 2025-08-29 17:25:15.805164 | orchestrator | 2025-08-29 17:25:15.805175 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.805186 | orchestrator | Friday 29 August 2025 17:25:12 +0000 (0:00:00.428) 0:00:03.380 ********* 2025-08-29 17:25:15.805197 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f) 2025-08-29 17:25:15.805207 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f) 2025-08-29 17:25:15.805218 | orchestrator | 2025-08-29 17:25:15.805229 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.805247 | orchestrator | Friday 29 August 2025 17:25:12 +0000 (0:00:00.531) 0:00:03.912 ********* 2025-08-29 17:25:15.805259 | orchestrator | ok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5) 2025-08-29 17:25:15.805269 | orchestrator | ok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5) 2025-08-29 17:25:15.805280 | orchestrator | 2025-08-29 17:25:15.805291 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:15.805317 | orchestrator | Friday 29 August 2025 17:25:13 +0000 (0:00:00.515) 0:00:04.428 ********* 2025-08-29 17:25:15.805327 | orchestrator | ok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001) 2025-08-29 17:25:15.805338 | orchestrator | 2025-08-29 17:25:15.805349 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805359 | orchestrator | Friday 29 August 2025 17:25:13 +0000 (0:00:00.536) 0:00:04.965 ********* 2025-08-29 17:25:15.805370 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0) 2025-08-29 17:25:15.805381 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1) 2025-08-29 17:25:15.805392 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2) 2025-08-29 17:25:15.805402 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3) 2025-08-29 17:25:15.805427 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4) 2025-08-29 17:25:15.805439 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5) 2025-08-29 17:25:15.805450 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6) 2025-08-29 17:25:15.805461 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7) 2025-08-29 17:25:15.805471 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda) 2025-08-29 17:25:15.805482 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb) 2025-08-29 17:25:15.805493 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc) 2025-08-29 17:25:15.805504 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd) 2025-08-29 17:25:15.805519 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0) 2025-08-29 17:25:15.805530 | orchestrator | 2025-08-29 17:25:15.805541 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805552 | orchestrator | Friday 29 August 2025 17:25:14 +0000 (0:00:00.363) 0:00:05.328 ********* 2025-08-29 17:25:15.805563 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805574 | orchestrator | 2025-08-29 17:25:15.805585 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805595 | orchestrator | Friday 29 August 2025 17:25:14 +0000 (0:00:00.237) 0:00:05.565 ********* 2025-08-29 17:25:15.805606 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805617 | orchestrator | 2025-08-29 17:25:15.805627 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805638 | orchestrator | Friday 29 August 2025 17:25:14 +0000 (0:00:00.177) 0:00:05.743 ********* 2025-08-29 17:25:15.805649 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805659 | orchestrator | 2025-08-29 17:25:15.805670 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805681 | orchestrator | Friday 29 August 2025 17:25:14 +0000 (0:00:00.183) 0:00:05.926 ********* 2025-08-29 17:25:15.805692 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805702 | orchestrator | 2025-08-29 17:25:15.805713 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805724 | orchestrator | Friday 29 August 2025 17:25:15 +0000 (0:00:00.184) 0:00:06.110 ********* 2025-08-29 17:25:15.805741 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805752 | orchestrator | 2025-08-29 17:25:15.805763 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805774 | orchestrator | Friday 29 August 2025 17:25:15 +0000 (0:00:00.192) 0:00:06.303 ********* 2025-08-29 17:25:15.805785 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805795 | orchestrator | 2025-08-29 17:25:15.805806 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805817 | orchestrator | Friday 29 August 2025 17:25:15 +0000 (0:00:00.175) 0:00:06.479 ********* 2025-08-29 17:25:15.805828 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:15.805838 | orchestrator | 2025-08-29 17:25:15.805849 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:15.805860 | orchestrator | Friday 29 August 2025 17:25:15 +0000 (0:00:00.172) 0:00:06.652 ********* 2025-08-29 17:25:15.805877 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.795428 | orchestrator | 2025-08-29 17:25:23.795543 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:23.795560 | orchestrator | Friday 29 August 2025 17:25:15 +0000 (0:00:00.176) 0:00:06.829 ********* 2025-08-29 17:25:23.795574 | orchestrator | ok: [testbed-node-3] => (item=sda1) 2025-08-29 17:25:23.795595 | orchestrator | ok: [testbed-node-3] => (item=sda14) 2025-08-29 17:25:23.795614 | orchestrator | ok: [testbed-node-3] => (item=sda15) 2025-08-29 17:25:23.795632 | orchestrator | ok: [testbed-node-3] => (item=sda16) 2025-08-29 17:25:23.795650 | orchestrator | 2025-08-29 17:25:23.795669 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:23.795687 | orchestrator | Friday 29 August 2025 17:25:16 +0000 (0:00:00.840) 0:00:07.669 ********* 2025-08-29 17:25:23.795705 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.795725 | orchestrator | 2025-08-29 17:25:23.795743 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:23.795762 | orchestrator | Friday 29 August 2025 17:25:16 +0000 (0:00:00.185) 0:00:07.854 ********* 2025-08-29 17:25:23.795781 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.795800 | orchestrator | 2025-08-29 17:25:23.795818 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:23.795837 | orchestrator | Friday 29 August 2025 17:25:17 +0000 (0:00:00.204) 0:00:08.058 ********* 2025-08-29 17:25:23.795858 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.795878 | orchestrator | 2025-08-29 17:25:23.795897 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:23.795917 | orchestrator | Friday 29 August 2025 17:25:17 +0000 (0:00:00.202) 0:00:08.260 ********* 2025-08-29 17:25:23.795937 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.795957 | orchestrator | 2025-08-29 17:25:23.796011 | orchestrator | TASK [Check whether ceph_db_wal_devices is used exclusively] ******************* 2025-08-29 17:25:23.796034 | orchestrator | Friday 29 August 2025 17:25:17 +0000 (0:00:00.174) 0:00:08.435 ********* 2025-08-29 17:25:23.796057 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796079 | orchestrator | 2025-08-29 17:25:23.796103 | orchestrator | TASK [Create dict of block VGs -> PVs from ceph_osd_devices] ******************* 2025-08-29 17:25:23.796124 | orchestrator | Friday 29 August 2025 17:25:17 +0000 (0:00:00.115) 0:00:08.551 ********* 2025-08-29 17:25:23.796144 | orchestrator | ok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}}) 2025-08-29 17:25:23.796164 | orchestrator | ok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '67ff47d9-d75a-55af-b095-c4dbbf8f796a'}}) 2025-08-29 17:25:23.796182 | orchestrator | 2025-08-29 17:25:23.796201 | orchestrator | TASK [Create block VGs] ******************************************************** 2025-08-29 17:25:23.796219 | orchestrator | Friday 29 August 2025 17:25:17 +0000 (0:00:00.198) 0:00:08.750 ********* 2025-08-29 17:25:23.796242 | orchestrator | changed: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}) 2025-08-29 17:25:23.796287 | orchestrator | changed: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'}) 2025-08-29 17:25:23.796325 | orchestrator | 2025-08-29 17:25:23.796337 | orchestrator | TASK [Print 'Create block VGs'] ************************************************ 2025-08-29 17:25:23.796348 | orchestrator | Friday 29 August 2025 17:25:19 +0000 (0:00:02.003) 0:00:10.754 ********* 2025-08-29 17:25:23.796359 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.796372 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.796382 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796393 | orchestrator | 2025-08-29 17:25:23.796404 | orchestrator | TASK [Create block LVs] ******************************************************** 2025-08-29 17:25:23.796414 | orchestrator | Friday 29 August 2025 17:25:19 +0000 (0:00:00.168) 0:00:10.922 ********* 2025-08-29 17:25:23.796425 | orchestrator | changed: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}) 2025-08-29 17:25:23.796436 | orchestrator | changed: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'}) 2025-08-29 17:25:23.796447 | orchestrator | 2025-08-29 17:25:23.796457 | orchestrator | TASK [Print 'Create block LVs'] ************************************************ 2025-08-29 17:25:23.796468 | orchestrator | Friday 29 August 2025 17:25:21 +0000 (0:00:01.509) 0:00:12.432 ********* 2025-08-29 17:25:23.796479 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.796490 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.796501 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796512 | orchestrator | 2025-08-29 17:25:23.796522 | orchestrator | TASK [Create DB VGs] *********************************************************** 2025-08-29 17:25:23.796533 | orchestrator | Friday 29 August 2025 17:25:21 +0000 (0:00:00.158) 0:00:12.590 ********* 2025-08-29 17:25:23.796544 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796555 | orchestrator | 2025-08-29 17:25:23.796566 | orchestrator | TASK [Print 'Create DB VGs'] *************************************************** 2025-08-29 17:25:23.796597 | orchestrator | Friday 29 August 2025 17:25:21 +0000 (0:00:00.145) 0:00:12.735 ********* 2025-08-29 17:25:23.796608 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.796619 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.796630 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796640 | orchestrator | 2025-08-29 17:25:23.796651 | orchestrator | TASK [Create WAL VGs] ********************************************************** 2025-08-29 17:25:23.796662 | orchestrator | Friday 29 August 2025 17:25:22 +0000 (0:00:00.484) 0:00:13.220 ********* 2025-08-29 17:25:23.796672 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796683 | orchestrator | 2025-08-29 17:25:23.796693 | orchestrator | TASK [Print 'Create WAL VGs'] ************************************************** 2025-08-29 17:25:23.796705 | orchestrator | Friday 29 August 2025 17:25:22 +0000 (0:00:00.160) 0:00:13.381 ********* 2025-08-29 17:25:23.796724 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.796754 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.796774 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796793 | orchestrator | 2025-08-29 17:25:23.796808 | orchestrator | TASK [Create DB+WAL VGs] ******************************************************* 2025-08-29 17:25:23.796819 | orchestrator | Friday 29 August 2025 17:25:22 +0000 (0:00:00.167) 0:00:13.549 ********* 2025-08-29 17:25:23.796829 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796840 | orchestrator | 2025-08-29 17:25:23.796851 | orchestrator | TASK [Print 'Create DB+WAL VGs'] *********************************************** 2025-08-29 17:25:23.796862 | orchestrator | Friday 29 August 2025 17:25:22 +0000 (0:00:00.148) 0:00:13.697 ********* 2025-08-29 17:25:23.796872 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.796883 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.796894 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.796905 | orchestrator | 2025-08-29 17:25:23.796915 | orchestrator | TASK [Prepare variables for OSD count check] *********************************** 2025-08-29 17:25:23.796926 | orchestrator | Friday 29 August 2025 17:25:22 +0000 (0:00:00.161) 0:00:13.858 ********* 2025-08-29 17:25:23.796937 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:23.796948 | orchestrator | 2025-08-29 17:25:23.796959 | orchestrator | TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] **************** 2025-08-29 17:25:23.796969 | orchestrator | Friday 29 August 2025 17:25:22 +0000 (0:00:00.149) 0:00:14.008 ********* 2025-08-29 17:25:23.796997 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.797013 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.797025 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.797035 | orchestrator | 2025-08-29 17:25:23.797046 | orchestrator | TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] *************** 2025-08-29 17:25:23.797057 | orchestrator | Friday 29 August 2025 17:25:23 +0000 (0:00:00.174) 0:00:14.183 ********* 2025-08-29 17:25:23.797067 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.797078 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.797089 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.797100 | orchestrator | 2025-08-29 17:25:23.797111 | orchestrator | TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************ 2025-08-29 17:25:23.797121 | orchestrator | Friday 29 August 2025 17:25:23 +0000 (0:00:00.170) 0:00:14.353 ********* 2025-08-29 17:25:23.797132 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:23.797143 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:23.797154 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.797164 | orchestrator | 2025-08-29 17:25:23.797175 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a DB VG] ********************* 2025-08-29 17:25:23.797186 | orchestrator | Friday 29 August 2025 17:25:23 +0000 (0:00:00.172) 0:00:14.526 ********* 2025-08-29 17:25:23.797197 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.797207 | orchestrator | 2025-08-29 17:25:23.797218 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ******************** 2025-08-29 17:25:23.797236 | orchestrator | Friday 29 August 2025 17:25:23 +0000 (0:00:00.147) 0:00:14.673 ********* 2025-08-29 17:25:23.797246 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:23.797257 | orchestrator | 2025-08-29 17:25:23.797274 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] ***************** 2025-08-29 17:25:30.710575 | orchestrator | Friday 29 August 2025 17:25:23 +0000 (0:00:00.147) 0:00:14.821 ********* 2025-08-29 17:25:30.710686 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.710701 | orchestrator | 2025-08-29 17:25:30.710714 | orchestrator | TASK [Print number of OSDs wanted per DB VG] *********************************** 2025-08-29 17:25:30.710726 | orchestrator | Friday 29 August 2025 17:25:23 +0000 (0:00:00.145) 0:00:14.966 ********* 2025-08-29 17:25:30.710737 | orchestrator | ok: [testbed-node-3] => { 2025-08-29 17:25:30.710748 | orchestrator |  "_num_osds_wanted_per_db_vg": {} 2025-08-29 17:25:30.710760 | orchestrator | } 2025-08-29 17:25:30.710771 | orchestrator | 2025-08-29 17:25:30.710782 | orchestrator | TASK [Print number of OSDs wanted per WAL VG] ********************************** 2025-08-29 17:25:30.710794 | orchestrator | Friday 29 August 2025 17:25:24 +0000 (0:00:00.450) 0:00:15.416 ********* 2025-08-29 17:25:30.710805 | orchestrator | ok: [testbed-node-3] => { 2025-08-29 17:25:30.710816 | orchestrator |  "_num_osds_wanted_per_wal_vg": {} 2025-08-29 17:25:30.710827 | orchestrator | } 2025-08-29 17:25:30.710838 | orchestrator | 2025-08-29 17:25:30.710848 | orchestrator | TASK [Print number of OSDs wanted per DB+WAL VG] ******************************* 2025-08-29 17:25:30.710859 | orchestrator | Friday 29 August 2025 17:25:24 +0000 (0:00:00.147) 0:00:15.564 ********* 2025-08-29 17:25:30.710870 | orchestrator | ok: [testbed-node-3] => { 2025-08-29 17:25:30.710881 | orchestrator |  "_num_osds_wanted_per_db_wal_vg": {} 2025-08-29 17:25:30.710892 | orchestrator | } 2025-08-29 17:25:30.710903 | orchestrator | 2025-08-29 17:25:30.710915 | orchestrator | TASK [Gather DB VGs with total and available size in bytes] ******************** 2025-08-29 17:25:30.710927 | orchestrator | Friday 29 August 2025 17:25:24 +0000 (0:00:00.147) 0:00:15.711 ********* 2025-08-29 17:25:30.710938 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:30.710949 | orchestrator | 2025-08-29 17:25:30.710960 | orchestrator | TASK [Gather WAL VGs with total and available size in bytes] ******************* 2025-08-29 17:25:30.710971 | orchestrator | Friday 29 August 2025 17:25:25 +0000 (0:00:00.721) 0:00:16.432 ********* 2025-08-29 17:25:30.710981 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:30.710992 | orchestrator | 2025-08-29 17:25:30.711003 | orchestrator | TASK [Gather DB+WAL VGs with total and available size in bytes] **************** 2025-08-29 17:25:30.711014 | orchestrator | Friday 29 August 2025 17:25:25 +0000 (0:00:00.538) 0:00:16.971 ********* 2025-08-29 17:25:30.711025 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:30.711036 | orchestrator | 2025-08-29 17:25:30.711047 | orchestrator | TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] ************************* 2025-08-29 17:25:30.711058 | orchestrator | Friday 29 August 2025 17:25:26 +0000 (0:00:00.537) 0:00:17.508 ********* 2025-08-29 17:25:30.711069 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:30.711080 | orchestrator | 2025-08-29 17:25:30.711091 | orchestrator | TASK [Calculate VG sizes (without buffer)] ************************************* 2025-08-29 17:25:30.711103 | orchestrator | Friday 29 August 2025 17:25:26 +0000 (0:00:00.157) 0:00:17.666 ********* 2025-08-29 17:25:30.711116 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711128 | orchestrator | 2025-08-29 17:25:30.711140 | orchestrator | TASK [Calculate VG sizes (with buffer)] **************************************** 2025-08-29 17:25:30.711153 | orchestrator | Friday 29 August 2025 17:25:26 +0000 (0:00:00.144) 0:00:17.811 ********* 2025-08-29 17:25:30.711165 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711178 | orchestrator | 2025-08-29 17:25:30.711190 | orchestrator | TASK [Print LVM VGs report data] *********************************************** 2025-08-29 17:25:30.711203 | orchestrator | Friday 29 August 2025 17:25:26 +0000 (0:00:00.125) 0:00:17.936 ********* 2025-08-29 17:25:30.711215 | orchestrator | ok: [testbed-node-3] => { 2025-08-29 17:25:30.711250 | orchestrator |  "vgs_report": { 2025-08-29 17:25:30.711263 | orchestrator |  "vg": [] 2025-08-29 17:25:30.711275 | orchestrator |  } 2025-08-29 17:25:30.711330 | orchestrator | } 2025-08-29 17:25:30.711344 | orchestrator | 2025-08-29 17:25:30.711356 | orchestrator | TASK [Print LVM VG sizes] ****************************************************** 2025-08-29 17:25:30.711369 | orchestrator | Friday 29 August 2025 17:25:27 +0000 (0:00:00.146) 0:00:18.082 ********* 2025-08-29 17:25:30.711381 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711393 | orchestrator | 2025-08-29 17:25:30.711405 | orchestrator | TASK [Calculate size needed for LVs on ceph_db_devices] ************************ 2025-08-29 17:25:30.711418 | orchestrator | Friday 29 August 2025 17:25:27 +0000 (0:00:00.131) 0:00:18.214 ********* 2025-08-29 17:25:30.711430 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711442 | orchestrator | 2025-08-29 17:25:30.711454 | orchestrator | TASK [Print size needed for LVs on ceph_db_devices] **************************** 2025-08-29 17:25:30.711465 | orchestrator | Friday 29 August 2025 17:25:27 +0000 (0:00:00.126) 0:00:18.340 ********* 2025-08-29 17:25:30.711475 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711486 | orchestrator | 2025-08-29 17:25:30.711497 | orchestrator | TASK [Fail if size of DB LVs on ceph_db_devices > available] ******************* 2025-08-29 17:25:30.711507 | orchestrator | Friday 29 August 2025 17:25:27 +0000 (0:00:00.481) 0:00:18.822 ********* 2025-08-29 17:25:30.711518 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711529 | orchestrator | 2025-08-29 17:25:30.711540 | orchestrator | TASK [Calculate size needed for LVs on ceph_wal_devices] *********************** 2025-08-29 17:25:30.711550 | orchestrator | Friday 29 August 2025 17:25:27 +0000 (0:00:00.138) 0:00:18.961 ********* 2025-08-29 17:25:30.711561 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711572 | orchestrator | 2025-08-29 17:25:30.711583 | orchestrator | TASK [Print size needed for LVs on ceph_wal_devices] *************************** 2025-08-29 17:25:30.711594 | orchestrator | Friday 29 August 2025 17:25:28 +0000 (0:00:00.130) 0:00:19.091 ********* 2025-08-29 17:25:30.711604 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711615 | orchestrator | 2025-08-29 17:25:30.711626 | orchestrator | TASK [Fail if size of WAL LVs on ceph_wal_devices > available] ***************** 2025-08-29 17:25:30.711637 | orchestrator | Friday 29 August 2025 17:25:28 +0000 (0:00:00.140) 0:00:19.231 ********* 2025-08-29 17:25:30.711647 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711658 | orchestrator | 2025-08-29 17:25:30.711669 | orchestrator | TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] **************** 2025-08-29 17:25:30.711680 | orchestrator | Friday 29 August 2025 17:25:28 +0000 (0:00:00.150) 0:00:19.382 ********* 2025-08-29 17:25:30.711690 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711701 | orchestrator | 2025-08-29 17:25:30.711712 | orchestrator | TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ******************** 2025-08-29 17:25:30.711739 | orchestrator | Friday 29 August 2025 17:25:28 +0000 (0:00:00.148) 0:00:19.531 ********* 2025-08-29 17:25:30.711750 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711761 | orchestrator | 2025-08-29 17:25:30.711772 | orchestrator | TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] ***************** 2025-08-29 17:25:30.711783 | orchestrator | Friday 29 August 2025 17:25:28 +0000 (0:00:00.144) 0:00:19.675 ********* 2025-08-29 17:25:30.711793 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711804 | orchestrator | 2025-08-29 17:25:30.711815 | orchestrator | TASK [Print size needed for DB LVs on ceph_db_wal_devices] ********************* 2025-08-29 17:25:30.711826 | orchestrator | Friday 29 August 2025 17:25:28 +0000 (0:00:00.133) 0:00:19.809 ********* 2025-08-29 17:25:30.711837 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711847 | orchestrator | 2025-08-29 17:25:30.711858 | orchestrator | TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] *********** 2025-08-29 17:25:30.711869 | orchestrator | Friday 29 August 2025 17:25:28 +0000 (0:00:00.146) 0:00:19.956 ********* 2025-08-29 17:25:30.711880 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711890 | orchestrator | 2025-08-29 17:25:30.711901 | orchestrator | TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] ************************* 2025-08-29 17:25:30.711921 | orchestrator | Friday 29 August 2025 17:25:29 +0000 (0:00:00.162) 0:00:20.118 ********* 2025-08-29 17:25:30.711931 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711942 | orchestrator | 2025-08-29 17:25:30.711953 | orchestrator | TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] ********************* 2025-08-29 17:25:30.711964 | orchestrator | Friday 29 August 2025 17:25:29 +0000 (0:00:00.150) 0:00:20.268 ********* 2025-08-29 17:25:30.711975 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.711986 | orchestrator | 2025-08-29 17:25:30.711997 | orchestrator | TASK [Create DB LVs for ceph_db_devices] *************************************** 2025-08-29 17:25:30.712008 | orchestrator | Friday 29 August 2025 17:25:29 +0000 (0:00:00.157) 0:00:20.426 ********* 2025-08-29 17:25:30.712019 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:30.712032 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:30.712043 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.712054 | orchestrator | 2025-08-29 17:25:30.712065 | orchestrator | TASK [Print 'Create DB LVs for ceph_db_devices'] ******************************* 2025-08-29 17:25:30.712076 | orchestrator | Friday 29 August 2025 17:25:29 +0000 (0:00:00.153) 0:00:20.579 ********* 2025-08-29 17:25:30.712086 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:30.712098 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:30.712108 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.712119 | orchestrator | 2025-08-29 17:25:30.712130 | orchestrator | TASK [Create WAL LVs for ceph_wal_devices] ************************************* 2025-08-29 17:25:30.712141 | orchestrator | Friday 29 August 2025 17:25:30 +0000 (0:00:00.486) 0:00:21.066 ********* 2025-08-29 17:25:30.712152 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:30.712163 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:30.712174 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.712185 | orchestrator | 2025-08-29 17:25:30.712195 | orchestrator | TASK [Print 'Create WAL LVs for ceph_wal_devices'] ***************************** 2025-08-29 17:25:30.712206 | orchestrator | Friday 29 August 2025 17:25:30 +0000 (0:00:00.171) 0:00:21.237 ********* 2025-08-29 17:25:30.712217 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:30.712228 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:30.712239 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.712250 | orchestrator | 2025-08-29 17:25:30.712261 | orchestrator | TASK [Create WAL LVs for ceph_db_wal_devices] ********************************** 2025-08-29 17:25:30.712272 | orchestrator | Friday 29 August 2025 17:25:30 +0000 (0:00:00.164) 0:00:21.402 ********* 2025-08-29 17:25:30.712282 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:30.712293 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:30.712337 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:30.712348 | orchestrator | 2025-08-29 17:25:30.712359 | orchestrator | TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] ************************** 2025-08-29 17:25:30.712377 | orchestrator | Friday 29 August 2025 17:25:30 +0000 (0:00:00.157) 0:00:21.560 ********* 2025-08-29 17:25:30.712396 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:30.712413 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:36.623949 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:36.624049 | orchestrator | 2025-08-29 17:25:36.624064 | orchestrator | TASK [Create DB LVs for ceph_db_wal_devices] *********************************** 2025-08-29 17:25:36.624076 | orchestrator | Friday 29 August 2025 17:25:30 +0000 (0:00:00.175) 0:00:21.736 ********* 2025-08-29 17:25:36.624088 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:36.624101 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:36.624112 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:36.624123 | orchestrator | 2025-08-29 17:25:36.624134 | orchestrator | TASK [Print 'Create DB LVs for ceph_db_wal_devices'] *************************** 2025-08-29 17:25:36.624145 | orchestrator | Friday 29 August 2025 17:25:30 +0000 (0:00:00.166) 0:00:21.903 ********* 2025-08-29 17:25:36.624156 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:36.624167 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:36.624178 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:36.624189 | orchestrator | 2025-08-29 17:25:36.624200 | orchestrator | TASK [Get list of Ceph LVs with associated VGs] ******************************** 2025-08-29 17:25:36.624211 | orchestrator | Friday 29 August 2025 17:25:31 +0000 (0:00:00.141) 0:00:22.044 ********* 2025-08-29 17:25:36.624223 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:36.624234 | orchestrator | 2025-08-29 17:25:36.624245 | orchestrator | TASK [Get list of Ceph PVs with associated VGs] ******************************** 2025-08-29 17:25:36.624256 | orchestrator | Friday 29 August 2025 17:25:31 +0000 (0:00:00.543) 0:00:22.588 ********* 2025-08-29 17:25:36.624267 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:36.624278 | orchestrator | 2025-08-29 17:25:36.624288 | orchestrator | TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] *********************** 2025-08-29 17:25:36.624332 | orchestrator | Friday 29 August 2025 17:25:32 +0000 (0:00:00.550) 0:00:23.139 ********* 2025-08-29 17:25:36.624344 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:25:36.624355 | orchestrator | 2025-08-29 17:25:36.624366 | orchestrator | TASK [Create list of VG/LV names] ********************************************** 2025-08-29 17:25:36.624377 | orchestrator | Friday 29 August 2025 17:25:32 +0000 (0:00:00.147) 0:00:23.286 ********* 2025-08-29 17:25:36.624388 | orchestrator | ok: [testbed-node-3] => (item={'lv_name': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'vg_name': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}) 2025-08-29 17:25:36.624400 | orchestrator | ok: [testbed-node-3] => (item={'lv_name': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'vg_name': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'}) 2025-08-29 17:25:36.624411 | orchestrator | 2025-08-29 17:25:36.624439 | orchestrator | TASK [Fail if block LV defined in lvm_volumes is missing] ********************** 2025-08-29 17:25:36.624451 | orchestrator | Friday 29 August 2025 17:25:32 +0000 (0:00:00.191) 0:00:23.477 ********* 2025-08-29 17:25:36.624462 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:36.624473 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:36.624505 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:36.624518 | orchestrator | 2025-08-29 17:25:36.624530 | orchestrator | TASK [Fail if DB LV defined in lvm_volumes is missing] ************************* 2025-08-29 17:25:36.624542 | orchestrator | Friday 29 August 2025 17:25:32 +0000 (0:00:00.165) 0:00:23.643 ********* 2025-08-29 17:25:36.624555 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:36.624567 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:36.624580 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:36.624592 | orchestrator | 2025-08-29 17:25:36.624604 | orchestrator | TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************ 2025-08-29 17:25:36.624616 | orchestrator | Friday 29 August 2025 17:25:33 +0000 (0:00:00.512) 0:00:24.156 ********* 2025-08-29 17:25:36.624629 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'})  2025-08-29 17:25:36.624642 | orchestrator | skipping: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'})  2025-08-29 17:25:36.624654 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:25:36.624667 | orchestrator | 2025-08-29 17:25:36.624679 | orchestrator | TASK [Print LVM report data] *************************************************** 2025-08-29 17:25:36.624691 | orchestrator | Friday 29 August 2025 17:25:33 +0000 (0:00:00.149) 0:00:24.305 ********* 2025-08-29 17:25:36.624702 | orchestrator | ok: [testbed-node-3] => { 2025-08-29 17:25:36.624713 | orchestrator |  "lvm_report": { 2025-08-29 17:25:36.624724 | orchestrator |  "lv": [ 2025-08-29 17:25:36.624735 | orchestrator |  { 2025-08-29 17:25:36.624761 | orchestrator |  "lv_name": "osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b", 2025-08-29 17:25:36.624773 | orchestrator |  "vg_name": "ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b" 2025-08-29 17:25:36.624784 | orchestrator |  }, 2025-08-29 17:25:36.624795 | orchestrator |  { 2025-08-29 17:25:36.624806 | orchestrator |  "lv_name": "osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a", 2025-08-29 17:25:36.624816 | orchestrator |  "vg_name": "ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a" 2025-08-29 17:25:36.624827 | orchestrator |  } 2025-08-29 17:25:36.624838 | orchestrator |  ], 2025-08-29 17:25:36.624849 | orchestrator |  "pv": [ 2025-08-29 17:25:36.624859 | orchestrator |  { 2025-08-29 17:25:36.624870 | orchestrator |  "pv_name": "/dev/sdb", 2025-08-29 17:25:36.624881 | orchestrator |  "vg_name": "ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b" 2025-08-29 17:25:36.624892 | orchestrator |  }, 2025-08-29 17:25:36.624902 | orchestrator |  { 2025-08-29 17:25:36.624913 | orchestrator |  "pv_name": "/dev/sdc", 2025-08-29 17:25:36.624924 | orchestrator |  "vg_name": "ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a" 2025-08-29 17:25:36.624935 | orchestrator |  } 2025-08-29 17:25:36.624946 | orchestrator |  ] 2025-08-29 17:25:36.624956 | orchestrator |  } 2025-08-29 17:25:36.624967 | orchestrator | } 2025-08-29 17:25:36.624979 | orchestrator | 2025-08-29 17:25:36.624990 | orchestrator | PLAY [Ceph create LVM devices] ************************************************* 2025-08-29 17:25:36.625000 | orchestrator | 2025-08-29 17:25:36.625011 | orchestrator | TASK [Get extra vars for Ceph configuration] *********************************** 2025-08-29 17:25:36.625022 | orchestrator | Friday 29 August 2025 17:25:33 +0000 (0:00:00.299) 0:00:24.605 ********* 2025-08-29 17:25:36.625033 | orchestrator | ok: [testbed-node-4 -> testbed-manager(192.168.16.5)] 2025-08-29 17:25:36.625044 | orchestrator | 2025-08-29 17:25:36.625062 | orchestrator | TASK [Get initial list of available block devices] ***************************** 2025-08-29 17:25:36.625073 | orchestrator | Friday 29 August 2025 17:25:33 +0000 (0:00:00.303) 0:00:24.908 ********* 2025-08-29 17:25:36.625084 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:36.625095 | orchestrator | 2025-08-29 17:25:36.625106 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625116 | orchestrator | Friday 29 August 2025 17:25:34 +0000 (0:00:00.269) 0:00:25.178 ********* 2025-08-29 17:25:36.625127 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0) 2025-08-29 17:25:36.625138 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1) 2025-08-29 17:25:36.625149 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2) 2025-08-29 17:25:36.625160 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3) 2025-08-29 17:25:36.625171 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4) 2025-08-29 17:25:36.625182 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5) 2025-08-29 17:25:36.625192 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6) 2025-08-29 17:25:36.625203 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7) 2025-08-29 17:25:36.625219 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda) 2025-08-29 17:25:36.625231 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb) 2025-08-29 17:25:36.625241 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc) 2025-08-29 17:25:36.625252 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd) 2025-08-29 17:25:36.625263 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0) 2025-08-29 17:25:36.625274 | orchestrator | 2025-08-29 17:25:36.625285 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625310 | orchestrator | Friday 29 August 2025 17:25:34 +0000 (0:00:00.414) 0:00:25.592 ********* 2025-08-29 17:25:36.625322 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:36.625333 | orchestrator | 2025-08-29 17:25:36.625344 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625354 | orchestrator | Friday 29 August 2025 17:25:34 +0000 (0:00:00.205) 0:00:25.797 ********* 2025-08-29 17:25:36.625365 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:36.625375 | orchestrator | 2025-08-29 17:25:36.625386 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625397 | orchestrator | Friday 29 August 2025 17:25:34 +0000 (0:00:00.187) 0:00:25.985 ********* 2025-08-29 17:25:36.625407 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:36.625418 | orchestrator | 2025-08-29 17:25:36.625429 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625439 | orchestrator | Friday 29 August 2025 17:25:35 +0000 (0:00:00.229) 0:00:26.214 ********* 2025-08-29 17:25:36.625450 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:36.625461 | orchestrator | 2025-08-29 17:25:36.625471 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625482 | orchestrator | Friday 29 August 2025 17:25:35 +0000 (0:00:00.728) 0:00:26.943 ********* 2025-08-29 17:25:36.625493 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:36.625503 | orchestrator | 2025-08-29 17:25:36.625514 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625525 | orchestrator | Friday 29 August 2025 17:25:36 +0000 (0:00:00.229) 0:00:27.173 ********* 2025-08-29 17:25:36.625535 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:36.625546 | orchestrator | 2025-08-29 17:25:36.625556 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:36.625574 | orchestrator | Friday 29 August 2025 17:25:36 +0000 (0:00:00.258) 0:00:27.431 ********* 2025-08-29 17:25:36.625585 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:36.625596 | orchestrator | 2025-08-29 17:25:36.625613 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:46.908243 | orchestrator | Friday 29 August 2025 17:25:36 +0000 (0:00:00.215) 0:00:27.646 ********* 2025-08-29 17:25:46.908381 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.908398 | orchestrator | 2025-08-29 17:25:46.908412 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:46.908423 | orchestrator | Friday 29 August 2025 17:25:36 +0000 (0:00:00.214) 0:00:27.860 ********* 2025-08-29 17:25:46.908434 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320) 2025-08-29 17:25:46.908446 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320) 2025-08-29 17:25:46.908457 | orchestrator | 2025-08-29 17:25:46.908468 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:46.908479 | orchestrator | Friday 29 August 2025 17:25:37 +0000 (0:00:00.412) 0:00:28.273 ********* 2025-08-29 17:25:46.908490 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3) 2025-08-29 17:25:46.908501 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3) 2025-08-29 17:25:46.908512 | orchestrator | 2025-08-29 17:25:46.908522 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:46.908533 | orchestrator | Friday 29 August 2025 17:25:37 +0000 (0:00:00.418) 0:00:28.692 ********* 2025-08-29 17:25:46.908544 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6) 2025-08-29 17:25:46.908555 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6) 2025-08-29 17:25:46.908566 | orchestrator | 2025-08-29 17:25:46.908577 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:46.908588 | orchestrator | Friday 29 August 2025 17:25:38 +0000 (0:00:00.452) 0:00:29.144 ********* 2025-08-29 17:25:46.908599 | orchestrator | ok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474) 2025-08-29 17:25:46.908610 | orchestrator | ok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474) 2025-08-29 17:25:46.908621 | orchestrator | 2025-08-29 17:25:46.908632 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:25:46.908643 | orchestrator | Friday 29 August 2025 17:25:38 +0000 (0:00:00.496) 0:00:29.641 ********* 2025-08-29 17:25:46.908653 | orchestrator | ok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001) 2025-08-29 17:25:46.908664 | orchestrator | 2025-08-29 17:25:46.908675 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.908686 | orchestrator | Friday 29 August 2025 17:25:39 +0000 (0:00:00.430) 0:00:30.071 ********* 2025-08-29 17:25:46.908697 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0) 2025-08-29 17:25:46.908709 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1) 2025-08-29 17:25:46.908720 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2) 2025-08-29 17:25:46.908730 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3) 2025-08-29 17:25:46.908741 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4) 2025-08-29 17:25:46.908752 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5) 2025-08-29 17:25:46.908779 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6) 2025-08-29 17:25:46.908821 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7) 2025-08-29 17:25:46.908841 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda) 2025-08-29 17:25:46.908854 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb) 2025-08-29 17:25:46.908870 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc) 2025-08-29 17:25:46.908887 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd) 2025-08-29 17:25:46.908899 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0) 2025-08-29 17:25:46.908912 | orchestrator | 2025-08-29 17:25:46.908924 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.908936 | orchestrator | Friday 29 August 2025 17:25:39 +0000 (0:00:00.694) 0:00:30.765 ********* 2025-08-29 17:25:46.908948 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.908961 | orchestrator | 2025-08-29 17:25:46.908973 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.908985 | orchestrator | Friday 29 August 2025 17:25:39 +0000 (0:00:00.264) 0:00:31.030 ********* 2025-08-29 17:25:46.908997 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909010 | orchestrator | 2025-08-29 17:25:46.909022 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909035 | orchestrator | Friday 29 August 2025 17:25:40 +0000 (0:00:00.208) 0:00:31.238 ********* 2025-08-29 17:25:46.909047 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909060 | orchestrator | 2025-08-29 17:25:46.909072 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909085 | orchestrator | Friday 29 August 2025 17:25:40 +0000 (0:00:00.249) 0:00:31.488 ********* 2025-08-29 17:25:46.909097 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909110 | orchestrator | 2025-08-29 17:25:46.909136 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909149 | orchestrator | Friday 29 August 2025 17:25:40 +0000 (0:00:00.215) 0:00:31.703 ********* 2025-08-29 17:25:46.909160 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909170 | orchestrator | 2025-08-29 17:25:46.909181 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909191 | orchestrator | Friday 29 August 2025 17:25:40 +0000 (0:00:00.199) 0:00:31.903 ********* 2025-08-29 17:25:46.909202 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909212 | orchestrator | 2025-08-29 17:25:46.909222 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909233 | orchestrator | Friday 29 August 2025 17:25:41 +0000 (0:00:00.238) 0:00:32.141 ********* 2025-08-29 17:25:46.909244 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909254 | orchestrator | 2025-08-29 17:25:46.909265 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909275 | orchestrator | Friday 29 August 2025 17:25:41 +0000 (0:00:00.205) 0:00:32.347 ********* 2025-08-29 17:25:46.909286 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909296 | orchestrator | 2025-08-29 17:25:46.909325 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909336 | orchestrator | Friday 29 August 2025 17:25:41 +0000 (0:00:00.194) 0:00:32.542 ********* 2025-08-29 17:25:46.909347 | orchestrator | ok: [testbed-node-4] => (item=sda1) 2025-08-29 17:25:46.909357 | orchestrator | ok: [testbed-node-4] => (item=sda14) 2025-08-29 17:25:46.909368 | orchestrator | ok: [testbed-node-4] => (item=sda15) 2025-08-29 17:25:46.909379 | orchestrator | ok: [testbed-node-4] => (item=sda16) 2025-08-29 17:25:46.909389 | orchestrator | 2025-08-29 17:25:46.909400 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909424 | orchestrator | Friday 29 August 2025 17:25:42 +0000 (0:00:00.820) 0:00:33.363 ********* 2025-08-29 17:25:46.909444 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909455 | orchestrator | 2025-08-29 17:25:46.909466 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909477 | orchestrator | Friday 29 August 2025 17:25:42 +0000 (0:00:00.187) 0:00:33.550 ********* 2025-08-29 17:25:46.909487 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909498 | orchestrator | 2025-08-29 17:25:46.909509 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909519 | orchestrator | Friday 29 August 2025 17:25:42 +0000 (0:00:00.178) 0:00:33.729 ********* 2025-08-29 17:25:46.909530 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909541 | orchestrator | 2025-08-29 17:25:46.909552 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:25:46.909562 | orchestrator | Friday 29 August 2025 17:25:43 +0000 (0:00:00.488) 0:00:34.218 ********* 2025-08-29 17:25:46.909573 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909584 | orchestrator | 2025-08-29 17:25:46.909595 | orchestrator | TASK [Check whether ceph_db_wal_devices is used exclusively] ******************* 2025-08-29 17:25:46.909606 | orchestrator | Friday 29 August 2025 17:25:43 +0000 (0:00:00.190) 0:00:34.409 ********* 2025-08-29 17:25:46.909617 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909627 | orchestrator | 2025-08-29 17:25:46.909643 | orchestrator | TASK [Create dict of block VGs -> PVs from ceph_osd_devices] ******************* 2025-08-29 17:25:46.909654 | orchestrator | Friday 29 August 2025 17:25:43 +0000 (0:00:00.128) 0:00:34.537 ********* 2025-08-29 17:25:46.909665 | orchestrator | ok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'b89384ec-5219-5f2a-8735-84f78c8179d2'}}) 2025-08-29 17:25:46.909676 | orchestrator | ok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '860f9296-4089-56f8-8238-0f24b03dbca2'}}) 2025-08-29 17:25:46.909687 | orchestrator | 2025-08-29 17:25:46.909698 | orchestrator | TASK [Create block VGs] ******************************************************** 2025-08-29 17:25:46.909709 | orchestrator | Friday 29 August 2025 17:25:43 +0000 (0:00:00.166) 0:00:34.704 ********* 2025-08-29 17:25:46.909720 | orchestrator | changed: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'}) 2025-08-29 17:25:46.909732 | orchestrator | changed: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'}) 2025-08-29 17:25:46.909743 | orchestrator | 2025-08-29 17:25:46.909754 | orchestrator | TASK [Print 'Create block VGs'] ************************************************ 2025-08-29 17:25:46.909764 | orchestrator | Friday 29 August 2025 17:25:45 +0000 (0:00:01.784) 0:00:36.489 ********* 2025-08-29 17:25:46.909775 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:46.909787 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:46.909798 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:46.909809 | orchestrator | 2025-08-29 17:25:46.909819 | orchestrator | TASK [Create block LVs] ******************************************************** 2025-08-29 17:25:46.909830 | orchestrator | Friday 29 August 2025 17:25:45 +0000 (0:00:00.135) 0:00:36.624 ********* 2025-08-29 17:25:46.909841 | orchestrator | changed: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'}) 2025-08-29 17:25:46.909852 | orchestrator | changed: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'}) 2025-08-29 17:25:46.909863 | orchestrator | 2025-08-29 17:25:46.909881 | orchestrator | TASK [Print 'Create block LVs'] ************************************************ 2025-08-29 17:25:52.479205 | orchestrator | Friday 29 August 2025 17:25:46 +0000 (0:00:01.306) 0:00:37.930 ********* 2025-08-29 17:25:52.479411 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:52.479434 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:52.479446 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.479458 | orchestrator | 2025-08-29 17:25:52.479471 | orchestrator | TASK [Create DB VGs] *********************************************************** 2025-08-29 17:25:52.479482 | orchestrator | Friday 29 August 2025 17:25:47 +0000 (0:00:00.165) 0:00:38.096 ********* 2025-08-29 17:25:52.479493 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.479503 | orchestrator | 2025-08-29 17:25:52.479514 | orchestrator | TASK [Print 'Create DB VGs'] *************************************************** 2025-08-29 17:25:52.479525 | orchestrator | Friday 29 August 2025 17:25:47 +0000 (0:00:00.130) 0:00:38.226 ********* 2025-08-29 17:25:52.479542 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:52.479560 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:52.479579 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.479596 | orchestrator | 2025-08-29 17:25:52.479616 | orchestrator | TASK [Create WAL VGs] ********************************************************** 2025-08-29 17:25:52.479635 | orchestrator | Friday 29 August 2025 17:25:47 +0000 (0:00:00.206) 0:00:38.433 ********* 2025-08-29 17:25:52.479653 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.479672 | orchestrator | 2025-08-29 17:25:52.479685 | orchestrator | TASK [Print 'Create WAL VGs'] ************************************************** 2025-08-29 17:25:52.479695 | orchestrator | Friday 29 August 2025 17:25:47 +0000 (0:00:00.177) 0:00:38.611 ********* 2025-08-29 17:25:52.479706 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:52.479716 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:52.479729 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.479741 | orchestrator | 2025-08-29 17:25:52.479753 | orchestrator | TASK [Create DB+WAL VGs] ******************************************************* 2025-08-29 17:25:52.479765 | orchestrator | Friday 29 August 2025 17:25:47 +0000 (0:00:00.175) 0:00:38.786 ********* 2025-08-29 17:25:52.479778 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.479790 | orchestrator | 2025-08-29 17:25:52.479817 | orchestrator | TASK [Print 'Create DB+WAL VGs'] *********************************************** 2025-08-29 17:25:52.479830 | orchestrator | Friday 29 August 2025 17:25:48 +0000 (0:00:00.338) 0:00:39.124 ********* 2025-08-29 17:25:52.479842 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:52.479854 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:52.479867 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.479879 | orchestrator | 2025-08-29 17:25:52.479891 | orchestrator | TASK [Prepare variables for OSD count check] *********************************** 2025-08-29 17:25:52.479904 | orchestrator | Friday 29 August 2025 17:25:48 +0000 (0:00:00.175) 0:00:39.300 ********* 2025-08-29 17:25:52.479916 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:52.479929 | orchestrator | 2025-08-29 17:25:52.479941 | orchestrator | TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] **************** 2025-08-29 17:25:52.479953 | orchestrator | Friday 29 August 2025 17:25:48 +0000 (0:00:00.150) 0:00:39.450 ********* 2025-08-29 17:25:52.479976 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:52.479989 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:52.480003 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480022 | orchestrator | 2025-08-29 17:25:52.480042 | orchestrator | TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] *************** 2025-08-29 17:25:52.480060 | orchestrator | Friday 29 August 2025 17:25:48 +0000 (0:00:00.165) 0:00:39.616 ********* 2025-08-29 17:25:52.480079 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:52.480098 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:52.480110 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480120 | orchestrator | 2025-08-29 17:25:52.480131 | orchestrator | TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************ 2025-08-29 17:25:52.480142 | orchestrator | Friday 29 August 2025 17:25:48 +0000 (0:00:00.157) 0:00:39.773 ********* 2025-08-29 17:25:52.480171 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:52.480183 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:52.480193 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480204 | orchestrator | 2025-08-29 17:25:52.480215 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a DB VG] ********************* 2025-08-29 17:25:52.480225 | orchestrator | Friday 29 August 2025 17:25:48 +0000 (0:00:00.155) 0:00:39.928 ********* 2025-08-29 17:25:52.480236 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480246 | orchestrator | 2025-08-29 17:25:52.480257 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ******************** 2025-08-29 17:25:52.480267 | orchestrator | Friday 29 August 2025 17:25:49 +0000 (0:00:00.140) 0:00:40.068 ********* 2025-08-29 17:25:52.480278 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480288 | orchestrator | 2025-08-29 17:25:52.480329 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] ***************** 2025-08-29 17:25:52.480351 | orchestrator | Friday 29 August 2025 17:25:49 +0000 (0:00:00.130) 0:00:40.199 ********* 2025-08-29 17:25:52.480369 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480385 | orchestrator | 2025-08-29 17:25:52.480395 | orchestrator | TASK [Print number of OSDs wanted per DB VG] *********************************** 2025-08-29 17:25:52.480406 | orchestrator | Friday 29 August 2025 17:25:49 +0000 (0:00:00.125) 0:00:40.325 ********* 2025-08-29 17:25:52.480417 | orchestrator | ok: [testbed-node-4] => { 2025-08-29 17:25:52.480428 | orchestrator |  "_num_osds_wanted_per_db_vg": {} 2025-08-29 17:25:52.480439 | orchestrator | } 2025-08-29 17:25:52.480450 | orchestrator | 2025-08-29 17:25:52.480461 | orchestrator | TASK [Print number of OSDs wanted per WAL VG] ********************************** 2025-08-29 17:25:52.480472 | orchestrator | Friday 29 August 2025 17:25:49 +0000 (0:00:00.134) 0:00:40.459 ********* 2025-08-29 17:25:52.480482 | orchestrator | ok: [testbed-node-4] => { 2025-08-29 17:25:52.480493 | orchestrator |  "_num_osds_wanted_per_wal_vg": {} 2025-08-29 17:25:52.480504 | orchestrator | } 2025-08-29 17:25:52.480514 | orchestrator | 2025-08-29 17:25:52.480525 | orchestrator | TASK [Print number of OSDs wanted per DB+WAL VG] ******************************* 2025-08-29 17:25:52.480536 | orchestrator | Friday 29 August 2025 17:25:49 +0000 (0:00:00.138) 0:00:40.597 ********* 2025-08-29 17:25:52.480546 | orchestrator | ok: [testbed-node-4] => { 2025-08-29 17:25:52.480557 | orchestrator |  "_num_osds_wanted_per_db_wal_vg": {} 2025-08-29 17:25:52.480568 | orchestrator | } 2025-08-29 17:25:52.480591 | orchestrator | 2025-08-29 17:25:52.480602 | orchestrator | TASK [Gather DB VGs with total and available size in bytes] ******************** 2025-08-29 17:25:52.480613 | orchestrator | Friday 29 August 2025 17:25:49 +0000 (0:00:00.133) 0:00:40.731 ********* 2025-08-29 17:25:52.480624 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:52.480634 | orchestrator | 2025-08-29 17:25:52.480645 | orchestrator | TASK [Gather WAL VGs with total and available size in bytes] ******************* 2025-08-29 17:25:52.480656 | orchestrator | Friday 29 August 2025 17:25:50 +0000 (0:00:00.678) 0:00:41.410 ********* 2025-08-29 17:25:52.480666 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:52.480677 | orchestrator | 2025-08-29 17:25:52.480688 | orchestrator | TASK [Gather DB+WAL VGs with total and available size in bytes] **************** 2025-08-29 17:25:52.480699 | orchestrator | Friday 29 August 2025 17:25:50 +0000 (0:00:00.539) 0:00:41.949 ********* 2025-08-29 17:25:52.480710 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:52.480720 | orchestrator | 2025-08-29 17:25:52.480731 | orchestrator | TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] ************************* 2025-08-29 17:25:52.480742 | orchestrator | Friday 29 August 2025 17:25:51 +0000 (0:00:00.506) 0:00:42.455 ********* 2025-08-29 17:25:52.480752 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:52.480763 | orchestrator | 2025-08-29 17:25:52.480774 | orchestrator | TASK [Calculate VG sizes (without buffer)] ************************************* 2025-08-29 17:25:52.480784 | orchestrator | Friday 29 August 2025 17:25:51 +0000 (0:00:00.145) 0:00:42.600 ********* 2025-08-29 17:25:52.480795 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480806 | orchestrator | 2025-08-29 17:25:52.480816 | orchestrator | TASK [Calculate VG sizes (with buffer)] **************************************** 2025-08-29 17:25:52.480827 | orchestrator | Friday 29 August 2025 17:25:51 +0000 (0:00:00.112) 0:00:42.713 ********* 2025-08-29 17:25:52.480847 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480858 | orchestrator | 2025-08-29 17:25:52.480868 | orchestrator | TASK [Print LVM VGs report data] *********************************************** 2025-08-29 17:25:52.480879 | orchestrator | Friday 29 August 2025 17:25:51 +0000 (0:00:00.104) 0:00:42.818 ********* 2025-08-29 17:25:52.480890 | orchestrator | ok: [testbed-node-4] => { 2025-08-29 17:25:52.480901 | orchestrator |  "vgs_report": { 2025-08-29 17:25:52.480912 | orchestrator |  "vg": [] 2025-08-29 17:25:52.480922 | orchestrator |  } 2025-08-29 17:25:52.480933 | orchestrator | } 2025-08-29 17:25:52.480944 | orchestrator | 2025-08-29 17:25:52.480955 | orchestrator | TASK [Print LVM VG sizes] ****************************************************** 2025-08-29 17:25:52.480965 | orchestrator | Friday 29 August 2025 17:25:51 +0000 (0:00:00.131) 0:00:42.949 ********* 2025-08-29 17:25:52.480976 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.480987 | orchestrator | 2025-08-29 17:25:52.480997 | orchestrator | TASK [Calculate size needed for LVs on ceph_db_devices] ************************ 2025-08-29 17:25:52.481008 | orchestrator | Friday 29 August 2025 17:25:52 +0000 (0:00:00.131) 0:00:43.081 ********* 2025-08-29 17:25:52.481018 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.481029 | orchestrator | 2025-08-29 17:25:52.481040 | orchestrator | TASK [Print size needed for LVs on ceph_db_devices] **************************** 2025-08-29 17:25:52.481051 | orchestrator | Friday 29 August 2025 17:25:52 +0000 (0:00:00.141) 0:00:43.223 ********* 2025-08-29 17:25:52.481061 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.481072 | orchestrator | 2025-08-29 17:25:52.481083 | orchestrator | TASK [Fail if size of DB LVs on ceph_db_devices > available] ******************* 2025-08-29 17:25:52.481094 | orchestrator | Friday 29 August 2025 17:25:52 +0000 (0:00:00.135) 0:00:43.359 ********* 2025-08-29 17:25:52.481104 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:52.481115 | orchestrator | 2025-08-29 17:25:52.481126 | orchestrator | TASK [Calculate size needed for LVs on ceph_wal_devices] *********************** 2025-08-29 17:25:52.481144 | orchestrator | Friday 29 August 2025 17:25:52 +0000 (0:00:00.144) 0:00:43.503 ********* 2025-08-29 17:25:57.326399 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326497 | orchestrator | 2025-08-29 17:25:57.326515 | orchestrator | TASK [Print size needed for LVs on ceph_wal_devices] *************************** 2025-08-29 17:25:57.326550 | orchestrator | Friday 29 August 2025 17:25:52 +0000 (0:00:00.139) 0:00:43.642 ********* 2025-08-29 17:25:57.326562 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326573 | orchestrator | 2025-08-29 17:25:57.326584 | orchestrator | TASK [Fail if size of WAL LVs on ceph_wal_devices > available] ***************** 2025-08-29 17:25:57.326595 | orchestrator | Friday 29 August 2025 17:25:52 +0000 (0:00:00.332) 0:00:43.975 ********* 2025-08-29 17:25:57.326605 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326616 | orchestrator | 2025-08-29 17:25:57.326627 | orchestrator | TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] **************** 2025-08-29 17:25:57.326637 | orchestrator | Friday 29 August 2025 17:25:53 +0000 (0:00:00.139) 0:00:44.115 ********* 2025-08-29 17:25:57.326648 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326659 | orchestrator | 2025-08-29 17:25:57.326670 | orchestrator | TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ******************** 2025-08-29 17:25:57.326680 | orchestrator | Friday 29 August 2025 17:25:53 +0000 (0:00:00.146) 0:00:44.261 ********* 2025-08-29 17:25:57.326691 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326701 | orchestrator | 2025-08-29 17:25:57.326712 | orchestrator | TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] ***************** 2025-08-29 17:25:57.326723 | orchestrator | Friday 29 August 2025 17:25:53 +0000 (0:00:00.144) 0:00:44.406 ********* 2025-08-29 17:25:57.326733 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326744 | orchestrator | 2025-08-29 17:25:57.326754 | orchestrator | TASK [Print size needed for DB LVs on ceph_db_wal_devices] ********************* 2025-08-29 17:25:57.326765 | orchestrator | Friday 29 August 2025 17:25:53 +0000 (0:00:00.127) 0:00:44.534 ********* 2025-08-29 17:25:57.326776 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326786 | orchestrator | 2025-08-29 17:25:57.326797 | orchestrator | TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] *********** 2025-08-29 17:25:57.326808 | orchestrator | Friday 29 August 2025 17:25:53 +0000 (0:00:00.124) 0:00:44.658 ********* 2025-08-29 17:25:57.326819 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326829 | orchestrator | 2025-08-29 17:25:57.326840 | orchestrator | TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] ************************* 2025-08-29 17:25:57.326850 | orchestrator | Friday 29 August 2025 17:25:53 +0000 (0:00:00.148) 0:00:44.807 ********* 2025-08-29 17:25:57.326861 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326872 | orchestrator | 2025-08-29 17:25:57.326884 | orchestrator | TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] ********************* 2025-08-29 17:25:57.326897 | orchestrator | Friday 29 August 2025 17:25:53 +0000 (0:00:00.145) 0:00:44.952 ********* 2025-08-29 17:25:57.326909 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.326922 | orchestrator | 2025-08-29 17:25:57.326934 | orchestrator | TASK [Create DB LVs for ceph_db_devices] *************************************** 2025-08-29 17:25:57.326946 | orchestrator | Friday 29 August 2025 17:25:54 +0000 (0:00:00.131) 0:00:45.084 ********* 2025-08-29 17:25:57.326973 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.326987 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327001 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327021 | orchestrator | 2025-08-29 17:25:57.327039 | orchestrator | TASK [Print 'Create DB LVs for ceph_db_devices'] ******************************* 2025-08-29 17:25:57.327058 | orchestrator | Friday 29 August 2025 17:25:54 +0000 (0:00:00.159) 0:00:45.244 ********* 2025-08-29 17:25:57.327076 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327095 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327125 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327146 | orchestrator | 2025-08-29 17:25:57.327167 | orchestrator | TASK [Create WAL LVs for ceph_wal_devices] ************************************* 2025-08-29 17:25:57.327180 | orchestrator | Friday 29 August 2025 17:25:54 +0000 (0:00:00.163) 0:00:45.408 ********* 2025-08-29 17:25:57.327193 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327205 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327218 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327230 | orchestrator | 2025-08-29 17:25:57.327241 | orchestrator | TASK [Print 'Create WAL LVs for ceph_wal_devices'] ***************************** 2025-08-29 17:25:57.327252 | orchestrator | Friday 29 August 2025 17:25:54 +0000 (0:00:00.178) 0:00:45.586 ********* 2025-08-29 17:25:57.327262 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327273 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327284 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327295 | orchestrator | 2025-08-29 17:25:57.327325 | orchestrator | TASK [Create WAL LVs for ceph_db_wal_devices] ********************************** 2025-08-29 17:25:57.327355 | orchestrator | Friday 29 August 2025 17:25:54 +0000 (0:00:00.435) 0:00:46.022 ********* 2025-08-29 17:25:57.327366 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327377 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327388 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327399 | orchestrator | 2025-08-29 17:25:57.327410 | orchestrator | TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] ************************** 2025-08-29 17:25:57.327421 | orchestrator | Friday 29 August 2025 17:25:55 +0000 (0:00:00.197) 0:00:46.219 ********* 2025-08-29 17:25:57.327431 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327442 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327452 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327463 | orchestrator | 2025-08-29 17:25:57.327474 | orchestrator | TASK [Create DB LVs for ceph_db_wal_devices] *********************************** 2025-08-29 17:25:57.327485 | orchestrator | Friday 29 August 2025 17:25:55 +0000 (0:00:00.179) 0:00:46.398 ********* 2025-08-29 17:25:57.327496 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327506 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327517 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327527 | orchestrator | 2025-08-29 17:25:57.327538 | orchestrator | TASK [Print 'Create DB LVs for ceph_db_wal_devices'] *************************** 2025-08-29 17:25:57.327548 | orchestrator | Friday 29 August 2025 17:25:55 +0000 (0:00:00.159) 0:00:46.558 ********* 2025-08-29 17:25:57.327559 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327569 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327587 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327598 | orchestrator | 2025-08-29 17:25:57.327608 | orchestrator | TASK [Get list of Ceph LVs with associated VGs] ******************************** 2025-08-29 17:25:57.327624 | orchestrator | Friday 29 August 2025 17:25:55 +0000 (0:00:00.144) 0:00:46.702 ********* 2025-08-29 17:25:57.327635 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:57.327646 | orchestrator | 2025-08-29 17:25:57.327657 | orchestrator | TASK [Get list of Ceph PVs with associated VGs] ******************************** 2025-08-29 17:25:57.327667 | orchestrator | Friday 29 August 2025 17:25:56 +0000 (0:00:00.532) 0:00:47.234 ********* 2025-08-29 17:25:57.327678 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:57.327688 | orchestrator | 2025-08-29 17:25:57.327699 | orchestrator | TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] *********************** 2025-08-29 17:25:57.327709 | orchestrator | Friday 29 August 2025 17:25:56 +0000 (0:00:00.505) 0:00:47.740 ********* 2025-08-29 17:25:57.327720 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:25:57.327730 | orchestrator | 2025-08-29 17:25:57.327741 | orchestrator | TASK [Create list of VG/LV names] ********************************************** 2025-08-29 17:25:57.327752 | orchestrator | Friday 29 August 2025 17:25:56 +0000 (0:00:00.161) 0:00:47.901 ********* 2025-08-29 17:25:57.327762 | orchestrator | ok: [testbed-node-4] => (item={'lv_name': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'vg_name': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'}) 2025-08-29 17:25:57.327773 | orchestrator | ok: [testbed-node-4] => (item={'lv_name': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'vg_name': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'}) 2025-08-29 17:25:57.327784 | orchestrator | 2025-08-29 17:25:57.327795 | orchestrator | TASK [Fail if block LV defined in lvm_volumes is missing] ********************** 2025-08-29 17:25:57.327805 | orchestrator | Friday 29 August 2025 17:25:57 +0000 (0:00:00.157) 0:00:48.059 ********* 2025-08-29 17:25:57.327816 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327826 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327837 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:25:57.327847 | orchestrator | 2025-08-29 17:25:57.327858 | orchestrator | TASK [Fail if DB LV defined in lvm_volumes is missing] ************************* 2025-08-29 17:25:57.327869 | orchestrator | Friday 29 August 2025 17:25:57 +0000 (0:00:00.137) 0:00:48.196 ********* 2025-08-29 17:25:57.327879 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:25:57.327890 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:25:57.327907 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:26:02.444187 | orchestrator | 2025-08-29 17:26:02.444335 | orchestrator | TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************ 2025-08-29 17:26:02.444355 | orchestrator | Friday 29 August 2025 17:25:57 +0000 (0:00:00.153) 0:00:48.349 ********* 2025-08-29 17:26:02.444409 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'})  2025-08-29 17:26:02.444424 | orchestrator | skipping: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'})  2025-08-29 17:26:02.444435 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:26:02.444448 | orchestrator | 2025-08-29 17:26:02.444459 | orchestrator | TASK [Print LVM report data] *************************************************** 2025-08-29 17:26:02.444470 | orchestrator | Friday 29 August 2025 17:25:57 +0000 (0:00:00.151) 0:00:48.501 ********* 2025-08-29 17:26:02.444502 | orchestrator | ok: [testbed-node-4] => { 2025-08-29 17:26:02.444513 | orchestrator |  "lvm_report": { 2025-08-29 17:26:02.444525 | orchestrator |  "lv": [ 2025-08-29 17:26:02.444536 | orchestrator |  { 2025-08-29 17:26:02.444548 | orchestrator |  "lv_name": "osd-block-860f9296-4089-56f8-8238-0f24b03dbca2", 2025-08-29 17:26:02.444559 | orchestrator |  "vg_name": "ceph-860f9296-4089-56f8-8238-0f24b03dbca2" 2025-08-29 17:26:02.444570 | orchestrator |  }, 2025-08-29 17:26:02.444581 | orchestrator |  { 2025-08-29 17:26:02.444592 | orchestrator |  "lv_name": "osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2", 2025-08-29 17:26:02.444602 | orchestrator |  "vg_name": "ceph-b89384ec-5219-5f2a-8735-84f78c8179d2" 2025-08-29 17:26:02.444613 | orchestrator |  } 2025-08-29 17:26:02.444624 | orchestrator |  ], 2025-08-29 17:26:02.444635 | orchestrator |  "pv": [ 2025-08-29 17:26:02.444645 | orchestrator |  { 2025-08-29 17:26:02.444656 | orchestrator |  "pv_name": "/dev/sdb", 2025-08-29 17:26:02.444667 | orchestrator |  "vg_name": "ceph-b89384ec-5219-5f2a-8735-84f78c8179d2" 2025-08-29 17:26:02.444678 | orchestrator |  }, 2025-08-29 17:26:02.444689 | orchestrator |  { 2025-08-29 17:26:02.444700 | orchestrator |  "pv_name": "/dev/sdc", 2025-08-29 17:26:02.444711 | orchestrator |  "vg_name": "ceph-860f9296-4089-56f8-8238-0f24b03dbca2" 2025-08-29 17:26:02.444721 | orchestrator |  } 2025-08-29 17:26:02.444732 | orchestrator |  ] 2025-08-29 17:26:02.444745 | orchestrator |  } 2025-08-29 17:26:02.444758 | orchestrator | } 2025-08-29 17:26:02.444771 | orchestrator | 2025-08-29 17:26:02.444783 | orchestrator | PLAY [Ceph create LVM devices] ************************************************* 2025-08-29 17:26:02.444796 | orchestrator | 2025-08-29 17:26:02.444808 | orchestrator | TASK [Get extra vars for Ceph configuration] *********************************** 2025-08-29 17:26:02.444820 | orchestrator | Friday 29 August 2025 17:25:57 +0000 (0:00:00.370) 0:00:48.872 ********* 2025-08-29 17:26:02.444833 | orchestrator | ok: [testbed-node-5 -> testbed-manager(192.168.16.5)] 2025-08-29 17:26:02.444846 | orchestrator | 2025-08-29 17:26:02.444859 | orchestrator | TASK [Get initial list of available block devices] ***************************** 2025-08-29 17:26:02.444872 | orchestrator | Friday 29 August 2025 17:25:58 +0000 (0:00:00.208) 0:00:49.080 ********* 2025-08-29 17:26:02.444884 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:02.444896 | orchestrator | 2025-08-29 17:26:02.444909 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.444922 | orchestrator | Friday 29 August 2025 17:25:58 +0000 (0:00:00.203) 0:00:49.284 ********* 2025-08-29 17:26:02.444934 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0) 2025-08-29 17:26:02.444946 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1) 2025-08-29 17:26:02.444958 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2) 2025-08-29 17:26:02.444971 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3) 2025-08-29 17:26:02.444983 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4) 2025-08-29 17:26:02.444996 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5) 2025-08-29 17:26:02.445008 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6) 2025-08-29 17:26:02.445021 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7) 2025-08-29 17:26:02.445033 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda) 2025-08-29 17:26:02.445045 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb) 2025-08-29 17:26:02.445057 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc) 2025-08-29 17:26:02.445079 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd) 2025-08-29 17:26:02.445092 | orchestrator | included: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0) 2025-08-29 17:26:02.445103 | orchestrator | 2025-08-29 17:26:02.445113 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445124 | orchestrator | Friday 29 August 2025 17:25:58 +0000 (0:00:00.379) 0:00:49.664 ********* 2025-08-29 17:26:02.445135 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445145 | orchestrator | 2025-08-29 17:26:02.445160 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445171 | orchestrator | Friday 29 August 2025 17:25:58 +0000 (0:00:00.171) 0:00:49.836 ********* 2025-08-29 17:26:02.445182 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445193 | orchestrator | 2025-08-29 17:26:02.445204 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445230 | orchestrator | Friday 29 August 2025 17:25:58 +0000 (0:00:00.180) 0:00:50.016 ********* 2025-08-29 17:26:02.445242 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445253 | orchestrator | 2025-08-29 17:26:02.445264 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445275 | orchestrator | Friday 29 August 2025 17:25:59 +0000 (0:00:00.179) 0:00:50.196 ********* 2025-08-29 17:26:02.445286 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445296 | orchestrator | 2025-08-29 17:26:02.445326 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445337 | orchestrator | Friday 29 August 2025 17:25:59 +0000 (0:00:00.172) 0:00:50.368 ********* 2025-08-29 17:26:02.445348 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445359 | orchestrator | 2025-08-29 17:26:02.445409 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445422 | orchestrator | Friday 29 August 2025 17:25:59 +0000 (0:00:00.165) 0:00:50.534 ********* 2025-08-29 17:26:02.445433 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445443 | orchestrator | 2025-08-29 17:26:02.445454 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445465 | orchestrator | Friday 29 August 2025 17:25:59 +0000 (0:00:00.424) 0:00:50.959 ********* 2025-08-29 17:26:02.445476 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445487 | orchestrator | 2025-08-29 17:26:02.445498 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445508 | orchestrator | Friday 29 August 2025 17:26:00 +0000 (0:00:00.171) 0:00:51.130 ********* 2025-08-29 17:26:02.445519 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:02.445530 | orchestrator | 2025-08-29 17:26:02.445541 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445551 | orchestrator | Friday 29 August 2025 17:26:00 +0000 (0:00:00.169) 0:00:51.300 ********* 2025-08-29 17:26:02.445562 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f) 2025-08-29 17:26:02.445574 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f) 2025-08-29 17:26:02.445585 | orchestrator | 2025-08-29 17:26:02.445596 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445606 | orchestrator | Friday 29 August 2025 17:26:00 +0000 (0:00:00.374) 0:00:51.674 ********* 2025-08-29 17:26:02.445617 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6) 2025-08-29 17:26:02.445628 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6) 2025-08-29 17:26:02.445639 | orchestrator | 2025-08-29 17:26:02.445650 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445660 | orchestrator | Friday 29 August 2025 17:26:01 +0000 (0:00:00.363) 0:00:52.038 ********* 2025-08-29 17:26:02.445676 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6) 2025-08-29 17:26:02.445694 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6) 2025-08-29 17:26:02.445705 | orchestrator | 2025-08-29 17:26:02.445716 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445727 | orchestrator | Friday 29 August 2025 17:26:01 +0000 (0:00:00.379) 0:00:52.418 ********* 2025-08-29 17:26:02.445737 | orchestrator | ok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290) 2025-08-29 17:26:02.445748 | orchestrator | ok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290) 2025-08-29 17:26:02.445759 | orchestrator | 2025-08-29 17:26:02.445770 | orchestrator | TASK [Add known links to the list of available block devices] ****************** 2025-08-29 17:26:02.445780 | orchestrator | Friday 29 August 2025 17:26:01 +0000 (0:00:00.410) 0:00:52.828 ********* 2025-08-29 17:26:02.445791 | orchestrator | ok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001) 2025-08-29 17:26:02.445802 | orchestrator | 2025-08-29 17:26:02.445813 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:02.445823 | orchestrator | Friday 29 August 2025 17:26:02 +0000 (0:00:00.279) 0:00:53.108 ********* 2025-08-29 17:26:02.445834 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0) 2025-08-29 17:26:02.445845 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1) 2025-08-29 17:26:02.445856 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2) 2025-08-29 17:26:02.445866 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3) 2025-08-29 17:26:02.445877 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4) 2025-08-29 17:26:02.445888 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5) 2025-08-29 17:26:02.445899 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6) 2025-08-29 17:26:02.445909 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7) 2025-08-29 17:26:02.445920 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda) 2025-08-29 17:26:02.445931 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb) 2025-08-29 17:26:02.445942 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc) 2025-08-29 17:26:02.445959 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd) 2025-08-29 17:26:12.006420 | orchestrator | included: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0) 2025-08-29 17:26:12.006536 | orchestrator | 2025-08-29 17:26:12.006553 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006565 | orchestrator | Friday 29 August 2025 17:26:02 +0000 (0:00:00.351) 0:00:53.459 ********* 2025-08-29 17:26:12.006577 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006589 | orchestrator | 2025-08-29 17:26:12.006601 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006612 | orchestrator | Friday 29 August 2025 17:26:02 +0000 (0:00:00.185) 0:00:53.645 ********* 2025-08-29 17:26:12.006623 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006633 | orchestrator | 2025-08-29 17:26:12.006645 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006656 | orchestrator | Friday 29 August 2025 17:26:02 +0000 (0:00:00.206) 0:00:53.852 ********* 2025-08-29 17:26:12.006667 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006677 | orchestrator | 2025-08-29 17:26:12.006688 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006699 | orchestrator | Friday 29 August 2025 17:26:03 +0000 (0:00:00.686) 0:00:54.538 ********* 2025-08-29 17:26:12.006734 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006746 | orchestrator | 2025-08-29 17:26:12.006757 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006767 | orchestrator | Friday 29 August 2025 17:26:03 +0000 (0:00:00.212) 0:00:54.751 ********* 2025-08-29 17:26:12.006778 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006789 | orchestrator | 2025-08-29 17:26:12.006800 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006810 | orchestrator | Friday 29 August 2025 17:26:03 +0000 (0:00:00.230) 0:00:54.981 ********* 2025-08-29 17:26:12.006821 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006832 | orchestrator | 2025-08-29 17:26:12.006842 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006853 | orchestrator | Friday 29 August 2025 17:26:04 +0000 (0:00:00.217) 0:00:55.199 ********* 2025-08-29 17:26:12.006864 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006875 | orchestrator | 2025-08-29 17:26:12.006885 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006896 | orchestrator | Friday 29 August 2025 17:26:04 +0000 (0:00:00.241) 0:00:55.440 ********* 2025-08-29 17:26:12.006907 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.006917 | orchestrator | 2025-08-29 17:26:12.006928 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.006939 | orchestrator | Friday 29 August 2025 17:26:04 +0000 (0:00:00.200) 0:00:55.641 ********* 2025-08-29 17:26:12.006950 | orchestrator | ok: [testbed-node-5] => (item=sda1) 2025-08-29 17:26:12.006961 | orchestrator | ok: [testbed-node-5] => (item=sda14) 2025-08-29 17:26:12.006972 | orchestrator | ok: [testbed-node-5] => (item=sda15) 2025-08-29 17:26:12.006998 | orchestrator | ok: [testbed-node-5] => (item=sda16) 2025-08-29 17:26:12.007009 | orchestrator | 2025-08-29 17:26:12.007020 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.007031 | orchestrator | Friday 29 August 2025 17:26:05 +0000 (0:00:00.660) 0:00:56.301 ********* 2025-08-29 17:26:12.007042 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007053 | orchestrator | 2025-08-29 17:26:12.007063 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.007074 | orchestrator | Friday 29 August 2025 17:26:05 +0000 (0:00:00.226) 0:00:56.528 ********* 2025-08-29 17:26:12.007085 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007095 | orchestrator | 2025-08-29 17:26:12.007106 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.007118 | orchestrator | Friday 29 August 2025 17:26:05 +0000 (0:00:00.210) 0:00:56.738 ********* 2025-08-29 17:26:12.007128 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007139 | orchestrator | 2025-08-29 17:26:12.007150 | orchestrator | TASK [Add known partitions to the list of available block devices] ************* 2025-08-29 17:26:12.007161 | orchestrator | Friday 29 August 2025 17:26:05 +0000 (0:00:00.206) 0:00:56.945 ********* 2025-08-29 17:26:12.007171 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007182 | orchestrator | 2025-08-29 17:26:12.007193 | orchestrator | TASK [Check whether ceph_db_wal_devices is used exclusively] ******************* 2025-08-29 17:26:12.007204 | orchestrator | Friday 29 August 2025 17:26:06 +0000 (0:00:00.225) 0:00:57.171 ********* 2025-08-29 17:26:12.007214 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007225 | orchestrator | 2025-08-29 17:26:12.007236 | orchestrator | TASK [Create dict of block VGs -> PVs from ceph_osd_devices] ******************* 2025-08-29 17:26:12.007246 | orchestrator | Friday 29 August 2025 17:26:06 +0000 (0:00:00.405) 0:00:57.576 ********* 2025-08-29 17:26:12.007257 | orchestrator | ok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': 'e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}}) 2025-08-29 17:26:12.007269 | orchestrator | ok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '21b183ef-757d-561c-bead-7bb3aee28288'}}) 2025-08-29 17:26:12.007287 | orchestrator | 2025-08-29 17:26:12.007298 | orchestrator | TASK [Create block VGs] ******************************************************** 2025-08-29 17:26:12.007347 | orchestrator | Friday 29 August 2025 17:26:06 +0000 (0:00:00.215) 0:00:57.792 ********* 2025-08-29 17:26:12.007360 | orchestrator | changed: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}) 2025-08-29 17:26:12.007373 | orchestrator | changed: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'}) 2025-08-29 17:26:12.007384 | orchestrator | 2025-08-29 17:26:12.007395 | orchestrator | TASK [Print 'Create block VGs'] ************************************************ 2025-08-29 17:26:12.007422 | orchestrator | Friday 29 August 2025 17:26:08 +0000 (0:00:01.926) 0:00:59.718 ********* 2025-08-29 17:26:12.007434 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:12.007446 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:12.007457 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007468 | orchestrator | 2025-08-29 17:26:12.007479 | orchestrator | TASK [Create block LVs] ******************************************************** 2025-08-29 17:26:12.007490 | orchestrator | Friday 29 August 2025 17:26:08 +0000 (0:00:00.164) 0:00:59.882 ********* 2025-08-29 17:26:12.007501 | orchestrator | changed: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}) 2025-08-29 17:26:12.007512 | orchestrator | changed: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'}) 2025-08-29 17:26:12.007523 | orchestrator | 2025-08-29 17:26:12.007535 | orchestrator | TASK [Print 'Create block LVs'] ************************************************ 2025-08-29 17:26:12.007546 | orchestrator | Friday 29 August 2025 17:26:10 +0000 (0:00:01.439) 0:01:01.322 ********* 2025-08-29 17:26:12.007557 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:12.007568 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:12.007579 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007590 | orchestrator | 2025-08-29 17:26:12.007601 | orchestrator | TASK [Create DB VGs] *********************************************************** 2025-08-29 17:26:12.007612 | orchestrator | Friday 29 August 2025 17:26:10 +0000 (0:00:00.156) 0:01:01.478 ********* 2025-08-29 17:26:12.007622 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007633 | orchestrator | 2025-08-29 17:26:12.007644 | orchestrator | TASK [Print 'Create DB VGs'] *************************************************** 2025-08-29 17:26:12.007655 | orchestrator | Friday 29 August 2025 17:26:10 +0000 (0:00:00.145) 0:01:01.624 ********* 2025-08-29 17:26:12.007666 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:12.007683 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:12.007694 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007705 | orchestrator | 2025-08-29 17:26:12.007716 | orchestrator | TASK [Create WAL VGs] ********************************************************** 2025-08-29 17:26:12.007727 | orchestrator | Friday 29 August 2025 17:26:10 +0000 (0:00:00.155) 0:01:01.779 ********* 2025-08-29 17:26:12.007738 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007749 | orchestrator | 2025-08-29 17:26:12.007760 | orchestrator | TASK [Print 'Create WAL VGs'] ************************************************** 2025-08-29 17:26:12.007778 | orchestrator | Friday 29 August 2025 17:26:10 +0000 (0:00:00.145) 0:01:01.925 ********* 2025-08-29 17:26:12.007789 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:12.007800 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:12.007811 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007822 | orchestrator | 2025-08-29 17:26:12.007833 | orchestrator | TASK [Create DB+WAL VGs] ******************************************************* 2025-08-29 17:26:12.007843 | orchestrator | Friday 29 August 2025 17:26:11 +0000 (0:00:00.170) 0:01:02.096 ********* 2025-08-29 17:26:12.007854 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007865 | orchestrator | 2025-08-29 17:26:12.007876 | orchestrator | TASK [Print 'Create DB+WAL VGs'] *********************************************** 2025-08-29 17:26:12.007887 | orchestrator | Friday 29 August 2025 17:26:11 +0000 (0:00:00.142) 0:01:02.238 ********* 2025-08-29 17:26:12.007898 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:12.007909 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:12.007920 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:12.007930 | orchestrator | 2025-08-29 17:26:12.007941 | orchestrator | TASK [Prepare variables for OSD count check] *********************************** 2025-08-29 17:26:12.007952 | orchestrator | Friday 29 August 2025 17:26:11 +0000 (0:00:00.169) 0:01:02.408 ********* 2025-08-29 17:26:12.007963 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:12.007974 | orchestrator | 2025-08-29 17:26:12.007985 | orchestrator | TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] **************** 2025-08-29 17:26:12.007996 | orchestrator | Friday 29 August 2025 17:26:11 +0000 (0:00:00.157) 0:01:02.566 ********* 2025-08-29 17:26:12.008013 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:18.184674 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:18.184804 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.184830 | orchestrator | 2025-08-29 17:26:18.184849 | orchestrator | TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] *************** 2025-08-29 17:26:18.184868 | orchestrator | Friday 29 August 2025 17:26:11 +0000 (0:00:00.463) 0:01:03.029 ********* 2025-08-29 17:26:18.184885 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:18.184903 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:18.184920 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.184933 | orchestrator | 2025-08-29 17:26:18.184944 | orchestrator | TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************ 2025-08-29 17:26:18.184954 | orchestrator | Friday 29 August 2025 17:26:12 +0000 (0:00:00.176) 0:01:03.206 ********* 2025-08-29 17:26:18.184964 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:18.184974 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:18.184984 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.184993 | orchestrator | 2025-08-29 17:26:18.185026 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a DB VG] ********************* 2025-08-29 17:26:18.185037 | orchestrator | Friday 29 August 2025 17:26:12 +0000 (0:00:00.154) 0:01:03.360 ********* 2025-08-29 17:26:18.185046 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.185056 | orchestrator | 2025-08-29 17:26:18.185065 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ******************** 2025-08-29 17:26:18.185074 | orchestrator | Friday 29 August 2025 17:26:12 +0000 (0:00:00.153) 0:01:03.515 ********* 2025-08-29 17:26:18.185083 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.185093 | orchestrator | 2025-08-29 17:26:18.185102 | orchestrator | TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] ***************** 2025-08-29 17:26:18.185111 | orchestrator | Friday 29 August 2025 17:26:12 +0000 (0:00:00.121) 0:01:03.636 ********* 2025-08-29 17:26:18.185120 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.185130 | orchestrator | 2025-08-29 17:26:18.185139 | orchestrator | TASK [Print number of OSDs wanted per DB VG] *********************************** 2025-08-29 17:26:18.185149 | orchestrator | Friday 29 August 2025 17:26:12 +0000 (0:00:00.139) 0:01:03.776 ********* 2025-08-29 17:26:18.185158 | orchestrator | ok: [testbed-node-5] => { 2025-08-29 17:26:18.185169 | orchestrator |  "_num_osds_wanted_per_db_vg": {} 2025-08-29 17:26:18.185178 | orchestrator | } 2025-08-29 17:26:18.185188 | orchestrator | 2025-08-29 17:26:18.185199 | orchestrator | TASK [Print number of OSDs wanted per WAL VG] ********************************** 2025-08-29 17:26:18.185214 | orchestrator | Friday 29 August 2025 17:26:12 +0000 (0:00:00.168) 0:01:03.944 ********* 2025-08-29 17:26:18.185227 | orchestrator | ok: [testbed-node-5] => { 2025-08-29 17:26:18.185238 | orchestrator |  "_num_osds_wanted_per_wal_vg": {} 2025-08-29 17:26:18.185249 | orchestrator | } 2025-08-29 17:26:18.185260 | orchestrator | 2025-08-29 17:26:18.185271 | orchestrator | TASK [Print number of OSDs wanted per DB+WAL VG] ******************************* 2025-08-29 17:26:18.185281 | orchestrator | Friday 29 August 2025 17:26:13 +0000 (0:00:00.147) 0:01:04.091 ********* 2025-08-29 17:26:18.185292 | orchestrator | ok: [testbed-node-5] => { 2025-08-29 17:26:18.185341 | orchestrator |  "_num_osds_wanted_per_db_wal_vg": {} 2025-08-29 17:26:18.185355 | orchestrator | } 2025-08-29 17:26:18.185366 | orchestrator | 2025-08-29 17:26:18.185383 | orchestrator | TASK [Gather DB VGs with total and available size in bytes] ******************** 2025-08-29 17:26:18.185401 | orchestrator | Friday 29 August 2025 17:26:13 +0000 (0:00:00.143) 0:01:04.235 ********* 2025-08-29 17:26:18.185416 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:18.185427 | orchestrator | 2025-08-29 17:26:18.185438 | orchestrator | TASK [Gather WAL VGs with total and available size in bytes] ******************* 2025-08-29 17:26:18.185449 | orchestrator | Friday 29 August 2025 17:26:13 +0000 (0:00:00.527) 0:01:04.762 ********* 2025-08-29 17:26:18.185466 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:18.185485 | orchestrator | 2025-08-29 17:26:18.185504 | orchestrator | TASK [Gather DB+WAL VGs with total and available size in bytes] **************** 2025-08-29 17:26:18.185522 | orchestrator | Friday 29 August 2025 17:26:14 +0000 (0:00:00.532) 0:01:05.295 ********* 2025-08-29 17:26:18.185541 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:18.185558 | orchestrator | 2025-08-29 17:26:18.185571 | orchestrator | TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] ************************* 2025-08-29 17:26:18.185581 | orchestrator | Friday 29 August 2025 17:26:14 +0000 (0:00:00.542) 0:01:05.838 ********* 2025-08-29 17:26:18.185591 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:18.185600 | orchestrator | 2025-08-29 17:26:18.185610 | orchestrator | TASK [Calculate VG sizes (without buffer)] ************************************* 2025-08-29 17:26:18.185620 | orchestrator | Friday 29 August 2025 17:26:15 +0000 (0:00:00.324) 0:01:06.162 ********* 2025-08-29 17:26:18.185633 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.185650 | orchestrator | 2025-08-29 17:26:18.185665 | orchestrator | TASK [Calculate VG sizes (with buffer)] **************************************** 2025-08-29 17:26:18.185675 | orchestrator | Friday 29 August 2025 17:26:15 +0000 (0:00:00.111) 0:01:06.273 ********* 2025-08-29 17:26:18.185684 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.185703 | orchestrator | 2025-08-29 17:26:18.185714 | orchestrator | TASK [Print LVM VGs report data] *********************************************** 2025-08-29 17:26:18.185731 | orchestrator | Friday 29 August 2025 17:26:15 +0000 (0:00:00.111) 0:01:06.385 ********* 2025-08-29 17:26:18.185749 | orchestrator | ok: [testbed-node-5] => { 2025-08-29 17:26:18.185786 | orchestrator |  "vgs_report": { 2025-08-29 17:26:18.185804 | orchestrator |  "vg": [] 2025-08-29 17:26:18.185844 | orchestrator |  } 2025-08-29 17:26:18.185858 | orchestrator | } 2025-08-29 17:26:18.185867 | orchestrator | 2025-08-29 17:26:18.185877 | orchestrator | TASK [Print LVM VG sizes] ****************************************************** 2025-08-29 17:26:18.185889 | orchestrator | Friday 29 August 2025 17:26:15 +0000 (0:00:00.145) 0:01:06.530 ********* 2025-08-29 17:26:18.185907 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.185925 | orchestrator | 2025-08-29 17:26:18.185943 | orchestrator | TASK [Calculate size needed for LVs on ceph_db_devices] ************************ 2025-08-29 17:26:18.185961 | orchestrator | Friday 29 August 2025 17:26:15 +0000 (0:00:00.134) 0:01:06.665 ********* 2025-08-29 17:26:18.185979 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.185997 | orchestrator | 2025-08-29 17:26:18.186007 | orchestrator | TASK [Print size needed for LVs on ceph_db_devices] **************************** 2025-08-29 17:26:18.186068 | orchestrator | Friday 29 August 2025 17:26:15 +0000 (0:00:00.125) 0:01:06.791 ********* 2025-08-29 17:26:18.186079 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186089 | orchestrator | 2025-08-29 17:26:18.186098 | orchestrator | TASK [Fail if size of DB LVs on ceph_db_devices > available] ******************* 2025-08-29 17:26:18.186107 | orchestrator | Friday 29 August 2025 17:26:15 +0000 (0:00:00.132) 0:01:06.924 ********* 2025-08-29 17:26:18.186117 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186126 | orchestrator | 2025-08-29 17:26:18.186135 | orchestrator | TASK [Calculate size needed for LVs on ceph_wal_devices] *********************** 2025-08-29 17:26:18.186145 | orchestrator | Friday 29 August 2025 17:26:16 +0000 (0:00:00.127) 0:01:07.051 ********* 2025-08-29 17:26:18.186154 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186164 | orchestrator | 2025-08-29 17:26:18.186173 | orchestrator | TASK [Print size needed for LVs on ceph_wal_devices] *************************** 2025-08-29 17:26:18.186182 | orchestrator | Friday 29 August 2025 17:26:16 +0000 (0:00:00.142) 0:01:07.193 ********* 2025-08-29 17:26:18.186192 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186201 | orchestrator | 2025-08-29 17:26:18.186211 | orchestrator | TASK [Fail if size of WAL LVs on ceph_wal_devices > available] ***************** 2025-08-29 17:26:18.186220 | orchestrator | Friday 29 August 2025 17:26:16 +0000 (0:00:00.152) 0:01:07.346 ********* 2025-08-29 17:26:18.186231 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186246 | orchestrator | 2025-08-29 17:26:18.186256 | orchestrator | TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] **************** 2025-08-29 17:26:18.186265 | orchestrator | Friday 29 August 2025 17:26:16 +0000 (0:00:00.129) 0:01:07.475 ********* 2025-08-29 17:26:18.186275 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186284 | orchestrator | 2025-08-29 17:26:18.186298 | orchestrator | TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ******************** 2025-08-29 17:26:18.186351 | orchestrator | Friday 29 August 2025 17:26:16 +0000 (0:00:00.139) 0:01:07.615 ********* 2025-08-29 17:26:18.186369 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186387 | orchestrator | 2025-08-29 17:26:18.186404 | orchestrator | TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] ***************** 2025-08-29 17:26:18.186420 | orchestrator | Friday 29 August 2025 17:26:16 +0000 (0:00:00.293) 0:01:07.908 ********* 2025-08-29 17:26:18.186444 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186463 | orchestrator | 2025-08-29 17:26:18.186480 | orchestrator | TASK [Print size needed for DB LVs on ceph_db_wal_devices] ********************* 2025-08-29 17:26:18.186498 | orchestrator | Friday 29 August 2025 17:26:17 +0000 (0:00:00.157) 0:01:08.066 ********* 2025-08-29 17:26:18.186516 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186534 | orchestrator | 2025-08-29 17:26:18.186552 | orchestrator | TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] *********** 2025-08-29 17:26:18.186581 | orchestrator | Friday 29 August 2025 17:26:17 +0000 (0:00:00.162) 0:01:08.228 ********* 2025-08-29 17:26:18.186598 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186615 | orchestrator | 2025-08-29 17:26:18.186632 | orchestrator | TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] ************************* 2025-08-29 17:26:18.186650 | orchestrator | Friday 29 August 2025 17:26:17 +0000 (0:00:00.159) 0:01:08.388 ********* 2025-08-29 17:26:18.186668 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186686 | orchestrator | 2025-08-29 17:26:18.186703 | orchestrator | TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] ********************* 2025-08-29 17:26:18.186720 | orchestrator | Friday 29 August 2025 17:26:17 +0000 (0:00:00.154) 0:01:08.542 ********* 2025-08-29 17:26:18.186738 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186752 | orchestrator | 2025-08-29 17:26:18.186762 | orchestrator | TASK [Create DB LVs for ceph_db_devices] *************************************** 2025-08-29 17:26:18.186775 | orchestrator | Friday 29 August 2025 17:26:17 +0000 (0:00:00.134) 0:01:08.676 ********* 2025-08-29 17:26:18.186793 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:18.186811 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:18.186828 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186847 | orchestrator | 2025-08-29 17:26:18.186868 | orchestrator | TASK [Print 'Create DB LVs for ceph_db_devices'] ******************************* 2025-08-29 17:26:18.186887 | orchestrator | Friday 29 August 2025 17:26:17 +0000 (0:00:00.185) 0:01:08.862 ********* 2025-08-29 17:26:18.186904 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:18.186920 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:18.186937 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:18.186953 | orchestrator | 2025-08-29 17:26:18.186970 | orchestrator | TASK [Create WAL LVs for ceph_wal_devices] ************************************* 2025-08-29 17:26:18.186987 | orchestrator | Friday 29 August 2025 17:26:18 +0000 (0:00:00.169) 0:01:09.032 ********* 2025-08-29 17:26:18.187016 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.290505 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.290615 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.290639 | orchestrator | 2025-08-29 17:26:21.290652 | orchestrator | TASK [Print 'Create WAL LVs for ceph_wal_devices'] ***************************** 2025-08-29 17:26:21.290665 | orchestrator | Friday 29 August 2025 17:26:18 +0000 (0:00:00.178) 0:01:09.210 ********* 2025-08-29 17:26:21.290676 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.290687 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.290698 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.290708 | orchestrator | 2025-08-29 17:26:21.290719 | orchestrator | TASK [Create WAL LVs for ceph_db_wal_devices] ********************************** 2025-08-29 17:26:21.290730 | orchestrator | Friday 29 August 2025 17:26:18 +0000 (0:00:00.151) 0:01:09.362 ********* 2025-08-29 17:26:21.290740 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.290776 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.290787 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.290798 | orchestrator | 2025-08-29 17:26:21.290808 | orchestrator | TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] ************************** 2025-08-29 17:26:21.290819 | orchestrator | Friday 29 August 2025 17:26:18 +0000 (0:00:00.156) 0:01:09.518 ********* 2025-08-29 17:26:21.290829 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.290840 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.290851 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.290861 | orchestrator | 2025-08-29 17:26:21.290872 | orchestrator | TASK [Create DB LVs for ceph_db_wal_devices] *********************************** 2025-08-29 17:26:21.290897 | orchestrator | Friday 29 August 2025 17:26:18 +0000 (0:00:00.153) 0:01:09.672 ********* 2025-08-29 17:26:21.290908 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.290919 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.290929 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.290939 | orchestrator | 2025-08-29 17:26:21.290950 | orchestrator | TASK [Print 'Create DB LVs for ceph_db_wal_devices'] *************************** 2025-08-29 17:26:21.290960 | orchestrator | Friday 29 August 2025 17:26:18 +0000 (0:00:00.335) 0:01:10.007 ********* 2025-08-29 17:26:21.290971 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.290982 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.290993 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.291003 | orchestrator | 2025-08-29 17:26:21.291014 | orchestrator | TASK [Get list of Ceph LVs with associated VGs] ******************************** 2025-08-29 17:26:21.291024 | orchestrator | Friday 29 August 2025 17:26:19 +0000 (0:00:00.167) 0:01:10.175 ********* 2025-08-29 17:26:21.291035 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:21.291048 | orchestrator | 2025-08-29 17:26:21.291061 | orchestrator | TASK [Get list of Ceph PVs with associated VGs] ******************************** 2025-08-29 17:26:21.291073 | orchestrator | Friday 29 August 2025 17:26:19 +0000 (0:00:00.545) 0:01:10.720 ********* 2025-08-29 17:26:21.291085 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:21.291098 | orchestrator | 2025-08-29 17:26:21.291110 | orchestrator | TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] *********************** 2025-08-29 17:26:21.291123 | orchestrator | Friday 29 August 2025 17:26:20 +0000 (0:00:00.599) 0:01:11.320 ********* 2025-08-29 17:26:21.291134 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:21.291146 | orchestrator | 2025-08-29 17:26:21.291158 | orchestrator | TASK [Create list of VG/LV names] ********************************************** 2025-08-29 17:26:21.291170 | orchestrator | Friday 29 August 2025 17:26:20 +0000 (0:00:00.154) 0:01:11.474 ********* 2025-08-29 17:26:21.291183 | orchestrator | ok: [testbed-node-5] => (item={'lv_name': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'vg_name': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'}) 2025-08-29 17:26:21.291197 | orchestrator | ok: [testbed-node-5] => (item={'lv_name': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'vg_name': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}) 2025-08-29 17:26:21.291209 | orchestrator | 2025-08-29 17:26:21.291221 | orchestrator | TASK [Fail if block LV defined in lvm_volumes is missing] ********************** 2025-08-29 17:26:21.291240 | orchestrator | Friday 29 August 2025 17:26:20 +0000 (0:00:00.178) 0:01:11.653 ********* 2025-08-29 17:26:21.291270 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.291283 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.291295 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.291340 | orchestrator | 2025-08-29 17:26:21.291361 | orchestrator | TASK [Fail if DB LV defined in lvm_volumes is missing] ************************* 2025-08-29 17:26:21.291382 | orchestrator | Friday 29 August 2025 17:26:20 +0000 (0:00:00.168) 0:01:11.821 ********* 2025-08-29 17:26:21.291402 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.291413 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.291424 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.291435 | orchestrator | 2025-08-29 17:26:21.291446 | orchestrator | TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************ 2025-08-29 17:26:21.291457 | orchestrator | Friday 29 August 2025 17:26:20 +0000 (0:00:00.143) 0:01:11.965 ********* 2025-08-29 17:26:21.291468 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'})  2025-08-29 17:26:21.291479 | orchestrator | skipping: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'})  2025-08-29 17:26:21.291489 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:21.291500 | orchestrator | 2025-08-29 17:26:21.291510 | orchestrator | TASK [Print LVM report data] *************************************************** 2025-08-29 17:26:21.291521 | orchestrator | Friday 29 August 2025 17:26:21 +0000 (0:00:00.176) 0:01:12.141 ********* 2025-08-29 17:26:21.291531 | orchestrator | ok: [testbed-node-5] => { 2025-08-29 17:26:21.291542 | orchestrator |  "lvm_report": { 2025-08-29 17:26:21.291553 | orchestrator |  "lv": [ 2025-08-29 17:26:21.291563 | orchestrator |  { 2025-08-29 17:26:21.291574 | orchestrator |  "lv_name": "osd-block-21b183ef-757d-561c-bead-7bb3aee28288", 2025-08-29 17:26:21.291585 | orchestrator |  "vg_name": "ceph-21b183ef-757d-561c-bead-7bb3aee28288" 2025-08-29 17:26:21.291596 | orchestrator |  }, 2025-08-29 17:26:21.291612 | orchestrator |  { 2025-08-29 17:26:21.291623 | orchestrator |  "lv_name": "osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe", 2025-08-29 17:26:21.291634 | orchestrator |  "vg_name": "ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe" 2025-08-29 17:26:21.291644 | orchestrator |  } 2025-08-29 17:26:21.291655 | orchestrator |  ], 2025-08-29 17:26:21.291665 | orchestrator |  "pv": [ 2025-08-29 17:26:21.291676 | orchestrator |  { 2025-08-29 17:26:21.291686 | orchestrator |  "pv_name": "/dev/sdb", 2025-08-29 17:26:21.291697 | orchestrator |  "vg_name": "ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe" 2025-08-29 17:26:21.291707 | orchestrator |  }, 2025-08-29 17:26:21.291718 | orchestrator |  { 2025-08-29 17:26:21.291728 | orchestrator |  "pv_name": "/dev/sdc", 2025-08-29 17:26:21.291739 | orchestrator |  "vg_name": "ceph-21b183ef-757d-561c-bead-7bb3aee28288" 2025-08-29 17:26:21.291749 | orchestrator |  } 2025-08-29 17:26:21.291760 | orchestrator |  ] 2025-08-29 17:26:21.291770 | orchestrator |  } 2025-08-29 17:26:21.291781 | orchestrator | } 2025-08-29 17:26:21.291792 | orchestrator | 2025-08-29 17:26:21.291802 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:26:21.291813 | orchestrator | testbed-node-3 : ok=51  changed=2  unreachable=0 failed=0 skipped=62  rescued=0 ignored=0 2025-08-29 17:26:21.291831 | orchestrator | testbed-node-4 : ok=51  changed=2  unreachable=0 failed=0 skipped=62  rescued=0 ignored=0 2025-08-29 17:26:21.291842 | orchestrator | testbed-node-5 : ok=51  changed=2  unreachable=0 failed=0 skipped=62  rescued=0 ignored=0 2025-08-29 17:26:21.291853 | orchestrator | 2025-08-29 17:26:21.291863 | orchestrator | 2025-08-29 17:26:21.291874 | orchestrator | 2025-08-29 17:26:21.291884 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:26:21.291895 | orchestrator | Friday 29 August 2025 17:26:21 +0000 (0:00:00.146) 0:01:12.288 ********* 2025-08-29 17:26:21.291905 | orchestrator | =============================================================================== 2025-08-29 17:26:21.291916 | orchestrator | Create block VGs -------------------------------------------------------- 5.72s 2025-08-29 17:26:21.291926 | orchestrator | Create block LVs -------------------------------------------------------- 4.26s 2025-08-29 17:26:21.291936 | orchestrator | Gather DB VGs with total and available size in bytes -------------------- 1.93s 2025-08-29 17:26:21.291947 | orchestrator | Get list of Ceph PVs with associated VGs -------------------------------- 1.66s 2025-08-29 17:26:21.291958 | orchestrator | Get list of Ceph LVs with associated VGs -------------------------------- 1.62s 2025-08-29 17:26:21.291968 | orchestrator | Gather WAL VGs with total and available size in bytes ------------------- 1.61s 2025-08-29 17:26:21.291979 | orchestrator | Gather DB+WAL VGs with total and available size in bytes ---------------- 1.59s 2025-08-29 17:26:21.291990 | orchestrator | Add known partitions to the list of available block devices ------------- 1.41s 2025-08-29 17:26:21.292008 | orchestrator | Add known links to the list of available block devices ------------------ 1.15s 2025-08-29 17:26:21.614809 | orchestrator | Print 'Create DB VGs' --------------------------------------------------- 0.85s 2025-08-29 17:26:21.614908 | orchestrator | Add known partitions to the list of available block devices ------------- 0.84s 2025-08-29 17:26:21.614922 | orchestrator | Add known partitions to the list of available block devices ------------- 0.82s 2025-08-29 17:26:21.614937 | orchestrator | Print 'Create DB LVs for ceph_db_devices' ------------------------------- 0.82s 2025-08-29 17:26:21.614956 | orchestrator | Print LVM report data --------------------------------------------------- 0.82s 2025-08-29 17:26:21.614975 | orchestrator | Fail if DB LV defined in lvm_volumes is missing ------------------------- 0.81s 2025-08-29 17:26:21.614993 | orchestrator | Count OSDs put on ceph_db_devices defined in lvm_volumes ---------------- 0.80s 2025-08-29 17:26:21.615010 | orchestrator | Print number of OSDs wanted per DB VG ----------------------------------- 0.75s 2025-08-29 17:26:21.615038 | orchestrator | Print 'Create WAL LVs for ceph_wal_devices' ----------------------------- 0.75s 2025-08-29 17:26:21.615057 | orchestrator | Print size needed for LVs on ceph_db_devices ---------------------------- 0.75s 2025-08-29 17:26:21.615075 | orchestrator | Add known links to the list of available block devices ------------------ 0.73s 2025-08-29 17:26:33.765484 | orchestrator | 2025-08-29 17:26:33 | INFO  | Task 1fe50d96-1025-4ee4-9d83-0f45e94cffd9 (facts) was prepared for execution. 2025-08-29 17:26:33.765597 | orchestrator | 2025-08-29 17:26:33 | INFO  | It takes a moment until task 1fe50d96-1025-4ee4-9d83-0f45e94cffd9 (facts) has been started and output is visible here. 2025-08-29 17:26:45.988456 | orchestrator | 2025-08-29 17:26:45.988559 | orchestrator | PLAY [Apply role facts] ******************************************************** 2025-08-29 17:26:45.988572 | orchestrator | 2025-08-29 17:26:45.988582 | orchestrator | TASK [osism.commons.facts : Create custom facts directory] ********************* 2025-08-29 17:26:45.988591 | orchestrator | Friday 29 August 2025 17:26:37 +0000 (0:00:00.257) 0:00:00.257 ********* 2025-08-29 17:26:45.988600 | orchestrator | ok: [testbed-manager] 2025-08-29 17:26:45.988610 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:26:45.988619 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:26:45.988650 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:26:45.988659 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:26:45.988668 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:26:45.988676 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:45.988685 | orchestrator | 2025-08-29 17:26:45.988694 | orchestrator | TASK [osism.commons.facts : Copy fact files] *********************************** 2025-08-29 17:26:45.988702 | orchestrator | Friday 29 August 2025 17:26:38 +0000 (0:00:01.050) 0:00:01.307 ********* 2025-08-29 17:26:45.988711 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:26:45.988721 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:26:45.988730 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:26:45.988739 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:26:45.988748 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:26:45.988757 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:26:45.988766 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:45.988775 | orchestrator | 2025-08-29 17:26:45.988784 | orchestrator | PLAY [Gather facts for all hosts] ********************************************** 2025-08-29 17:26:45.988792 | orchestrator | 2025-08-29 17:26:45.988801 | orchestrator | TASK [Gathers facts about hosts] *********************************************** 2025-08-29 17:26:45.988810 | orchestrator | Friday 29 August 2025 17:26:39 +0000 (0:00:01.202) 0:00:02.510 ********* 2025-08-29 17:26:45.988818 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:26:45.988827 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:26:45.988836 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:26:45.988844 | orchestrator | ok: [testbed-manager] 2025-08-29 17:26:45.988853 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:26:45.988862 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:26:45.988870 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:26:45.988879 | orchestrator | 2025-08-29 17:26:45.988888 | orchestrator | PLAY [Gather facts for all hosts if using --limit] ***************************** 2025-08-29 17:26:45.988897 | orchestrator | 2025-08-29 17:26:45.988905 | orchestrator | TASK [Gather facts for all hosts] ********************************************** 2025-08-29 17:26:45.988914 | orchestrator | Friday 29 August 2025 17:26:45 +0000 (0:00:05.014) 0:00:07.525 ********* 2025-08-29 17:26:45.988923 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:26:45.988931 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:26:45.988940 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:26:45.988949 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:26:45.988957 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:26:45.988966 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:26:45.988975 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:26:45.988983 | orchestrator | 2025-08-29 17:26:45.988992 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:26:45.989001 | orchestrator | testbed-manager : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:26:45.989012 | orchestrator | testbed-node-0 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:26:45.989022 | orchestrator | testbed-node-1 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:26:45.989032 | orchestrator | testbed-node-2 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:26:45.989043 | orchestrator | testbed-node-3 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:26:45.989053 | orchestrator | testbed-node-4 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:26:45.989063 | orchestrator | testbed-node-5 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:26:45.989073 | orchestrator | 2025-08-29 17:26:45.989083 | orchestrator | 2025-08-29 17:26:45.989100 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:26:45.989110 | orchestrator | Friday 29 August 2025 17:26:45 +0000 (0:00:00.546) 0:00:08.072 ********* 2025-08-29 17:26:45.989120 | orchestrator | =============================================================================== 2025-08-29 17:26:45.989130 | orchestrator | Gathers facts about hosts ----------------------------------------------- 5.02s 2025-08-29 17:26:45.989140 | orchestrator | osism.commons.facts : Copy fact files ----------------------------------- 1.20s 2025-08-29 17:26:45.989151 | orchestrator | osism.commons.facts : Create custom facts directory --------------------- 1.05s 2025-08-29 17:26:45.989160 | orchestrator | Gather facts for all hosts ---------------------------------------------- 0.55s 2025-08-29 17:26:58.315700 | orchestrator | 2025-08-29 17:26:58 | INFO  | Task 7f75f5df-2a11-408d-9e4a-71b59cc4d68b (frr) was prepared for execution. 2025-08-29 17:26:58.315801 | orchestrator | 2025-08-29 17:26:58 | INFO  | It takes a moment until task 7f75f5df-2a11-408d-9e4a-71b59cc4d68b (frr) has been started and output is visible here. 2025-08-29 17:27:23.979675 | orchestrator | 2025-08-29 17:27:23.979767 | orchestrator | PLAY [Apply role frr] ********************************************************** 2025-08-29 17:27:23.979774 | orchestrator | 2025-08-29 17:27:23.979778 | orchestrator | TASK [osism.services.frr : Include distribution specific install tasks] ******** 2025-08-29 17:27:23.979783 | orchestrator | Friday 29 August 2025 17:27:02 +0000 (0:00:00.232) 0:00:00.232 ********* 2025-08-29 17:27:23.979801 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager 2025-08-29 17:27:23.979807 | orchestrator | 2025-08-29 17:27:23.979811 | orchestrator | TASK [osism.services.frr : Pin frr package version] **************************** 2025-08-29 17:27:23.979815 | orchestrator | Friday 29 August 2025 17:27:02 +0000 (0:00:00.216) 0:00:00.449 ********* 2025-08-29 17:27:23.979819 | orchestrator | changed: [testbed-manager] 2025-08-29 17:27:23.979824 | orchestrator | 2025-08-29 17:27:23.979828 | orchestrator | TASK [osism.services.frr : Install frr package] ******************************** 2025-08-29 17:27:23.979833 | orchestrator | Friday 29 August 2025 17:27:03 +0000 (0:00:01.155) 0:00:01.604 ********* 2025-08-29 17:27:23.979836 | orchestrator | changed: [testbed-manager] 2025-08-29 17:27:23.979840 | orchestrator | 2025-08-29 17:27:23.979844 | orchestrator | TASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] ********************* 2025-08-29 17:27:23.979850 | orchestrator | Friday 29 August 2025 17:27:13 +0000 (0:00:09.767) 0:00:11.371 ********* 2025-08-29 17:27:23.979854 | orchestrator | ok: [testbed-manager] 2025-08-29 17:27:23.979859 | orchestrator | 2025-08-29 17:27:23.979863 | orchestrator | TASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************ 2025-08-29 17:27:23.979866 | orchestrator | Friday 29 August 2025 17:27:14 +0000 (0:00:01.306) 0:00:12.678 ********* 2025-08-29 17:27:23.979870 | orchestrator | changed: [testbed-manager] 2025-08-29 17:27:23.979874 | orchestrator | 2025-08-29 17:27:23.979878 | orchestrator | TASK [osism.services.frr : Set _frr_uplinks fact] ****************************** 2025-08-29 17:27:23.979881 | orchestrator | Friday 29 August 2025 17:27:15 +0000 (0:00:00.967) 0:00:13.645 ********* 2025-08-29 17:27:23.979885 | orchestrator | ok: [testbed-manager] 2025-08-29 17:27:23.979889 | orchestrator | 2025-08-29 17:27:23.979893 | orchestrator | TASK [osism.services.frr : Check for frr.conf file in the configuration repository] *** 2025-08-29 17:27:23.979897 | orchestrator | Friday 29 August 2025 17:27:16 +0000 (0:00:01.166) 0:00:14.812 ********* 2025-08-29 17:27:23.979901 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:27:23.979905 | orchestrator | 2025-08-29 17:27:23.979908 | orchestrator | TASK [osism.services.frr : Copy file from the configuration repository: /etc/frr/frr.conf] *** 2025-08-29 17:27:23.979912 | orchestrator | Friday 29 August 2025 17:27:17 +0000 (0:00:00.817) 0:00:15.629 ********* 2025-08-29 17:27:23.979916 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:27:23.979920 | orchestrator | 2025-08-29 17:27:23.979923 | orchestrator | TASK [osism.services.frr : Copy file from the role: /etc/frr/frr.conf] ********* 2025-08-29 17:27:23.979942 | orchestrator | Friday 29 August 2025 17:27:17 +0000 (0:00:00.145) 0:00:15.775 ********* 2025-08-29 17:27:23.979946 | orchestrator | changed: [testbed-manager] 2025-08-29 17:27:23.979949 | orchestrator | 2025-08-29 17:27:23.979953 | orchestrator | TASK [osism.services.frr : Set sysctl parameters] ****************************** 2025-08-29 17:27:23.979957 | orchestrator | Friday 29 August 2025 17:27:18 +0000 (0:00:01.020) 0:00:16.795 ********* 2025-08-29 17:27:23.979961 | orchestrator | changed: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1}) 2025-08-29 17:27:23.979965 | orchestrator | changed: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0}) 2025-08-29 17:27:23.979970 | orchestrator | changed: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0}) 2025-08-29 17:27:23.979974 | orchestrator | changed: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1}) 2025-08-29 17:27:23.979977 | orchestrator | changed: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1}) 2025-08-29 17:27:23.979981 | orchestrator | changed: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2}) 2025-08-29 17:27:23.979985 | orchestrator | 2025-08-29 17:27:23.979989 | orchestrator | TASK [osism.services.frr : Manage frr service] ********************************* 2025-08-29 17:27:23.979992 | orchestrator | Friday 29 August 2025 17:27:20 +0000 (0:00:02.195) 0:00:18.990 ********* 2025-08-29 17:27:23.979996 | orchestrator | ok: [testbed-manager] 2025-08-29 17:27:23.980000 | orchestrator | 2025-08-29 17:27:23.980003 | orchestrator | RUNNING HANDLER [osism.services.frr : Restart frr service] ********************* 2025-08-29 17:27:23.980007 | orchestrator | Friday 29 August 2025 17:27:22 +0000 (0:00:01.415) 0:00:20.406 ********* 2025-08-29 17:27:23.980011 | orchestrator | changed: [testbed-manager] 2025-08-29 17:27:23.980015 | orchestrator | 2025-08-29 17:27:23.980018 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:27:23.980023 | orchestrator | testbed-manager : ok=11  changed=6  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:27:23.980026 | orchestrator | 2025-08-29 17:27:23.980030 | orchestrator | 2025-08-29 17:27:23.980034 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:27:23.980038 | orchestrator | Friday 29 August 2025 17:27:23 +0000 (0:00:01.351) 0:00:21.757 ********* 2025-08-29 17:27:23.980041 | orchestrator | =============================================================================== 2025-08-29 17:27:23.980045 | orchestrator | osism.services.frr : Install frr package -------------------------------- 9.77s 2025-08-29 17:27:23.980049 | orchestrator | osism.services.frr : Set sysctl parameters ------------------------------ 2.20s 2025-08-29 17:27:23.980052 | orchestrator | osism.services.frr : Manage frr service --------------------------------- 1.42s 2025-08-29 17:27:23.980056 | orchestrator | osism.services.frr : Restart frr service -------------------------------- 1.35s 2025-08-29 17:27:23.980071 | orchestrator | osism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.31s 2025-08-29 17:27:23.980075 | orchestrator | osism.services.frr : Set _frr_uplinks fact ------------------------------ 1.17s 2025-08-29 17:27:23.980078 | orchestrator | osism.services.frr : Pin frr package version ---------------------------- 1.16s 2025-08-29 17:27:23.980082 | orchestrator | osism.services.frr : Copy file from the role: /etc/frr/frr.conf --------- 1.02s 2025-08-29 17:27:23.980086 | orchestrator | osism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.97s 2025-08-29 17:27:23.980089 | orchestrator | osism.services.frr : Check for frr.conf file in the configuration repository --- 0.82s 2025-08-29 17:27:23.980093 | orchestrator | osism.services.frr : Include distribution specific install tasks -------- 0.22s 2025-08-29 17:27:23.980097 | orchestrator | osism.services.frr : Copy file from the configuration repository: /etc/frr/frr.conf --- 0.15s 2025-08-29 17:27:24.246492 | orchestrator | 2025-08-29 17:27:24.249443 | orchestrator | --> DEPLOY IN A NUTSHELL -- START -- Fri Aug 29 17:27:24 UTC 2025 2025-08-29 17:27:24.249459 | orchestrator | 2025-08-29 17:27:26.001145 | orchestrator | 2025-08-29 17:27:25 | INFO  | Collection nutshell is prepared for execution 2025-08-29 17:27:26.001199 | orchestrator | 2025-08-29 17:27:25 | INFO  | D [0] - dotfiles 2025-08-29 17:27:36.189693 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [0] - homer 2025-08-29 17:27:36.189798 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [0] - netdata 2025-08-29 17:27:36.189812 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [0] - openstackclient 2025-08-29 17:27:36.189823 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [0] - phpmyadmin 2025-08-29 17:27:36.189834 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [0] - common 2025-08-29 17:27:36.191866 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [1] -- loadbalancer 2025-08-29 17:27:36.191906 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [2] --- opensearch 2025-08-29 17:27:36.192099 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [2] --- mariadb-ng 2025-08-29 17:27:36.192400 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [3] ---- horizon 2025-08-29 17:27:36.192432 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [3] ---- keystone 2025-08-29 17:27:36.192935 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [4] ----- neutron 2025-08-29 17:27:36.192957 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [5] ------ wait-for-nova 2025-08-29 17:27:36.192969 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [5] ------ octavia 2025-08-29 17:27:36.194195 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [4] ----- barbican 2025-08-29 17:27:36.194225 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [4] ----- designate 2025-08-29 17:27:36.194237 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [4] ----- ironic 2025-08-29 17:27:36.194666 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [4] ----- placement 2025-08-29 17:27:36.194689 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [4] ----- magnum 2025-08-29 17:27:36.195445 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [1] -- openvswitch 2025-08-29 17:27:36.195475 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [2] --- ovn 2025-08-29 17:27:36.195782 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [1] -- memcached 2025-08-29 17:27:36.195978 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [1] -- redis 2025-08-29 17:27:36.195997 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [1] -- rabbitmq-ng 2025-08-29 17:27:36.196217 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [0] - kubernetes 2025-08-29 17:27:36.198516 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [1] -- kubeconfig 2025-08-29 17:27:36.198655 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [1] -- copy-kubeconfig 2025-08-29 17:27:36.198928 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [0] - ceph 2025-08-29 17:27:36.201049 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [1] -- ceph-pools 2025-08-29 17:27:36.201139 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [2] --- copy-ceph-keys 2025-08-29 17:27:36.201155 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [3] ---- cephclient 2025-08-29 17:27:36.201247 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [4] ----- ceph-bootstrap-dashboard 2025-08-29 17:27:36.201529 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [4] ----- wait-for-keystone 2025-08-29 17:27:36.201550 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [5] ------ kolla-ceph-rgw 2025-08-29 17:27:36.201933 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [5] ------ glance 2025-08-29 17:27:36.201962 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [5] ------ cinder 2025-08-29 17:27:36.201974 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [5] ------ nova 2025-08-29 17:27:36.201985 | orchestrator | 2025-08-29 17:27:36 | INFO  | A [4] ----- prometheus 2025-08-29 17:27:36.202149 | orchestrator | 2025-08-29 17:27:36 | INFO  | D [5] ------ grafana 2025-08-29 17:27:36.436729 | orchestrator | 2025-08-29 17:27:36 | INFO  | All tasks of the collection nutshell are prepared for execution 2025-08-29 17:27:36.436829 | orchestrator | 2025-08-29 17:27:36 | INFO  | Tasks are running in the background 2025-08-29 17:27:38.998289 | orchestrator | 2025-08-29 17:27:38 | INFO  | No task IDs specified, wait for all currently running tasks 2025-08-29 17:27:41.096734 | orchestrator | 2025-08-29 17:27:41 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:27:41.096834 | orchestrator | 2025-08-29 17:27:41 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:27:41.097269 | orchestrator | 2025-08-29 17:27:41 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state STARTED 2025-08-29 17:27:41.099698 | orchestrator | 2025-08-29 17:27:41 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:27:41.101899 | orchestrator | 2025-08-29 17:27:41 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:27:41.102278 | orchestrator | 2025-08-29 17:27:41 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:27:41.102838 | orchestrator | 2025-08-29 17:27:41 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:27:41.102861 | orchestrator | 2025-08-29 17:27:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:27:44.151771 | orchestrator | 2025-08-29 17:27:44 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:27:44.151952 | orchestrator | 2025-08-29 17:27:44 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:27:44.151972 | orchestrator | 2025-08-29 17:27:44 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state STARTED 2025-08-29 17:27:44.151985 | orchestrator | 2025-08-29 17:27:44 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:27:44.152007 | orchestrator | 2025-08-29 17:27:44 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:27:44.152478 | orchestrator | 2025-08-29 17:27:44 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:27:44.152931 | orchestrator | 2025-08-29 17:27:44 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:27:44.154260 | orchestrator | 2025-08-29 17:27:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:27:47.197855 | orchestrator | 2025-08-29 17:27:47 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:27:47.197929 | orchestrator | 2025-08-29 17:27:47 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:27:47.197944 | orchestrator | 2025-08-29 17:27:47 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state STARTED 2025-08-29 17:27:47.197956 | orchestrator | 2025-08-29 17:27:47 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:27:47.197967 | orchestrator | 2025-08-29 17:27:47 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:27:47.259526 | orchestrator | 2025-08-29 17:27:47 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:27:47.259587 | orchestrator | 2025-08-29 17:27:47 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:27:47.259599 | orchestrator | 2025-08-29 17:27:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:27:50.332527 | orchestrator | 2025-08-29 17:27:50 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:27:50.332866 | orchestrator | 2025-08-29 17:27:50 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:27:50.333404 | orchestrator | 2025-08-29 17:27:50 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state STARTED 2025-08-29 17:27:50.334251 | orchestrator | 2025-08-29 17:27:50 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:27:50.334750 | orchestrator | 2025-08-29 17:27:50 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:27:50.335947 | orchestrator | 2025-08-29 17:27:50 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:27:50.338375 | orchestrator | 2025-08-29 17:27:50 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:27:50.338429 | orchestrator | 2025-08-29 17:27:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:27:53.473484 | orchestrator | 2025-08-29 17:27:53 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:27:53.473859 | orchestrator | 2025-08-29 17:27:53 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:27:53.475937 | orchestrator | 2025-08-29 17:27:53 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state STARTED 2025-08-29 17:27:53.476247 | orchestrator | 2025-08-29 17:27:53 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:27:53.476852 | orchestrator | 2025-08-29 17:27:53 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:27:53.477718 | orchestrator | 2025-08-29 17:27:53 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:27:53.478564 | orchestrator | 2025-08-29 17:27:53 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:27:53.478599 | orchestrator | 2025-08-29 17:27:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:27:56.568637 | orchestrator | 2025-08-29 17:27:56 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:27:56.571943 | orchestrator | 2025-08-29 17:27:56 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:27:56.575567 | orchestrator | 2025-08-29 17:27:56 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state STARTED 2025-08-29 17:27:56.579540 | orchestrator | 2025-08-29 17:27:56 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:27:56.581406 | orchestrator | 2025-08-29 17:27:56 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:27:56.582669 | orchestrator | 2025-08-29 17:27:56 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:27:56.584025 | orchestrator | 2025-08-29 17:27:56 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:27:56.584246 | orchestrator | 2025-08-29 17:27:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:27:59.642509 | orchestrator | 2025-08-29 17:27:59 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:27:59.644621 | orchestrator | 2025-08-29 17:27:59 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:27:59.647738 | orchestrator | 2025-08-29 17:27:59 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state STARTED 2025-08-29 17:27:59.650554 | orchestrator | 2025-08-29 17:27:59 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:27:59.653542 | orchestrator | 2025-08-29 17:27:59 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:27:59.659788 | orchestrator | 2025-08-29 17:27:59 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:27:59.659843 | orchestrator | 2025-08-29 17:27:59 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:27:59.659864 | orchestrator | 2025-08-29 17:27:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:02.704155 | orchestrator | 2025-08-29 17:28:02 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:02.713832 | orchestrator | 2025-08-29 17:28:02.713909 | orchestrator | PLAY [Apply role geerlingguy.dotfiles] ***************************************** 2025-08-29 17:28:02.713930 | orchestrator | 2025-08-29 17:28:02.713945 | orchestrator | TASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] **** 2025-08-29 17:28:02.713960 | orchestrator | Friday 29 August 2025 17:27:48 +0000 (0:00:01.107) 0:00:01.107 ********* 2025-08-29 17:28:02.713975 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:28:02.713999 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:28:02.714069 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:28:02.714089 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:02.714098 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:28:02.714107 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:28:02.714115 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:28:02.714124 | orchestrator | 2025-08-29 17:28:02.714133 | orchestrator | TASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ******** 2025-08-29 17:28:02.714141 | orchestrator | Friday 29 August 2025 17:27:52 +0000 (0:00:04.034) 0:00:05.142 ********* 2025-08-29 17:28:02.714150 | orchestrator | ok: [testbed-manager] => (item=.tmux.conf) 2025-08-29 17:28:02.714159 | orchestrator | ok: [testbed-node-0] => (item=.tmux.conf) 2025-08-29 17:28:02.714168 | orchestrator | ok: [testbed-node-1] => (item=.tmux.conf) 2025-08-29 17:28:02.714176 | orchestrator | ok: [testbed-node-2] => (item=.tmux.conf) 2025-08-29 17:28:02.714185 | orchestrator | ok: [testbed-node-3] => (item=.tmux.conf) 2025-08-29 17:28:02.714194 | orchestrator | ok: [testbed-node-4] => (item=.tmux.conf) 2025-08-29 17:28:02.714202 | orchestrator | ok: [testbed-node-5] => (item=.tmux.conf) 2025-08-29 17:28:02.714211 | orchestrator | 2025-08-29 17:28:02.714220 | orchestrator | TASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] *** 2025-08-29 17:28:02.714229 | orchestrator | Friday 29 August 2025 17:27:55 +0000 (0:00:02.592) 0:00:07.734 ********* 2025-08-29 17:28:02.714241 | orchestrator | ok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': "ls: cannot access '/home/dragon/.tmux.conf': No such file or directory", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2025-08-29 17:27:53.204566', 'end': '2025-08-29 17:27:53.211241', 'delta': '0:00:00.006675', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': ["ls: cannot access '/home/dragon/.tmux.conf': No such file or directory"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}]) 2025-08-29 17:28:02.714262 | orchestrator | ok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': "ls: cannot access '/home/dragon/.tmux.conf': No such file or directory", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2025-08-29 17:27:53.676926', 'end': '2025-08-29 17:27:53.685967', 'delta': '0:00:00.009041', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': ["ls: cannot access '/home/dragon/.tmux.conf': No such file or directory"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}]) 2025-08-29 17:28:02.714289 | orchestrator | ok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': "ls: cannot access '/home/dragon/.tmux.conf': No such file or directory", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2025-08-29 17:27:53.883442', 'end': '2025-08-29 17:27:53.892619', 'delta': '0:00:00.009177', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': ["ls: cannot access '/home/dragon/.tmux.conf': No such file or directory"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}]) 2025-08-29 17:28:02.714341 | orchestrator | ok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': "ls: cannot access '/home/dragon/.tmux.conf': No such file or directory", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2025-08-29 17:27:53.512235', 'end': '2025-08-29 17:27:53.518921', 'delta': '0:00:00.006686', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': ["ls: cannot access '/home/dragon/.tmux.conf': No such file or directory"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}]) 2025-08-29 17:28:02.714354 | orchestrator | ok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': "ls: cannot access '/home/dragon/.tmux.conf': No such file or directory", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2025-08-29 17:27:54.377544', 'end': '2025-08-29 17:27:54.383903', 'delta': '0:00:00.006359', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': ["ls: cannot access '/home/dragon/.tmux.conf': No such file or directory"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}]) 2025-08-29 17:28:02.714574 | orchestrator | ok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': "ls: cannot access '/home/dragon/.tmux.conf': No such file or directory", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2025-08-29 17:27:52.913194', 'end': '2025-08-29 17:27:52.916905', 'delta': '0:00:00.003711', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': ["ls: cannot access '/home/dragon/.tmux.conf': No such file or directory"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}]) 2025-08-29 17:28:02.714587 | orchestrator | ok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': "ls: cannot access '/home/dragon/.tmux.conf': No such file or directory", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2025-08-29 17:27:54.758962', 'end': '2025-08-29 17:27:54.768819', 'delta': '0:00:00.009857', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': ["ls: cannot access '/home/dragon/.tmux.conf': No such file or directory"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}]) 2025-08-29 17:28:02.714608 | orchestrator | 2025-08-29 17:28:02.714618 | orchestrator | TASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] **** 2025-08-29 17:28:02.714627 | orchestrator | Friday 29 August 2025 17:27:57 +0000 (0:00:01.983) 0:00:09.717 ********* 2025-08-29 17:28:02.714635 | orchestrator | ok: [testbed-manager] => (item=.tmux.conf) 2025-08-29 17:28:02.714644 | orchestrator | ok: [testbed-node-0] => (item=.tmux.conf) 2025-08-29 17:28:02.714652 | orchestrator | ok: [testbed-node-1] => (item=.tmux.conf) 2025-08-29 17:28:02.714661 | orchestrator | ok: [testbed-node-2] => (item=.tmux.conf) 2025-08-29 17:28:02.714669 | orchestrator | ok: [testbed-node-3] => (item=.tmux.conf) 2025-08-29 17:28:02.714678 | orchestrator | ok: [testbed-node-4] => (item=.tmux.conf) 2025-08-29 17:28:02.714686 | orchestrator | ok: [testbed-node-5] => (item=.tmux.conf) 2025-08-29 17:28:02.714694 | orchestrator | 2025-08-29 17:28:02.714703 | orchestrator | TASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ****************** 2025-08-29 17:28:02.714715 | orchestrator | Friday 29 August 2025 17:27:59 +0000 (0:00:02.387) 0:00:12.105 ********* 2025-08-29 17:28:02.714724 | orchestrator | changed: [testbed-node-0] => (item=.tmux.conf) 2025-08-29 17:28:02.714733 | orchestrator | changed: [testbed-node-1] => (item=.tmux.conf) 2025-08-29 17:28:02.714741 | orchestrator | changed: [testbed-node-3] => (item=.tmux.conf) 2025-08-29 17:28:02.714750 | orchestrator | changed: [testbed-node-2] => (item=.tmux.conf) 2025-08-29 17:28:02.714758 | orchestrator | changed: [testbed-manager] => (item=.tmux.conf) 2025-08-29 17:28:02.714767 | orchestrator | changed: [testbed-node-4] => (item=.tmux.conf) 2025-08-29 17:28:02.714775 | orchestrator | changed: [testbed-node-5] => (item=.tmux.conf) 2025-08-29 17:28:02.714784 | orchestrator | 2025-08-29 17:28:02.714792 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:28:02.714808 | orchestrator | testbed-manager : ok=5  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:02.714818 | orchestrator | testbed-node-0 : ok=5  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:02.714827 | orchestrator | testbed-node-1 : ok=5  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:02.714836 | orchestrator | testbed-node-2 : ok=5  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:02.714844 | orchestrator | testbed-node-3 : ok=5  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:02.714853 | orchestrator | testbed-node-4 : ok=5  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:02.714861 | orchestrator | testbed-node-5 : ok=5  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:02.714870 | orchestrator | 2025-08-29 17:28:02.714879 | orchestrator | 2025-08-29 17:28:02.714887 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:28:02.714896 | orchestrator | Friday 29 August 2025 17:28:01 +0000 (0:00:02.398) 0:00:14.503 ********* 2025-08-29 17:28:02.714905 | orchestrator | =============================================================================== 2025-08-29 17:28:02.714913 | orchestrator | geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 4.03s 2025-08-29 17:28:02.714922 | orchestrator | geerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 2.59s 2025-08-29 17:28:02.714936 | orchestrator | geerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 2.40s 2025-08-29 17:28:02.714945 | orchestrator | geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 2.39s 2025-08-29 17:28:02.714953 | orchestrator | geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 1.98s 2025-08-29 17:28:02.714962 | orchestrator | 2025-08-29 17:28:02 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:02.714971 | orchestrator | 2025-08-29 17:28:02 | INFO  | Task d66cb012-a6c0-4403-968f-ab1a4dc04783 is in state SUCCESS 2025-08-29 17:28:02.714980 | orchestrator | 2025-08-29 17:28:02 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:02.716836 | orchestrator | 2025-08-29 17:28:02 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:02.718004 | orchestrator | 2025-08-29 17:28:02 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:28:02.719539 | orchestrator | 2025-08-29 17:28:02 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:02.722302 | orchestrator | 2025-08-29 17:28:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:05.773568 | orchestrator | 2025-08-29 17:28:05 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:05.773651 | orchestrator | 2025-08-29 17:28:05 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:05.775618 | orchestrator | 2025-08-29 17:28:05 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:05.775648 | orchestrator | 2025-08-29 17:28:05 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:05.776400 | orchestrator | 2025-08-29 17:28:05 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:28:05.779364 | orchestrator | 2025-08-29 17:28:05 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:05.779392 | orchestrator | 2025-08-29 17:28:05 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:05.779404 | orchestrator | 2025-08-29 17:28:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:08.903669 | orchestrator | 2025-08-29 17:28:08 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:08.905972 | orchestrator | 2025-08-29 17:28:08 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:08.910231 | orchestrator | 2025-08-29 17:28:08 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:08.911559 | orchestrator | 2025-08-29 17:28:08 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:08.925449 | orchestrator | 2025-08-29 17:28:08 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:28:08.927541 | orchestrator | 2025-08-29 17:28:08 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:08.931393 | orchestrator | 2025-08-29 17:28:08 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:08.931422 | orchestrator | 2025-08-29 17:28:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:12.127189 | orchestrator | 2025-08-29 17:28:11 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:12.127270 | orchestrator | 2025-08-29 17:28:11 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:12.127284 | orchestrator | 2025-08-29 17:28:11 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:12.127372 | orchestrator | 2025-08-29 17:28:11 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:12.127386 | orchestrator | 2025-08-29 17:28:11 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:28:12.127398 | orchestrator | 2025-08-29 17:28:11 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:12.127409 | orchestrator | 2025-08-29 17:28:11 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:12.127419 | orchestrator | 2025-08-29 17:28:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:15.047681 | orchestrator | 2025-08-29 17:28:15 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:15.047777 | orchestrator | 2025-08-29 17:28:15 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:15.047795 | orchestrator | 2025-08-29 17:28:15 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:15.050681 | orchestrator | 2025-08-29 17:28:15 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:15.052513 | orchestrator | 2025-08-29 17:28:15 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:28:15.053745 | orchestrator | 2025-08-29 17:28:15 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:15.056636 | orchestrator | 2025-08-29 17:28:15 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:15.056678 | orchestrator | 2025-08-29 17:28:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:18.125354 | orchestrator | 2025-08-29 17:28:18 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:18.125698 | orchestrator | 2025-08-29 17:28:18 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:18.125725 | orchestrator | 2025-08-29 17:28:18 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:18.126243 | orchestrator | 2025-08-29 17:28:18 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:18.126563 | orchestrator | 2025-08-29 17:28:18 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:28:18.127007 | orchestrator | 2025-08-29 17:28:18 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:18.127501 | orchestrator | 2025-08-29 17:28:18 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:18.127528 | orchestrator | 2025-08-29 17:28:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:21.165482 | orchestrator | 2025-08-29 17:28:21 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:21.165569 | orchestrator | 2025-08-29 17:28:21 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:21.165583 | orchestrator | 2025-08-29 17:28:21 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:21.169669 | orchestrator | 2025-08-29 17:28:21 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:21.172639 | orchestrator | 2025-08-29 17:28:21 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state STARTED 2025-08-29 17:28:21.175293 | orchestrator | 2025-08-29 17:28:21 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:21.178742 | orchestrator | 2025-08-29 17:28:21 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:21.178769 | orchestrator | 2025-08-29 17:28:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:24.268031 | orchestrator | 2025-08-29 17:28:24 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:24.268117 | orchestrator | 2025-08-29 17:28:24 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:24.269788 | orchestrator | 2025-08-29 17:28:24 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:24.271048 | orchestrator | 2025-08-29 17:28:24 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:24.271697 | orchestrator | 2025-08-29 17:28:24 | INFO  | Task 77c33d3f-1394-4c3d-b5f8-969e72767e32 is in state SUCCESS 2025-08-29 17:28:24.342189 | orchestrator | 2025-08-29 17:28:24 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:24.342263 | orchestrator | 2025-08-29 17:28:24 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:24.342276 | orchestrator | 2025-08-29 17:28:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:27.372091 | orchestrator | 2025-08-29 17:28:27 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:27.372177 | orchestrator | 2025-08-29 17:28:27 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:27.372191 | orchestrator | 2025-08-29 17:28:27 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:27.374253 | orchestrator | 2025-08-29 17:28:27 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:27.375626 | orchestrator | 2025-08-29 17:28:27 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:27.377193 | orchestrator | 2025-08-29 17:28:27 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:27.377799 | orchestrator | 2025-08-29 17:28:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:30.515618 | orchestrator | 2025-08-29 17:28:30 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:30.614506 | orchestrator | 2025-08-29 17:28:30 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:30.614581 | orchestrator | 2025-08-29 17:28:30 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:30.614594 | orchestrator | 2025-08-29 17:28:30 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state STARTED 2025-08-29 17:28:30.614605 | orchestrator | 2025-08-29 17:28:30 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:30.614616 | orchestrator | 2025-08-29 17:28:30 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:30.614628 | orchestrator | 2025-08-29 17:28:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:33.565820 | orchestrator | 2025-08-29 17:28:33 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:33.566002 | orchestrator | 2025-08-29 17:28:33 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:33.566655 | orchestrator | 2025-08-29 17:28:33 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:33.566829 | orchestrator | 2025-08-29 17:28:33 | INFO  | Task 7ce891da-3fd3-4efd-ac67-fc559706d148 is in state SUCCESS 2025-08-29 17:28:33.567540 | orchestrator | 2025-08-29 17:28:33 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:33.568067 | orchestrator | 2025-08-29 17:28:33 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:33.568494 | orchestrator | 2025-08-29 17:28:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:36.598992 | orchestrator | 2025-08-29 17:28:36 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:36.599066 | orchestrator | 2025-08-29 17:28:36 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:36.599421 | orchestrator | 2025-08-29 17:28:36 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:36.600020 | orchestrator | 2025-08-29 17:28:36 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:36.600628 | orchestrator | 2025-08-29 17:28:36 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:36.600748 | orchestrator | 2025-08-29 17:28:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:39.655972 | orchestrator | 2025-08-29 17:28:39 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:39.656779 | orchestrator | 2025-08-29 17:28:39 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:39.658975 | orchestrator | 2025-08-29 17:28:39 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:39.660925 | orchestrator | 2025-08-29 17:28:39 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:39.664049 | orchestrator | 2025-08-29 17:28:39 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:39.664421 | orchestrator | 2025-08-29 17:28:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:42.699935 | orchestrator | 2025-08-29 17:28:42 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:42.704795 | orchestrator | 2025-08-29 17:28:42 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:42.704869 | orchestrator | 2025-08-29 17:28:42 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:42.704884 | orchestrator | 2025-08-29 17:28:42 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:42.704897 | orchestrator | 2025-08-29 17:28:42 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:42.705307 | orchestrator | 2025-08-29 17:28:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:45.744983 | orchestrator | 2025-08-29 17:28:45 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:45.745460 | orchestrator | 2025-08-29 17:28:45 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:45.746766 | orchestrator | 2025-08-29 17:28:45 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:45.748498 | orchestrator | 2025-08-29 17:28:45 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state STARTED 2025-08-29 17:28:45.748520 | orchestrator | 2025-08-29 17:28:45 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:45.748528 | orchestrator | 2025-08-29 17:28:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:48.780713 | orchestrator | 2025-08-29 17:28:48 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:48.781762 | orchestrator | 2025-08-29 17:28:48 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:48.783906 | orchestrator | 2025-08-29 17:28:48 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:48.784808 | orchestrator | 2025-08-29 17:28:48 | INFO  | Task 628ae254-8421-4c8e-b5c3-7e47d3a76466 is in state SUCCESS 2025-08-29 17:28:48.786875 | orchestrator | 2025-08-29 17:28:48.786916 | orchestrator | 2025-08-29 17:28:48.786928 | orchestrator | PLAY [Apply role homer] ******************************************************** 2025-08-29 17:28:48.786940 | orchestrator | 2025-08-29 17:28:48.786951 | orchestrator | TASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] *** 2025-08-29 17:28:48.786963 | orchestrator | Friday 29 August 2025 17:27:47 +0000 (0:00:00.754) 0:00:00.754 ********* 2025-08-29 17:28:48.786974 | orchestrator | ok: [testbed-manager] => { 2025-08-29 17:28:48.786987 | orchestrator |  "msg": "The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter." 2025-08-29 17:28:48.787000 | orchestrator | } 2025-08-29 17:28:48.787011 | orchestrator | 2025-08-29 17:28:48.787022 | orchestrator | TASK [osism.services.homer : Create traefik external network] ****************** 2025-08-29 17:28:48.787034 | orchestrator | Friday 29 August 2025 17:27:47 +0000 (0:00:00.466) 0:00:01.221 ********* 2025-08-29 17:28:48.787045 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.787056 | orchestrator | 2025-08-29 17:28:48.787067 | orchestrator | TASK [osism.services.homer : Create required directories] ********************** 2025-08-29 17:28:48.787078 | orchestrator | Friday 29 August 2025 17:27:49 +0000 (0:00:01.972) 0:00:03.193 ********* 2025-08-29 17:28:48.787088 | orchestrator | changed: [testbed-manager] => (item=/opt/homer/configuration) 2025-08-29 17:28:48.787099 | orchestrator | ok: [testbed-manager] => (item=/opt/homer) 2025-08-29 17:28:48.787110 | orchestrator | 2025-08-29 17:28:48.787121 | orchestrator | TASK [osism.services.homer : Copy config.yml configuration file] *************** 2025-08-29 17:28:48.787138 | orchestrator | Friday 29 August 2025 17:27:51 +0000 (0:00:01.091) 0:00:04.285 ********* 2025-08-29 17:28:48.787149 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.787160 | orchestrator | 2025-08-29 17:28:48.787171 | orchestrator | TASK [osism.services.homer : Copy docker-compose.yml file] ********************* 2025-08-29 17:28:48.787181 | orchestrator | Friday 29 August 2025 17:27:53 +0000 (0:00:02.533) 0:00:06.819 ********* 2025-08-29 17:28:48.787192 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.787202 | orchestrator | 2025-08-29 17:28:48.787213 | orchestrator | TASK [osism.services.homer : Manage homer service] ***************************** 2025-08-29 17:28:48.787224 | orchestrator | Friday 29 August 2025 17:27:55 +0000 (0:00:01.559) 0:00:08.379 ********* 2025-08-29 17:28:48.787234 | orchestrator | FAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left). 2025-08-29 17:28:48.787245 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.787255 | orchestrator | 2025-08-29 17:28:48.787266 | orchestrator | RUNNING HANDLER [osism.services.homer : Restart homer service] ***************** 2025-08-29 17:28:48.787277 | orchestrator | Friday 29 August 2025 17:28:19 +0000 (0:00:24.081) 0:00:32.461 ********* 2025-08-29 17:28:48.787287 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.787298 | orchestrator | 2025-08-29 17:28:48.787309 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:28:48.787320 | orchestrator | testbed-manager : ok=7  changed=4  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.787355 | orchestrator | 2025-08-29 17:28:48.787366 | orchestrator | 2025-08-29 17:28:48.787384 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:28:48.787401 | orchestrator | Friday 29 August 2025 17:28:21 +0000 (0:00:02.700) 0:00:35.161 ********* 2025-08-29 17:28:48.787412 | orchestrator | =============================================================================== 2025-08-29 17:28:48.787427 | orchestrator | osism.services.homer : Manage homer service ---------------------------- 24.08s 2025-08-29 17:28:48.787449 | orchestrator | osism.services.homer : Restart homer service ---------------------------- 2.70s 2025-08-29 17:28:48.787481 | orchestrator | osism.services.homer : Copy config.yml configuration file --------------- 2.53s 2025-08-29 17:28:48.787494 | orchestrator | osism.services.homer : Create traefik external network ------------------ 1.97s 2025-08-29 17:28:48.787506 | orchestrator | osism.services.homer : Copy docker-compose.yml file --------------------- 1.56s 2025-08-29 17:28:48.787518 | orchestrator | osism.services.homer : Create required directories ---------------------- 1.09s 2025-08-29 17:28:48.787531 | orchestrator | osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.47s 2025-08-29 17:28:48.787543 | orchestrator | 2025-08-29 17:28:48.787555 | orchestrator | 2025-08-29 17:28:48.787594 | orchestrator | PLAY [Apply role openstackclient] ********************************************** 2025-08-29 17:28:48.787607 | orchestrator | 2025-08-29 17:28:48.787619 | orchestrator | TASK [osism.services.openstackclient : Include tasks] ************************** 2025-08-29 17:28:48.787632 | orchestrator | Friday 29 August 2025 17:27:48 +0000 (0:00:00.584) 0:00:00.584 ********* 2025-08-29 17:28:48.787645 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager 2025-08-29 17:28:48.787659 | orchestrator | 2025-08-29 17:28:48.787671 | orchestrator | TASK [osism.services.openstackclient : Create required directories] ************ 2025-08-29 17:28:48.787683 | orchestrator | Friday 29 August 2025 17:27:48 +0000 (0:00:00.750) 0:00:01.335 ********* 2025-08-29 17:28:48.787696 | orchestrator | changed: [testbed-manager] => (item=/opt/configuration/environments/openstack) 2025-08-29 17:28:48.787708 | orchestrator | changed: [testbed-manager] => (item=/opt/openstackclient/data) 2025-08-29 17:28:48.787721 | orchestrator | ok: [testbed-manager] => (item=/opt/openstackclient) 2025-08-29 17:28:48.787733 | orchestrator | 2025-08-29 17:28:48.787746 | orchestrator | TASK [osism.services.openstackclient : Copy docker-compose.yml file] *********** 2025-08-29 17:28:48.787758 | orchestrator | Friday 29 August 2025 17:27:50 +0000 (0:00:01.526) 0:00:02.861 ********* 2025-08-29 17:28:48.787770 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.787783 | orchestrator | 2025-08-29 17:28:48.787796 | orchestrator | TASK [osism.services.openstackclient : Manage openstackclient service] ********* 2025-08-29 17:28:48.787816 | orchestrator | Friday 29 August 2025 17:27:52 +0000 (0:00:01.570) 0:00:04.431 ********* 2025-08-29 17:28:48.787845 | orchestrator | FAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left). 2025-08-29 17:28:48.787857 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.787868 | orchestrator | 2025-08-29 17:28:48.787885 | orchestrator | TASK [osism.services.openstackclient : Copy openstack wrapper script] ********** 2025-08-29 17:28:48.787903 | orchestrator | Friday 29 August 2025 17:28:24 +0000 (0:00:32.587) 0:00:37.019 ********* 2025-08-29 17:28:48.787921 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.787936 | orchestrator | 2025-08-29 17:28:48.787947 | orchestrator | TASK [osism.services.openstackclient : Remove ospurge wrapper script] ********** 2025-08-29 17:28:48.787958 | orchestrator | Friday 29 August 2025 17:28:25 +0000 (0:00:00.932) 0:00:37.952 ********* 2025-08-29 17:28:48.787968 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.787979 | orchestrator | 2025-08-29 17:28:48.787990 | orchestrator | RUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] *** 2025-08-29 17:28:48.788000 | orchestrator | Friday 29 August 2025 17:28:26 +0000 (0:00:00.574) 0:00:38.526 ********* 2025-08-29 17:28:48.788011 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.788021 | orchestrator | 2025-08-29 17:28:48.788032 | orchestrator | RUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] *** 2025-08-29 17:28:48.788042 | orchestrator | Friday 29 August 2025 17:28:28 +0000 (0:00:02.719) 0:00:41.245 ********* 2025-08-29 17:28:48.788053 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.788063 | orchestrator | 2025-08-29 17:28:48.788074 | orchestrator | RUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] *** 2025-08-29 17:28:48.788085 | orchestrator | Friday 29 August 2025 17:28:30 +0000 (0:00:01.435) 0:00:42.681 ********* 2025-08-29 17:28:48.788096 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.788113 | orchestrator | 2025-08-29 17:28:48.788124 | orchestrator | RUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] *** 2025-08-29 17:28:48.788134 | orchestrator | Friday 29 August 2025 17:28:31 +0000 (0:00:00.843) 0:00:43.524 ********* 2025-08-29 17:28:48.788145 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.788156 | orchestrator | 2025-08-29 17:28:48.788166 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:28:48.788177 | orchestrator | testbed-manager : ok=10  changed=6  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.788188 | orchestrator | 2025-08-29 17:28:48.788199 | orchestrator | 2025-08-29 17:28:48.788209 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:28:48.788220 | orchestrator | Friday 29 August 2025 17:28:31 +0000 (0:00:00.559) 0:00:44.083 ********* 2025-08-29 17:28:48.788260 | orchestrator | =============================================================================== 2025-08-29 17:28:48.788272 | orchestrator | osism.services.openstackclient : Manage openstackclient service -------- 32.59s 2025-08-29 17:28:48.788283 | orchestrator | osism.services.openstackclient : Restart openstackclient service -------- 2.72s 2025-08-29 17:28:48.788294 | orchestrator | osism.services.openstackclient : Copy docker-compose.yml file ----------- 1.57s 2025-08-29 17:28:48.788304 | orchestrator | osism.services.openstackclient : Create required directories ------------ 1.53s 2025-08-29 17:28:48.788315 | orchestrator | osism.services.openstackclient : Ensure that all containers are up ------ 1.44s 2025-08-29 17:28:48.788370 | orchestrator | osism.services.openstackclient : Copy openstack wrapper script ---------- 0.93s 2025-08-29 17:28:48.788383 | orchestrator | osism.services.openstackclient : Wait for an healthy service ------------ 0.84s 2025-08-29 17:28:48.788394 | orchestrator | osism.services.openstackclient : Include tasks -------------------------- 0.75s 2025-08-29 17:28:48.788405 | orchestrator | osism.services.openstackclient : Remove ospurge wrapper script ---------- 0.57s 2025-08-29 17:28:48.788415 | orchestrator | osism.services.openstackclient : Copy bash completion script ------------ 0.56s 2025-08-29 17:28:48.788426 | orchestrator | 2025-08-29 17:28:48.788437 | orchestrator | 2025-08-29 17:28:48.788447 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:28:48.788458 | orchestrator | 2025-08-29 17:28:48.788468 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:28:48.788479 | orchestrator | Friday 29 August 2025 17:27:46 +0000 (0:00:00.323) 0:00:00.323 ********* 2025-08-29 17:28:48.788490 | orchestrator | changed: [testbed-manager] => (item=enable_netdata_True) 2025-08-29 17:28:48.788501 | orchestrator | changed: [testbed-node-0] => (item=enable_netdata_True) 2025-08-29 17:28:48.788521 | orchestrator | changed: [testbed-node-1] => (item=enable_netdata_True) 2025-08-29 17:28:48.788534 | orchestrator | changed: [testbed-node-2] => (item=enable_netdata_True) 2025-08-29 17:28:48.788545 | orchestrator | changed: [testbed-node-3] => (item=enable_netdata_True) 2025-08-29 17:28:48.788555 | orchestrator | changed: [testbed-node-4] => (item=enable_netdata_True) 2025-08-29 17:28:48.788566 | orchestrator | changed: [testbed-node-5] => (item=enable_netdata_True) 2025-08-29 17:28:48.788577 | orchestrator | 2025-08-29 17:28:48.788587 | orchestrator | PLAY [Apply role netdata] ****************************************************** 2025-08-29 17:28:48.788598 | orchestrator | 2025-08-29 17:28:48.788609 | orchestrator | TASK [osism.services.netdata : Include distribution specific install tasks] **** 2025-08-29 17:28:48.788620 | orchestrator | Friday 29 August 2025 17:27:49 +0000 (0:00:02.483) 0:00:02.806 ********* 2025-08-29 17:28:48.788646 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:28:48.788667 | orchestrator | 2025-08-29 17:28:48.788678 | orchestrator | TASK [osism.services.netdata : Remove old architecture-dependent repository] *** 2025-08-29 17:28:48.788688 | orchestrator | Friday 29 August 2025 17:27:51 +0000 (0:00:01.824) 0:00:04.631 ********* 2025-08-29 17:28:48.788707 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.788718 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:28:48.788728 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:28:48.788739 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:28:48.788750 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:28:48.788766 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:28:48.788777 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:28:48.788788 | orchestrator | 2025-08-29 17:28:48.788799 | orchestrator | TASK [osism.services.netdata : Install apt-transport-https package] ************ 2025-08-29 17:28:48.788810 | orchestrator | Friday 29 August 2025 17:27:52 +0000 (0:00:01.666) 0:00:06.297 ********* 2025-08-29 17:28:48.788821 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.788831 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:28:48.788842 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:28:48.788853 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:28:48.788863 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:28:48.788874 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:28:48.788885 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:28:48.788895 | orchestrator | 2025-08-29 17:28:48.788906 | orchestrator | TASK [osism.services.netdata : Add repository gpg key] ************************* 2025-08-29 17:28:48.788917 | orchestrator | Friday 29 August 2025 17:27:55 +0000 (0:00:03.193) 0:00:09.491 ********* 2025-08-29 17:28:48.788928 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:28:48.788939 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:28:48.788949 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:28:48.788963 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.788982 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:28:48.788994 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:28:48.789005 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:28:48.789015 | orchestrator | 2025-08-29 17:28:48.789026 | orchestrator | TASK [osism.services.netdata : Add repository] ********************************* 2025-08-29 17:28:48.789036 | orchestrator | Friday 29 August 2025 17:27:58 +0000 (0:00:02.803) 0:00:12.294 ********* 2025-08-29 17:28:48.789051 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:28:48.789062 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.789073 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:28:48.789083 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:28:48.789094 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:28:48.789104 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:28:48.789115 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:28:48.789125 | orchestrator | 2025-08-29 17:28:48.789136 | orchestrator | TASK [osism.services.netdata : Install package netdata] ************************ 2025-08-29 17:28:48.789147 | orchestrator | Friday 29 August 2025 17:28:08 +0000 (0:00:09.712) 0:00:22.006 ********* 2025-08-29 17:28:48.789157 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:28:48.789168 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:28:48.789178 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:28:48.789189 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:28:48.789200 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:28:48.789210 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:28:48.789221 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.789249 | orchestrator | 2025-08-29 17:28:48.789268 | orchestrator | TASK [osism.services.netdata : Include config tasks] *************************** 2025-08-29 17:28:48.789286 | orchestrator | Friday 29 August 2025 17:28:27 +0000 (0:00:19.089) 0:00:41.096 ********* 2025-08-29 17:28:48.789305 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:28:48.789323 | orchestrator | 2025-08-29 17:28:48.789396 | orchestrator | TASK [osism.services.netdata : Copy configuration files] *********************** 2025-08-29 17:28:48.789413 | orchestrator | Friday 29 August 2025 17:28:28 +0000 (0:00:01.299) 0:00:42.395 ********* 2025-08-29 17:28:48.789433 | orchestrator | changed: [testbed-manager] => (item=netdata.conf) 2025-08-29 17:28:48.789463 | orchestrator | changed: [testbed-node-3] => (item=netdata.conf) 2025-08-29 17:28:48.789480 | orchestrator | changed: [testbed-node-1] => (item=netdata.conf) 2025-08-29 17:28:48.789498 | orchestrator | changed: [testbed-node-0] => (item=netdata.conf) 2025-08-29 17:28:48.789515 | orchestrator | changed: [testbed-node-2] => (item=netdata.conf) 2025-08-29 17:28:48.789526 | orchestrator | changed: [testbed-node-4] => (item=netdata.conf) 2025-08-29 17:28:48.789537 | orchestrator | changed: [testbed-node-5] => (item=netdata.conf) 2025-08-29 17:28:48.789548 | orchestrator | changed: [testbed-node-0] => (item=stream.conf) 2025-08-29 17:28:48.789558 | orchestrator | changed: [testbed-manager] => (item=stream.conf) 2025-08-29 17:28:48.789569 | orchestrator | changed: [testbed-node-5] => (item=stream.conf) 2025-08-29 17:28:48.789580 | orchestrator | changed: [testbed-node-3] => (item=stream.conf) 2025-08-29 17:28:48.789590 | orchestrator | changed: [testbed-node-4] => (item=stream.conf) 2025-08-29 17:28:48.789601 | orchestrator | changed: [testbed-node-2] => (item=stream.conf) 2025-08-29 17:28:48.789611 | orchestrator | changed: [testbed-node-1] => (item=stream.conf) 2025-08-29 17:28:48.789622 | orchestrator | 2025-08-29 17:28:48.789632 | orchestrator | TASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] *** 2025-08-29 17:28:48.789644 | orchestrator | Friday 29 August 2025 17:28:33 +0000 (0:00:05.016) 0:00:47.412 ********* 2025-08-29 17:28:48.789655 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.789665 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:28:48.789676 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:28:48.789687 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:28:48.789697 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:28:48.789708 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:28:48.789718 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:28:48.789729 | orchestrator | 2025-08-29 17:28:48.789740 | orchestrator | TASK [osism.services.netdata : Opt out from anonymous statistics] ************** 2025-08-29 17:28:48.789751 | orchestrator | Friday 29 August 2025 17:28:34 +0000 (0:00:01.089) 0:00:48.501 ********* 2025-08-29 17:28:48.789762 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.789772 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:28:48.789783 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:28:48.789793 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:28:48.789804 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:28:48.789815 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:28:48.789825 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:28:48.789836 | orchestrator | 2025-08-29 17:28:48.789846 | orchestrator | TASK [osism.services.netdata : Add netdata user to docker group] *************** 2025-08-29 17:28:48.789866 | orchestrator | Friday 29 August 2025 17:28:36 +0000 (0:00:01.547) 0:00:50.049 ********* 2025-08-29 17:28:48.789877 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:28:48.789888 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:28:48.789899 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:28:48.789909 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:28:48.789920 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.789931 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:28:48.789942 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:28:48.789953 | orchestrator | 2025-08-29 17:28:48.789963 | orchestrator | TASK [osism.services.netdata : Manage service netdata] ************************* 2025-08-29 17:28:48.789974 | orchestrator | Friday 29 August 2025 17:28:37 +0000 (0:00:01.400) 0:00:51.449 ********* 2025-08-29 17:28:48.789985 | orchestrator | ok: [testbed-manager] 2025-08-29 17:28:48.789996 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:28:48.790007 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:28:48.790058 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:28:48.790073 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:28:48.790084 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:28:48.790095 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:28:48.790105 | orchestrator | 2025-08-29 17:28:48.790116 | orchestrator | TASK [osism.services.netdata : Include host type specific tasks] *************** 2025-08-29 17:28:48.790140 | orchestrator | Friday 29 August 2025 17:28:39 +0000 (0:00:01.742) 0:00:53.192 ********* 2025-08-29 17:28:48.790158 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager 2025-08-29 17:28:48.790177 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:28:48.790189 | orchestrator | 2025-08-29 17:28:48.790200 | orchestrator | TASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] ********** 2025-08-29 17:28:48.790211 | orchestrator | Friday 29 August 2025 17:28:40 +0000 (0:00:01.220) 0:00:54.412 ********* 2025-08-29 17:28:48.790222 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.790232 | orchestrator | 2025-08-29 17:28:48.790243 | orchestrator | RUNNING HANDLER [osism.services.netdata : Restart service netdata] ************* 2025-08-29 17:28:48.790254 | orchestrator | Friday 29 August 2025 17:28:42 +0000 (0:00:01.718) 0:00:56.130 ********* 2025-08-29 17:28:48.790265 | orchestrator | changed: [testbed-manager] 2025-08-29 17:28:48.790275 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:28:48.790286 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:28:48.790297 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:28:48.790307 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:28:48.790318 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:28:48.790398 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:28:48.790412 | orchestrator | 2025-08-29 17:28:48.790423 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:28:48.790434 | orchestrator | testbed-manager : ok=16  changed=8  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.790446 | orchestrator | testbed-node-0 : ok=15  changed=7  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.790457 | orchestrator | testbed-node-1 : ok=15  changed=7  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.790468 | orchestrator | testbed-node-2 : ok=15  changed=7  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.790479 | orchestrator | testbed-node-3 : ok=15  changed=7  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.790490 | orchestrator | testbed-node-4 : ok=15  changed=7  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.790500 | orchestrator | testbed-node-5 : ok=15  changed=7  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:28:48.790511 | orchestrator | 2025-08-29 17:28:48.790530 | orchestrator | 2025-08-29 17:28:48.790542 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:28:48.790553 | orchestrator | Friday 29 August 2025 17:28:45 +0000 (0:00:03.201) 0:00:59.332 ********* 2025-08-29 17:28:48.790564 | orchestrator | =============================================================================== 2025-08-29 17:28:48.790575 | orchestrator | osism.services.netdata : Install package netdata ----------------------- 19.09s 2025-08-29 17:28:48.790586 | orchestrator | osism.services.netdata : Add repository --------------------------------- 9.71s 2025-08-29 17:28:48.790596 | orchestrator | osism.services.netdata : Copy configuration files ----------------------- 5.02s 2025-08-29 17:28:48.790607 | orchestrator | osism.services.netdata : Restart service netdata ------------------------ 3.20s 2025-08-29 17:28:48.790618 | orchestrator | osism.services.netdata : Install apt-transport-https package ------------ 3.19s 2025-08-29 17:28:48.790628 | orchestrator | osism.services.netdata : Add repository gpg key ------------------------- 2.80s 2025-08-29 17:28:48.790639 | orchestrator | Group hosts based on enabled services ----------------------------------- 2.48s 2025-08-29 17:28:48.790663 | orchestrator | osism.services.netdata : Include distribution specific install tasks ---- 1.82s 2025-08-29 17:28:48.790674 | orchestrator | osism.services.netdata : Manage service netdata ------------------------- 1.74s 2025-08-29 17:28:48.790684 | orchestrator | osism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 1.72s 2025-08-29 17:28:48.790695 | orchestrator | osism.services.netdata : Remove old architecture-dependent repository --- 1.67s 2025-08-29 17:28:48.790713 | orchestrator | osism.services.netdata : Opt out from anonymous statistics -------------- 1.55s 2025-08-29 17:28:48.790724 | orchestrator | osism.services.netdata : Add netdata user to docker group --------------- 1.40s 2025-08-29 17:28:48.790733 | orchestrator | osism.services.netdata : Include config tasks --------------------------- 1.30s 2025-08-29 17:28:48.790743 | orchestrator | osism.services.netdata : Include host type specific tasks --------------- 1.22s 2025-08-29 17:28:48.790752 | orchestrator | osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.09s 2025-08-29 17:28:48.790762 | orchestrator | 2025-08-29 17:28:48 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:48.790772 | orchestrator | 2025-08-29 17:28:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:51.829120 | orchestrator | 2025-08-29 17:28:51 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:51.829409 | orchestrator | 2025-08-29 17:28:51 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:51.831773 | orchestrator | 2025-08-29 17:28:51 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:51.832863 | orchestrator | 2025-08-29 17:28:51 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:51.832913 | orchestrator | 2025-08-29 17:28:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:54.866605 | orchestrator | 2025-08-29 17:28:54 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:54.866825 | orchestrator | 2025-08-29 17:28:54 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:54.867949 | orchestrator | 2025-08-29 17:28:54 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:54.869347 | orchestrator | 2025-08-29 17:28:54 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:54.869370 | orchestrator | 2025-08-29 17:28:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:28:57.921724 | orchestrator | 2025-08-29 17:28:57 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:28:57.922170 | orchestrator | 2025-08-29 17:28:57 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:28:57.922903 | orchestrator | 2025-08-29 17:28:57 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:28:57.924502 | orchestrator | 2025-08-29 17:28:57 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state STARTED 2025-08-29 17:28:57.924526 | orchestrator | 2025-08-29 17:28:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:00.971808 | orchestrator | 2025-08-29 17:29:00 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:00.974708 | orchestrator | 2025-08-29 17:29:00 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:00.976740 | orchestrator | 2025-08-29 17:29:00 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:00.978059 | orchestrator | 2025-08-29 17:29:00 | INFO  | Task 509fd92c-1fb8-4e63-8845-1c73718c21ea is in state SUCCESS 2025-08-29 17:29:00.978089 | orchestrator | 2025-08-29 17:29:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:04.019312 | orchestrator | 2025-08-29 17:29:04 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:04.019462 | orchestrator | 2025-08-29 17:29:04 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:04.020636 | orchestrator | 2025-08-29 17:29:04 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:04.021688 | orchestrator | 2025-08-29 17:29:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:07.055056 | orchestrator | 2025-08-29 17:29:07 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:07.055498 | orchestrator | 2025-08-29 17:29:07 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:07.056570 | orchestrator | 2025-08-29 17:29:07 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:07.056599 | orchestrator | 2025-08-29 17:29:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:10.101411 | orchestrator | 2025-08-29 17:29:10 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:10.103466 | orchestrator | 2025-08-29 17:29:10 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:10.104819 | orchestrator | 2025-08-29 17:29:10 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:10.104901 | orchestrator | 2025-08-29 17:29:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:13.150218 | orchestrator | 2025-08-29 17:29:13 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:13.152562 | orchestrator | 2025-08-29 17:29:13 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:13.153933 | orchestrator | 2025-08-29 17:29:13 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:13.154070 | orchestrator | 2025-08-29 17:29:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:16.211622 | orchestrator | 2025-08-29 17:29:16 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:16.212260 | orchestrator | 2025-08-29 17:29:16 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:16.214573 | orchestrator | 2025-08-29 17:29:16 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:16.214787 | orchestrator | 2025-08-29 17:29:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:19.276602 | orchestrator | 2025-08-29 17:29:19 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:19.278776 | orchestrator | 2025-08-29 17:29:19 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:19.279840 | orchestrator | 2025-08-29 17:29:19 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:19.279873 | orchestrator | 2025-08-29 17:29:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:22.319265 | orchestrator | 2025-08-29 17:29:22 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:22.320323 | orchestrator | 2025-08-29 17:29:22 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:22.321845 | orchestrator | 2025-08-29 17:29:22 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:22.321877 | orchestrator | 2025-08-29 17:29:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:25.366590 | orchestrator | 2025-08-29 17:29:25 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:25.368294 | orchestrator | 2025-08-29 17:29:25 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:25.369807 | orchestrator | 2025-08-29 17:29:25 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:25.369833 | orchestrator | 2025-08-29 17:29:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:28.416016 | orchestrator | 2025-08-29 17:29:28 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:28.417260 | orchestrator | 2025-08-29 17:29:28 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:28.419254 | orchestrator | 2025-08-29 17:29:28 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:28.419561 | orchestrator | 2025-08-29 17:29:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:31.470278 | orchestrator | 2025-08-29 17:29:31 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:31.470991 | orchestrator | 2025-08-29 17:29:31 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:31.472080 | orchestrator | 2025-08-29 17:29:31 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:31.472111 | orchestrator | 2025-08-29 17:29:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:34.521673 | orchestrator | 2025-08-29 17:29:34 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:34.522889 | orchestrator | 2025-08-29 17:29:34 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:34.522930 | orchestrator | 2025-08-29 17:29:34 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:34.522943 | orchestrator | 2025-08-29 17:29:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:37.568331 | orchestrator | 2025-08-29 17:29:37 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:37.570673 | orchestrator | 2025-08-29 17:29:37 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:37.572225 | orchestrator | 2025-08-29 17:29:37 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:37.572264 | orchestrator | 2025-08-29 17:29:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:40.616948 | orchestrator | 2025-08-29 17:29:40 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:40.618791 | orchestrator | 2025-08-29 17:29:40 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:40.620082 | orchestrator | 2025-08-29 17:29:40 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:40.620180 | orchestrator | 2025-08-29 17:29:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:43.671771 | orchestrator | 2025-08-29 17:29:43 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:43.672668 | orchestrator | 2025-08-29 17:29:43 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:43.677026 | orchestrator | 2025-08-29 17:29:43 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:43.677101 | orchestrator | 2025-08-29 17:29:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:46.718927 | orchestrator | 2025-08-29 17:29:46 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:46.719443 | orchestrator | 2025-08-29 17:29:46 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:46.721268 | orchestrator | 2025-08-29 17:29:46 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:46.721405 | orchestrator | 2025-08-29 17:29:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:49.770137 | orchestrator | 2025-08-29 17:29:49 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:49.772436 | orchestrator | 2025-08-29 17:29:49 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:49.774404 | orchestrator | 2025-08-29 17:29:49 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:49.774436 | orchestrator | 2025-08-29 17:29:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:52.821077 | orchestrator | 2025-08-29 17:29:52 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:52.822990 | orchestrator | 2025-08-29 17:29:52 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:52.825650 | orchestrator | 2025-08-29 17:29:52 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:52.825987 | orchestrator | 2025-08-29 17:29:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:55.863868 | orchestrator | 2025-08-29 17:29:55 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:55.865499 | orchestrator | 2025-08-29 17:29:55 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:55.866869 | orchestrator | 2025-08-29 17:29:55 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:55.866895 | orchestrator | 2025-08-29 17:29:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:29:58.899612 | orchestrator | 2025-08-29 17:29:58 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:29:58.900680 | orchestrator | 2025-08-29 17:29:58 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:29:58.902360 | orchestrator | 2025-08-29 17:29:58 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:29:58.902411 | orchestrator | 2025-08-29 17:29:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:01.956688 | orchestrator | 2025-08-29 17:30:01 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:01.958928 | orchestrator | 2025-08-29 17:30:01 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:30:01.960495 | orchestrator | 2025-08-29 17:30:01 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:01.960723 | orchestrator | 2025-08-29 17:30:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:04.997862 | orchestrator | 2025-08-29 17:30:04 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:04.998419 | orchestrator | 2025-08-29 17:30:04 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:30:04.999462 | orchestrator | 2025-08-29 17:30:04 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:04.999587 | orchestrator | 2025-08-29 17:30:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:08.030887 | orchestrator | 2025-08-29 17:30:08 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:08.032047 | orchestrator | 2025-08-29 17:30:08 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:30:08.033059 | orchestrator | 2025-08-29 17:30:08 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:08.033225 | orchestrator | 2025-08-29 17:30:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:11.065271 | orchestrator | 2025-08-29 17:30:11 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:11.067711 | orchestrator | 2025-08-29 17:30:11 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:30:11.070383 | orchestrator | 2025-08-29 17:30:11 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:11.070412 | orchestrator | 2025-08-29 17:30:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:14.111411 | orchestrator | 2025-08-29 17:30:14 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:14.112262 | orchestrator | 2025-08-29 17:30:14 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:30:14.113416 | orchestrator | 2025-08-29 17:30:14 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:14.113449 | orchestrator | 2025-08-29 17:30:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:17.158531 | orchestrator | 2025-08-29 17:30:17 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:17.159313 | orchestrator | 2025-08-29 17:30:17 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state STARTED 2025-08-29 17:30:17.161006 | orchestrator | 2025-08-29 17:30:17 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:17.161163 | orchestrator | 2025-08-29 17:30:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:20.196582 | orchestrator | 2025-08-29 17:30:20 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:20.202095 | orchestrator | 2025-08-29 17:30:20 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:20.202737 | orchestrator | 2025-08-29 17:30:20 | INFO  | Task d6b8f4c7-494f-4564-959f-e55c39bfa665 is in state SUCCESS 2025-08-29 17:30:20.203203 | orchestrator | 2025-08-29 17:30:20.203226 | orchestrator | 2025-08-29 17:30:20.203235 | orchestrator | PLAY [Apply role phpmyadmin] *************************************************** 2025-08-29 17:30:20.203244 | orchestrator | 2025-08-29 17:30:20.203252 | orchestrator | TASK [osism.services.phpmyadmin : Create traefik external network] ************* 2025-08-29 17:30:20.203261 | orchestrator | Friday 29 August 2025 17:28:07 +0000 (0:00:00.354) 0:00:00.354 ********* 2025-08-29 17:30:20.203269 | orchestrator | ok: [testbed-manager] 2025-08-29 17:30:20.203279 | orchestrator | 2025-08-29 17:30:20.203305 | orchestrator | TASK [osism.services.phpmyadmin : Create required directories] ***************** 2025-08-29 17:30:20.203313 | orchestrator | Friday 29 August 2025 17:28:08 +0000 (0:00:00.891) 0:00:01.245 ********* 2025-08-29 17:30:20.203321 | orchestrator | changed: [testbed-manager] => (item=/opt/phpmyadmin) 2025-08-29 17:30:20.203330 | orchestrator | 2025-08-29 17:30:20.203362 | orchestrator | TASK [osism.services.phpmyadmin : Copy docker-compose.yml file] **************** 2025-08-29 17:30:20.203373 | orchestrator | Friday 29 August 2025 17:28:09 +0000 (0:00:01.188) 0:00:02.434 ********* 2025-08-29 17:30:20.203381 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.203390 | orchestrator | 2025-08-29 17:30:20.203398 | orchestrator | TASK [osism.services.phpmyadmin : Manage phpmyadmin service] ******************* 2025-08-29 17:30:20.203406 | orchestrator | Friday 29 August 2025 17:28:10 +0000 (0:00:01.064) 0:00:03.498 ********* 2025-08-29 17:30:20.203414 | orchestrator | FAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left). 2025-08-29 17:30:20.203423 | orchestrator | ok: [testbed-manager] 2025-08-29 17:30:20.203521 | orchestrator | 2025-08-29 17:30:20.203534 | orchestrator | RUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] ******* 2025-08-29 17:30:20.203542 | orchestrator | Friday 29 August 2025 17:28:55 +0000 (0:00:44.715) 0:00:48.214 ********* 2025-08-29 17:30:20.203550 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.203601 | orchestrator | 2025-08-29 17:30:20.203611 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:30:20.203619 | orchestrator | testbed-manager : ok=5  changed=3  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:30:20.203628 | orchestrator | 2025-08-29 17:30:20.203636 | orchestrator | 2025-08-29 17:30:20.203644 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:30:20.203652 | orchestrator | Friday 29 August 2025 17:28:59 +0000 (0:00:03.904) 0:00:52.118 ********* 2025-08-29 17:30:20.203659 | orchestrator | =============================================================================== 2025-08-29 17:30:20.203667 | orchestrator | osism.services.phpmyadmin : Manage phpmyadmin service ------------------ 44.72s 2025-08-29 17:30:20.203675 | orchestrator | osism.services.phpmyadmin : Restart phpmyadmin service ------------------ 3.90s 2025-08-29 17:30:20.203683 | orchestrator | osism.services.phpmyadmin : Create required directories ----------------- 1.19s 2025-08-29 17:30:20.203691 | orchestrator | osism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 1.06s 2025-08-29 17:30:20.203698 | orchestrator | osism.services.phpmyadmin : Create traefik external network ------------- 0.89s 2025-08-29 17:30:20.203706 | orchestrator | 2025-08-29 17:30:20.205153 | orchestrator | 2025-08-29 17:30:20.205187 | orchestrator | PLAY [Apply role common] ******************************************************* 2025-08-29 17:30:20.205196 | orchestrator | 2025-08-29 17:30:20.205204 | orchestrator | TASK [common : include_tasks] ************************************************** 2025-08-29 17:30:20.205212 | orchestrator | Friday 29 August 2025 17:27:40 +0000 (0:00:00.263) 0:00:00.263 ********* 2025-08-29 17:30:20.205221 | orchestrator | included: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:30:20.205230 | orchestrator | 2025-08-29 17:30:20.205238 | orchestrator | TASK [common : Ensuring config directories exist] ****************************** 2025-08-29 17:30:20.205246 | orchestrator | Friday 29 August 2025 17:27:41 +0000 (0:00:01.038) 0:00:01.301 ********* 2025-08-29 17:30:20.205254 | orchestrator | changed: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron']) 2025-08-29 17:30:20.205265 | orchestrator | changed: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron']) 2025-08-29 17:30:20.205274 | orchestrator | changed: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron']) 2025-08-29 17:30:20.205281 | orchestrator | changed: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron']) 2025-08-29 17:30:20.205289 | orchestrator | changed: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd']) 2025-08-29 17:30:20.205297 | orchestrator | changed: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd']) 2025-08-29 17:30:20.205305 | orchestrator | changed: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron']) 2025-08-29 17:30:20.205312 | orchestrator | changed: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd']) 2025-08-29 17:30:20.205320 | orchestrator | changed: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd']) 2025-08-29 17:30:20.205328 | orchestrator | changed: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron']) 2025-08-29 17:30:20.205336 | orchestrator | changed: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron']) 2025-08-29 17:30:20.206176 | orchestrator | changed: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox']) 2025-08-29 17:30:20.206225 | orchestrator | changed: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox']) 2025-08-29 17:30:20.206235 | orchestrator | changed: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox']) 2025-08-29 17:30:20.206266 | orchestrator | changed: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd']) 2025-08-29 17:30:20.206274 | orchestrator | changed: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox']) 2025-08-29 17:30:20.206282 | orchestrator | changed: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd']) 2025-08-29 17:30:20.206289 | orchestrator | changed: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd']) 2025-08-29 17:30:20.206296 | orchestrator | changed: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox']) 2025-08-29 17:30:20.206304 | orchestrator | changed: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox']) 2025-08-29 17:30:20.206311 | orchestrator | changed: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox']) 2025-08-29 17:30:20.206319 | orchestrator | 2025-08-29 17:30:20.206326 | orchestrator | TASK [common : include_tasks] ************************************************** 2025-08-29 17:30:20.206334 | orchestrator | Friday 29 August 2025 17:27:45 +0000 (0:00:03.796) 0:00:05.098 ********* 2025-08-29 17:30:20.206357 | orchestrator | included: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:30:20.206366 | orchestrator | 2025-08-29 17:30:20.206374 | orchestrator | TASK [service-cert-copy : common | Copying over extra CA certificates] ********* 2025-08-29 17:30:20.206381 | orchestrator | Friday 29 August 2025 17:27:46 +0000 (0:00:01.229) 0:00:06.327 ********* 2025-08-29 17:30:20.206393 | orchestrator | changed: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.206405 | orchestrator | changed: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.206438 | orchestrator | changed: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.206467 | orchestrator | changed: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.206477 | orchestrator | changed: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206492 | orchestrator | changed: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.206500 | orchestrator | changed: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.206508 | orchestrator | changed: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206516 | orchestrator | changed: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.206532 | orchestrator | changed: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206545 | orchestrator | changed: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206557 | orchestrator | changed: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206572 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206579 | orchestrator | changed: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206588 | orchestrator | changed: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206595 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206608 | orchestrator | changed: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206616 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206628 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206640 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206648 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.206656 | orchestrator | 2025-08-29 17:30:20.206664 | orchestrator | TASK [service-cert-copy : common | Copying over backend internal TLS certificate] *** 2025-08-29 17:30:20.206671 | orchestrator | Friday 29 August 2025 17:27:51 +0000 (0:00:04.806) 0:00:11.133 ********* 2025-08-29 17:30:20.206679 | orchestrator | skipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206688 | orchestrator | skipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206696 | orchestrator | skipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206703 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:30:20.206716 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206728 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206740 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206748 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:30:20.206756 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206764 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206772 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206779 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:30:20.206787 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206795 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206813 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206829 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206838 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206845 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206853 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206861 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206869 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206876 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:30:20.206884 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:30:20.206891 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:30:20.206908 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206921 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206929 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206937 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:30:20.206944 | orchestrator | 2025-08-29 17:30:20.206951 | orchestrator | TASK [service-cert-copy : common | Copying over backend internal TLS key] ****** 2025-08-29 17:30:20.206959 | orchestrator | Friday 29 August 2025 17:27:53 +0000 (0:00:02.059) 0:00:13.193 ********* 2025-08-29 17:30:20.206967 | orchestrator | skipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.206978 | orchestrator | skipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206987 | orchestrator | skipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.206994 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.207006 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207019 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207026 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:30:20.207034 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:30:20.207048 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.207056 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207064 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207071 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:30:20.207079 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.207087 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207094 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207112 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.207124 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207131 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207139 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:30:20.207146 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:30:20.207154 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.207162 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207170 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207177 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:30:20.207185 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})  2025-08-29 17:30:20.207200 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207209 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.207216 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:30:20.207224 | orchestrator | 2025-08-29 17:30:20.207231 | orchestrator | TASK [common : Copying over /run subdirectories conf] ************************** 2025-08-29 17:30:20.207241 | orchestrator | Friday 29 August 2025 17:27:56 +0000 (0:00:03.318) 0:00:16.511 ********* 2025-08-29 17:30:20.207249 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:30:20.207256 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:30:20.207263 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:30:20.207271 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:30:20.207278 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:30:20.207285 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:30:20.207292 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:30:20.207299 | orchestrator | 2025-08-29 17:30:20.207307 | orchestrator | TASK [common : Restart systemd-tmpfiles] *************************************** 2025-08-29 17:30:20.207314 | orchestrator | Friday 29 August 2025 17:27:58 +0000 (0:00:01.261) 0:00:17.772 ********* 2025-08-29 17:30:20.207321 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:30:20.207329 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:30:20.207336 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:30:20.207356 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:30:20.207363 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:30:20.207371 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:30:20.207378 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:30:20.207385 | orchestrator | 2025-08-29 17:30:20.207393 | orchestrator | TASK [common : Copying over config.json files for services] ******************** 2025-08-29 17:30:20.207400 | orchestrator | Friday 29 August 2025 17:28:00 +0000 (0:00:01.932) 0:00:19.704 ********* 2025-08-29 17:30:20.207408 | orchestrator | changed: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.207415 | orchestrator | changed: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.207428 | orchestrator | changed: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.207436 | orchestrator | changed: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.207449 | orchestrator | changed: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.207461 | orchestrator | changed: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.207469 | orchestrator | changed: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.207476 | orchestrator | changed: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207484 | orchestrator | changed: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207496 | orchestrator | changed: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207504 | orchestrator | changed: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207521 | orchestrator | changed: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207533 | orchestrator | changed: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207541 | orchestrator | changed: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207549 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207561 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207569 | orchestrator | changed: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207577 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207585 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207601 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207610 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.207617 | orchestrator | 2025-08-29 17:30:20.207625 | orchestrator | TASK [common : Find custom fluentd input config files] ************************* 2025-08-29 17:30:20.207632 | orchestrator | Friday 29 August 2025 17:28:05 +0000 (0:00:05.373) 0:00:25.078 ********* 2025-08-29 17:30:20.207639 | orchestrator | [WARNING]: Skipped 2025-08-29 17:30:20.207650 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due 2025-08-29 17:30:20.207657 | orchestrator | to this access issue: 2025-08-29 17:30:20.207665 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a 2025-08-29 17:30:20.207672 | orchestrator | directory 2025-08-29 17:30:20.207679 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:30:20.207687 | orchestrator | 2025-08-29 17:30:20.207694 | orchestrator | TASK [common : Find custom fluentd filter config files] ************************ 2025-08-29 17:30:20.207701 | orchestrator | Friday 29 August 2025 17:28:07 +0000 (0:00:01.605) 0:00:26.684 ********* 2025-08-29 17:30:20.207708 | orchestrator | [WARNING]: Skipped 2025-08-29 17:30:20.207716 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due 2025-08-29 17:30:20.207723 | orchestrator | to this access issue: 2025-08-29 17:30:20.207730 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a 2025-08-29 17:30:20.207742 | orchestrator | directory 2025-08-29 17:30:20.207749 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:30:20.207756 | orchestrator | 2025-08-29 17:30:20.207763 | orchestrator | TASK [common : Find custom fluentd format config files] ************************ 2025-08-29 17:30:20.207771 | orchestrator | Friday 29 August 2025 17:28:08 +0000 (0:00:01.123) 0:00:27.808 ********* 2025-08-29 17:30:20.207778 | orchestrator | [WARNING]: Skipped 2025-08-29 17:30:20.207785 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due 2025-08-29 17:30:20.207792 | orchestrator | to this access issue: 2025-08-29 17:30:20.207799 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a 2025-08-29 17:30:20.207807 | orchestrator | directory 2025-08-29 17:30:20.207814 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:30:20.207821 | orchestrator | 2025-08-29 17:30:20.207828 | orchestrator | TASK [common : Find custom fluentd output config files] ************************ 2025-08-29 17:30:20.207836 | orchestrator | Friday 29 August 2025 17:28:09 +0000 (0:00:00.850) 0:00:28.658 ********* 2025-08-29 17:30:20.207843 | orchestrator | [WARNING]: Skipped 2025-08-29 17:30:20.207851 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due 2025-08-29 17:30:20.207858 | orchestrator | to this access issue: 2025-08-29 17:30:20.207865 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a 2025-08-29 17:30:20.207872 | orchestrator | directory 2025-08-29 17:30:20.207879 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:30:20.207887 | orchestrator | 2025-08-29 17:30:20.207894 | orchestrator | TASK [common : Copying over fluentd.conf] ************************************** 2025-08-29 17:30:20.207901 | orchestrator | Friday 29 August 2025 17:28:09 +0000 (0:00:00.781) 0:00:29.440 ********* 2025-08-29 17:30:20.207908 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:20.207916 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:20.207923 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.207930 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:30:20.207937 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:30:20.207944 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:30:20.207951 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:20.207959 | orchestrator | 2025-08-29 17:30:20.207966 | orchestrator | TASK [common : Copying over cron logrotate config file] ************************ 2025-08-29 17:30:20.207973 | orchestrator | Friday 29 August 2025 17:28:14 +0000 (0:00:04.198) 0:00:33.639 ********* 2025-08-29 17:30:20.207980 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2) 2025-08-29 17:30:20.207988 | orchestrator | changed: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2) 2025-08-29 17:30:20.207995 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2) 2025-08-29 17:30:20.208002 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2) 2025-08-29 17:30:20.208009 | orchestrator | changed: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2) 2025-08-29 17:30:20.208017 | orchestrator | changed: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2) 2025-08-29 17:30:20.208024 | orchestrator | changed: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2) 2025-08-29 17:30:20.208031 | orchestrator | 2025-08-29 17:30:20.208038 | orchestrator | TASK [common : Ensure RabbitMQ Erlang cookie exists] *************************** 2025-08-29 17:30:20.208046 | orchestrator | Friday 29 August 2025 17:28:16 +0000 (0:00:02.420) 0:00:36.059 ********* 2025-08-29 17:30:20.208053 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.208060 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:20.208067 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:30:20.208074 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:20.208090 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:20.208097 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:30:20.208105 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:30:20.208112 | orchestrator | 2025-08-29 17:30:20.208120 | orchestrator | TASK [common : Ensuring config directories have correct owner and permission] *** 2025-08-29 17:30:20.208127 | orchestrator | Friday 29 August 2025 17:28:18 +0000 (0:00:02.003) 0:00:38.063 ********* 2025-08-29 17:30:20.208138 | orchestrator | ok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208146 | orchestrator | skipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.208154 | orchestrator | ok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208171 | orchestrator | ok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208179 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.208187 | orchestrator | ok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208198 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.208210 | orchestrator | ok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208218 | orchestrator | ok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208226 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.208239 | orchestrator | ok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208247 | orchestrator | ok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208255 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.208262 | orchestrator | ok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208281 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.208293 | orchestrator | ok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208301 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:30:20.208308 | orchestrator | ok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208316 | orchestrator | ok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208324 | orchestrator | ok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208332 | orchestrator | ok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208366 | orchestrator | 2025-08-29 17:30:20.208374 | orchestrator | TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************ 2025-08-29 17:30:20.208388 | orchestrator | Friday 29 August 2025 17:28:21 +0000 (0:00:02.695) 0:00:40.758 ********* 2025-08-29 17:30:20.208395 | orchestrator | changed: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2) 2025-08-29 17:30:20.208403 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2) 2025-08-29 17:30:20.208410 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2) 2025-08-29 17:30:20.208417 | orchestrator | changed: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2) 2025-08-29 17:30:20.208425 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2) 2025-08-29 17:30:20.208432 | orchestrator | changed: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2) 2025-08-29 17:30:20.208439 | orchestrator | changed: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2) 2025-08-29 17:30:20.208446 | orchestrator | 2025-08-29 17:30:20.208457 | orchestrator | TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] ********************** 2025-08-29 17:30:20.208465 | orchestrator | Friday 29 August 2025 17:28:25 +0000 (0:00:04.133) 0:00:44.892 ********* 2025-08-29 17:30:20.208472 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2) 2025-08-29 17:30:20.208480 | orchestrator | changed: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2) 2025-08-29 17:30:20.208487 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2) 2025-08-29 17:30:20.208494 | orchestrator | changed: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2) 2025-08-29 17:30:20.208501 | orchestrator | changed: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2) 2025-08-29 17:30:20.208508 | orchestrator | changed: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2) 2025-08-29 17:30:20.208519 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2) 2025-08-29 17:30:20.208526 | orchestrator | 2025-08-29 17:30:20.208534 | orchestrator | TASK [common : Check common containers] **************************************** 2025-08-29 17:30:20.208541 | orchestrator | Friday 29 August 2025 17:28:27 +0000 (0:00:02.584) 0:00:47.476 ********* 2025-08-29 17:30:20.208548 | orchestrator | changed: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208556 | orchestrator | changed: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208564 | orchestrator | changed: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208577 | orchestrator | changed: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208585 | orchestrator | changed: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208764 | orchestrator | changed: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208783 | orchestrator | changed: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208791 | orchestrator | changed: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208799 | orchestrator | changed: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208807 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208821 | orchestrator | changed: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208828 | orchestrator | changed: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) 2025-08-29 17:30:20.208841 | orchestrator | changed: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208849 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208860 | orchestrator | changed: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208868 | orchestrator | changed: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208876 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208888 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208895 | orchestrator | changed: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208903 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208910 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:30:20.208918 | orchestrator | 2025-08-29 17:30:20.208925 | orchestrator | TASK [common : Creating log volume] ******************************************** 2025-08-29 17:30:20.208933 | orchestrator | Friday 29 August 2025 17:28:31 +0000 (0:00:03.713) 0:00:51.189 ********* 2025-08-29 17:30:20.208943 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.208951 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:20.208958 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:20.208965 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:20.208972 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:30:20.208979 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:30:20.208987 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:30:20.208994 | orchestrator | 2025-08-29 17:30:20.209001 | orchestrator | TASK [common : Link kolla_logs volume to /var/log/kolla] *********************** 2025-08-29 17:30:20.209008 | orchestrator | Friday 29 August 2025 17:28:33 +0000 (0:00:02.067) 0:00:53.257 ********* 2025-08-29 17:30:20.209015 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:20.209022 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.209029 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:20.209037 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:20.209044 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:30:20.209051 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:30:20.209058 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:30:20.209065 | orchestrator | 2025-08-29 17:30:20.209076 | orchestrator | TASK [common : Flush handlers] ************************************************* 2025-08-29 17:30:20.209083 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:01.398) 0:00:54.656 ********* 2025-08-29 17:30:20.209091 | orchestrator | 2025-08-29 17:30:20.209098 | orchestrator | TASK [common : Flush handlers] ************************************************* 2025-08-29 17:30:20.209105 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:00.050) 0:00:54.707 ********* 2025-08-29 17:30:20.209112 | orchestrator | 2025-08-29 17:30:20.209120 | orchestrator | TASK [common : Flush handlers] ************************************************* 2025-08-29 17:30:20.209127 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:00.048) 0:00:54.755 ********* 2025-08-29 17:30:20.209139 | orchestrator | 2025-08-29 17:30:20.209146 | orchestrator | TASK [common : Flush handlers] ************************************************* 2025-08-29 17:30:20.209153 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:00.165) 0:00:54.920 ********* 2025-08-29 17:30:20.209161 | orchestrator | 2025-08-29 17:30:20.209168 | orchestrator | TASK [common : Flush handlers] ************************************************* 2025-08-29 17:30:20.209175 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:00.050) 0:00:54.970 ********* 2025-08-29 17:30:20.209182 | orchestrator | 2025-08-29 17:30:20.209189 | orchestrator | TASK [common : Flush handlers] ************************************************* 2025-08-29 17:30:20.209197 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:00.050) 0:00:55.021 ********* 2025-08-29 17:30:20.209204 | orchestrator | 2025-08-29 17:30:20.209211 | orchestrator | TASK [common : Flush handlers] ************************************************* 2025-08-29 17:30:20.209218 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:00.048) 0:00:55.069 ********* 2025-08-29 17:30:20.209226 | orchestrator | 2025-08-29 17:30:20.209233 | orchestrator | RUNNING HANDLER [common : Restart fluentd container] *************************** 2025-08-29 17:30:20.209240 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:00.088) 0:00:55.158 ********* 2025-08-29 17:30:20.209247 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:20.209255 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:20.209262 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:30:20.209269 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:20.209276 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.209283 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:30:20.209290 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:30:20.209297 | orchestrator | 2025-08-29 17:30:20.209305 | orchestrator | RUNNING HANDLER [common : Restart kolla-toolbox container] ********************* 2025-08-29 17:30:20.209312 | orchestrator | Friday 29 August 2025 17:29:16 +0000 (0:00:41.403) 0:01:36.562 ********* 2025-08-29 17:30:20.209319 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:20.209326 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:20.209333 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:20.209376 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.209384 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:30:20.209392 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:30:20.209400 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:30:20.209456 | orchestrator | 2025-08-29 17:30:20.209465 | orchestrator | RUNNING HANDLER [common : Initializing toolbox container using normal user] **** 2025-08-29 17:30:20.209473 | orchestrator | Friday 29 August 2025 17:30:07 +0000 (0:00:50.219) 0:02:26.782 ********* 2025-08-29 17:30:20.209481 | orchestrator | ok: [testbed-manager] 2025-08-29 17:30:20.209490 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:30:20.209498 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:30:20.209506 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:30:20.209514 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:30:20.209522 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:30:20.209530 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:30:20.209538 | orchestrator | 2025-08-29 17:30:20.209546 | orchestrator | RUNNING HANDLER [common : Restart cron container] ****************************** 2025-08-29 17:30:20.209554 | orchestrator | Friday 29 August 2025 17:30:08 +0000 (0:00:01.757) 0:02:28.539 ********* 2025-08-29 17:30:20.209562 | orchestrator | changed: [testbed-manager] 2025-08-29 17:30:20.209570 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:20.209578 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:30:20.209586 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:20.209593 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:20.209602 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:30:20.209610 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:30:20.209618 | orchestrator | 2025-08-29 17:30:20.209626 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:30:20.209635 | orchestrator | testbed-manager : ok=22  changed=14  unreachable=0 failed=0 skipped=4  rescued=0 ignored=0 2025-08-29 17:30:20.209652 | orchestrator | testbed-node-0 : ok=18  changed=14  unreachable=0 failed=0 skipped=4  rescued=0 ignored=0 2025-08-29 17:30:20.209659 | orchestrator | testbed-node-1 : ok=18  changed=14  unreachable=0 failed=0 skipped=4  rescued=0 ignored=0 2025-08-29 17:30:20.209671 | orchestrator | testbed-node-2 : ok=18  changed=14  unreachable=0 failed=0 skipped=4  rescued=0 ignored=0 2025-08-29 17:30:20.209679 | orchestrator | testbed-node-3 : ok=18  changed=14  unreachable=0 failed=0 skipped=4  rescued=0 ignored=0 2025-08-29 17:30:20.209686 | orchestrator | testbed-node-4 : ok=18  changed=14  unreachable=0 failed=0 skipped=4  rescued=0 ignored=0 2025-08-29 17:30:20.209694 | orchestrator | testbed-node-5 : ok=18  changed=14  unreachable=0 failed=0 skipped=4  rescued=0 ignored=0 2025-08-29 17:30:20.209701 | orchestrator | 2025-08-29 17:30:20.209708 | orchestrator | 2025-08-29 17:30:20.209715 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:30:20.209728 | orchestrator | Friday 29 August 2025 17:30:18 +0000 (0:00:09.211) 0:02:37.751 ********* 2025-08-29 17:30:20.209735 | orchestrator | =============================================================================== 2025-08-29 17:30:20.209742 | orchestrator | common : Restart kolla-toolbox container ------------------------------- 50.22s 2025-08-29 17:30:20.209749 | orchestrator | common : Restart fluentd container ------------------------------------- 41.40s 2025-08-29 17:30:20.209757 | orchestrator | common : Restart cron container ----------------------------------------- 9.21s 2025-08-29 17:30:20.209764 | orchestrator | common : Copying over config.json files for services -------------------- 5.37s 2025-08-29 17:30:20.209771 | orchestrator | service-cert-copy : common | Copying over extra CA certificates --------- 4.81s 2025-08-29 17:30:20.209778 | orchestrator | common : Copying over fluentd.conf -------------------------------------- 4.20s 2025-08-29 17:30:20.209786 | orchestrator | common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 4.13s 2025-08-29 17:30:20.209793 | orchestrator | common : Ensuring config directories exist ------------------------------ 3.80s 2025-08-29 17:30:20.209800 | orchestrator | common : Check common containers ---------------------------------------- 3.71s 2025-08-29 17:30:20.209807 | orchestrator | service-cert-copy : common | Copying over backend internal TLS key ------ 3.32s 2025-08-29 17:30:20.209814 | orchestrator | common : Ensuring config directories have correct owner and permission --- 2.70s 2025-08-29 17:30:20.209821 | orchestrator | common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.58s 2025-08-29 17:30:20.209829 | orchestrator | common : Copying over cron logrotate config file ------------------------ 2.42s 2025-08-29 17:30:20.209836 | orchestrator | common : Creating log volume -------------------------------------------- 2.07s 2025-08-29 17:30:20.209843 | orchestrator | service-cert-copy : common | Copying over backend internal TLS certificate --- 2.06s 2025-08-29 17:30:20.209850 | orchestrator | common : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.00s 2025-08-29 17:30:20.209857 | orchestrator | common : Restart systemd-tmpfiles --------------------------------------- 1.93s 2025-08-29 17:30:20.209865 | orchestrator | common : Initializing toolbox container using normal user --------------- 1.76s 2025-08-29 17:30:20.209872 | orchestrator | common : Find custom fluentd input config files ------------------------- 1.61s 2025-08-29 17:30:20.209880 | orchestrator | common : Link kolla_logs volume to /var/log/kolla ----------------------- 1.40s 2025-08-29 17:30:20.209887 | orchestrator | 2025-08-29 17:30:20 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:20.209894 | orchestrator | 2025-08-29 17:30:20 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:20.209901 | orchestrator | 2025-08-29 17:30:20 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:20.209914 | orchestrator | 2025-08-29 17:30:20 | INFO  | Task 133758f0-a5da-46ab-bdd7-e7f4ca3f3070 is in state STARTED 2025-08-29 17:30:20.210497 | orchestrator | 2025-08-29 17:30:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:23.247607 | orchestrator | 2025-08-29 17:30:23 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:23.247864 | orchestrator | 2025-08-29 17:30:23 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:23.248478 | orchestrator | 2025-08-29 17:30:23 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:23.249079 | orchestrator | 2025-08-29 17:30:23 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:23.250510 | orchestrator | 2025-08-29 17:30:23 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:23.250554 | orchestrator | 2025-08-29 17:30:23 | INFO  | Task 133758f0-a5da-46ab-bdd7-e7f4ca3f3070 is in state STARTED 2025-08-29 17:30:23.250566 | orchestrator | 2025-08-29 17:30:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:26.275451 | orchestrator | 2025-08-29 17:30:26 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:26.275509 | orchestrator | 2025-08-29 17:30:26 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:26.275520 | orchestrator | 2025-08-29 17:30:26 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:26.276026 | orchestrator | 2025-08-29 17:30:26 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:26.276666 | orchestrator | 2025-08-29 17:30:26 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:26.277317 | orchestrator | 2025-08-29 17:30:26 | INFO  | Task 133758f0-a5da-46ab-bdd7-e7f4ca3f3070 is in state STARTED 2025-08-29 17:30:26.278431 | orchestrator | 2025-08-29 17:30:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:29.309525 | orchestrator | 2025-08-29 17:30:29 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:29.309756 | orchestrator | 2025-08-29 17:30:29 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:29.310268 | orchestrator | 2025-08-29 17:30:29 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:29.314785 | orchestrator | 2025-08-29 17:30:29 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:29.315365 | orchestrator | 2025-08-29 17:30:29 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:29.315889 | orchestrator | 2025-08-29 17:30:29 | INFO  | Task 133758f0-a5da-46ab-bdd7-e7f4ca3f3070 is in state STARTED 2025-08-29 17:30:29.315914 | orchestrator | 2025-08-29 17:30:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:32.344478 | orchestrator | 2025-08-29 17:30:32 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:32.345128 | orchestrator | 2025-08-29 17:30:32 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:32.345911 | orchestrator | 2025-08-29 17:30:32 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:32.346661 | orchestrator | 2025-08-29 17:30:32 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:32.347741 | orchestrator | 2025-08-29 17:30:32 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:32.348240 | orchestrator | 2025-08-29 17:30:32 | INFO  | Task 133758f0-a5da-46ab-bdd7-e7f4ca3f3070 is in state STARTED 2025-08-29 17:30:32.348405 | orchestrator | 2025-08-29 17:30:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:35.391645 | orchestrator | 2025-08-29 17:30:35 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:35.391838 | orchestrator | 2025-08-29 17:30:35 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:35.391870 | orchestrator | 2025-08-29 17:30:35 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:35.392727 | orchestrator | 2025-08-29 17:30:35 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:35.393305 | orchestrator | 2025-08-29 17:30:35 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:35.394761 | orchestrator | 2025-08-29 17:30:35 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:35.394807 | orchestrator | 2025-08-29 17:30:35 | INFO  | Task 133758f0-a5da-46ab-bdd7-e7f4ca3f3070 is in state SUCCESS 2025-08-29 17:30:35.394821 | orchestrator | 2025-08-29 17:30:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:38.455828 | orchestrator | 2025-08-29 17:30:38 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:38.456639 | orchestrator | 2025-08-29 17:30:38 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:38.459012 | orchestrator | 2025-08-29 17:30:38 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:38.462521 | orchestrator | 2025-08-29 17:30:38 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:38.463173 | orchestrator | 2025-08-29 17:30:38 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:38.464258 | orchestrator | 2025-08-29 17:30:38 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:38.465734 | orchestrator | 2025-08-29 17:30:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:41.587167 | orchestrator | 2025-08-29 17:30:41 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:41.591550 | orchestrator | 2025-08-29 17:30:41 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:41.593173 | orchestrator | 2025-08-29 17:30:41 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:41.593956 | orchestrator | 2025-08-29 17:30:41 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:41.594522 | orchestrator | 2025-08-29 17:30:41 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:41.595199 | orchestrator | 2025-08-29 17:30:41 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:41.596437 | orchestrator | 2025-08-29 17:30:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:44.656226 | orchestrator | 2025-08-29 17:30:44 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:44.656509 | orchestrator | 2025-08-29 17:30:44 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:44.657274 | orchestrator | 2025-08-29 17:30:44 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:44.657860 | orchestrator | 2025-08-29 17:30:44 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:44.658568 | orchestrator | 2025-08-29 17:30:44 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:44.659170 | orchestrator | 2025-08-29 17:30:44 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:44.659192 | orchestrator | 2025-08-29 17:30:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:47.694422 | orchestrator | 2025-08-29 17:30:47 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:47.695417 | orchestrator | 2025-08-29 17:30:47 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:47.697408 | orchestrator | 2025-08-29 17:30:47 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:47.699729 | orchestrator | 2025-08-29 17:30:47 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:47.703148 | orchestrator | 2025-08-29 17:30:47 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state STARTED 2025-08-29 17:30:47.705299 | orchestrator | 2025-08-29 17:30:47 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:47.705962 | orchestrator | 2025-08-29 17:30:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:50.737059 | orchestrator | 2025-08-29 17:30:50 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:50.737161 | orchestrator | 2025-08-29 17:30:50 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:50.737896 | orchestrator | 2025-08-29 17:30:50 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:50.738654 | orchestrator | 2025-08-29 17:30:50 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:50.739449 | orchestrator | 2025-08-29 17:30:50 | INFO  | Task 7ef92b42-74d6-41df-8f89-170f3d45dd2c is in state SUCCESS 2025-08-29 17:30:50.740490 | orchestrator | 2025-08-29 17:30:50.740520 | orchestrator | 2025-08-29 17:30:50.740533 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:30:50.740546 | orchestrator | 2025-08-29 17:30:50.740557 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:30:50.740592 | orchestrator | Friday 29 August 2025 17:30:23 +0000 (0:00:00.322) 0:00:00.322 ********* 2025-08-29 17:30:50.740606 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:30:50.740618 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:30:50.740629 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:30:50.740640 | orchestrator | 2025-08-29 17:30:50.740651 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:30:50.740662 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.450) 0:00:00.772 ********* 2025-08-29 17:30:50.740673 | orchestrator | ok: [testbed-node-0] => (item=enable_memcached_True) 2025-08-29 17:30:50.740685 | orchestrator | ok: [testbed-node-1] => (item=enable_memcached_True) 2025-08-29 17:30:50.740696 | orchestrator | ok: [testbed-node-2] => (item=enable_memcached_True) 2025-08-29 17:30:50.740707 | orchestrator | 2025-08-29 17:30:50.740718 | orchestrator | PLAY [Apply role memcached] **************************************************** 2025-08-29 17:30:50.740729 | orchestrator | 2025-08-29 17:30:50.740740 | orchestrator | TASK [memcached : include_tasks] *********************************************** 2025-08-29 17:30:50.740751 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.549) 0:00:01.322 ********* 2025-08-29 17:30:50.740762 | orchestrator | included: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:30:50.740773 | orchestrator | 2025-08-29 17:30:50.740784 | orchestrator | TASK [memcached : Ensuring config directories exist] *************************** 2025-08-29 17:30:50.740796 | orchestrator | Friday 29 August 2025 17:30:25 +0000 (0:00:00.468) 0:00:01.791 ********* 2025-08-29 17:30:50.740832 | orchestrator | changed: [testbed-node-0] => (item=memcached) 2025-08-29 17:30:50.740844 | orchestrator | changed: [testbed-node-1] => (item=memcached) 2025-08-29 17:30:50.740855 | orchestrator | changed: [testbed-node-2] => (item=memcached) 2025-08-29 17:30:50.740866 | orchestrator | 2025-08-29 17:30:50.740877 | orchestrator | TASK [memcached : Copying over config.json files for services] ***************** 2025-08-29 17:30:50.740888 | orchestrator | Friday 29 August 2025 17:30:26 +0000 (0:00:00.842) 0:00:02.633 ********* 2025-08-29 17:30:50.740898 | orchestrator | changed: [testbed-node-1] => (item=memcached) 2025-08-29 17:30:50.740910 | orchestrator | changed: [testbed-node-2] => (item=memcached) 2025-08-29 17:30:50.740920 | orchestrator | changed: [testbed-node-0] => (item=memcached) 2025-08-29 17:30:50.740931 | orchestrator | 2025-08-29 17:30:50.740942 | orchestrator | TASK [memcached : Check memcached container] *********************************** 2025-08-29 17:30:50.740953 | orchestrator | Friday 29 August 2025 17:30:28 +0000 (0:00:02.091) 0:00:04.725 ********* 2025-08-29 17:30:50.740964 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:50.740976 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:50.740987 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:50.740997 | orchestrator | 2025-08-29 17:30:50.741008 | orchestrator | RUNNING HANDLER [memcached : Restart memcached container] ********************** 2025-08-29 17:30:50.741019 | orchestrator | Friday 29 August 2025 17:30:30 +0000 (0:00:02.722) 0:00:07.450 ********* 2025-08-29 17:30:50.741030 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:50.741041 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:50.741052 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:50.741063 | orchestrator | 2025-08-29 17:30:50.741076 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:30:50.741089 | orchestrator | testbed-node-0 : ok=7  changed=4  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:30:50.741102 | orchestrator | testbed-node-1 : ok=7  changed=4  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:30:50.741129 | orchestrator | testbed-node-2 : ok=7  changed=4  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:30:50.741143 | orchestrator | 2025-08-29 17:30:50.741155 | orchestrator | 2025-08-29 17:30:50.741168 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:30:50.741180 | orchestrator | Friday 29 August 2025 17:30:33 +0000 (0:00:02.676) 0:00:10.126 ********* 2025-08-29 17:30:50.741193 | orchestrator | =============================================================================== 2025-08-29 17:30:50.741205 | orchestrator | memcached : Check memcached container ----------------------------------- 2.73s 2025-08-29 17:30:50.741217 | orchestrator | memcached : Restart memcached container --------------------------------- 2.68s 2025-08-29 17:30:50.741230 | orchestrator | memcached : Copying over config.json files for services ----------------- 2.09s 2025-08-29 17:30:50.741243 | orchestrator | memcached : Ensuring config directories exist --------------------------- 0.84s 2025-08-29 17:30:50.741255 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.55s 2025-08-29 17:30:50.741267 | orchestrator | memcached : include_tasks ----------------------------------------------- 0.47s 2025-08-29 17:30:50.741280 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.45s 2025-08-29 17:30:50.741293 | orchestrator | 2025-08-29 17:30:50.741306 | orchestrator | 2025-08-29 17:30:50.741318 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:30:50.741331 | orchestrator | 2025-08-29 17:30:50.741359 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:30:50.741371 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.418) 0:00:00.418 ********* 2025-08-29 17:30:50.741384 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:30:50.741396 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:30:50.741415 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:30:50.741426 | orchestrator | 2025-08-29 17:30:50.741437 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:30:50.741459 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.338) 0:00:00.757 ********* 2025-08-29 17:30:50.741470 | orchestrator | ok: [testbed-node-0] => (item=enable_redis_True) 2025-08-29 17:30:50.741481 | orchestrator | ok: [testbed-node-1] => (item=enable_redis_True) 2025-08-29 17:30:50.741492 | orchestrator | ok: [testbed-node-2] => (item=enable_redis_True) 2025-08-29 17:30:50.741503 | orchestrator | 2025-08-29 17:30:50.741514 | orchestrator | PLAY [Apply role redis] ******************************************************** 2025-08-29 17:30:50.741525 | orchestrator | 2025-08-29 17:30:50.741536 | orchestrator | TASK [redis : include_tasks] *************************************************** 2025-08-29 17:30:50.741547 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.504) 0:00:01.262 ********* 2025-08-29 17:30:50.741557 | orchestrator | included: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:30:50.741568 | orchestrator | 2025-08-29 17:30:50.741579 | orchestrator | TASK [redis : Ensuring config directories exist] ******************************* 2025-08-29 17:30:50.741590 | orchestrator | Friday 29 August 2025 17:30:25 +0000 (0:00:00.726) 0:00:01.988 ********* 2025-08-29 17:30:50.741603 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741620 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741637 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741650 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741662 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741687 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741700 | orchestrator | 2025-08-29 17:30:50.741711 | orchestrator | TASK [redis : Copying over default config.json files] ************************** 2025-08-29 17:30:50.741722 | orchestrator | Friday 29 August 2025 17:30:27 +0000 (0:00:01.473) 0:00:03.461 ********* 2025-08-29 17:30:50.741734 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741746 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741762 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741773 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741785 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741809 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741821 | orchestrator | 2025-08-29 17:30:50.741832 | orchestrator | TASK [redis : Copying over redis config files] ********************************* 2025-08-29 17:30:50.741844 | orchestrator | Friday 29 August 2025 17:30:30 +0000 (0:00:03.040) 0:00:06.502 ********* 2025-08-29 17:30:50.741855 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741867 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741882 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741894 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741912 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741930 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741942 | orchestrator | 2025-08-29 17:30:50.741953 | orchestrator | TASK [redis : Check redis containers] ****************************************** 2025-08-29 17:30:50.741964 | orchestrator | Friday 29 August 2025 17:30:32 +0000 (0:00:02.664) 0:00:09.167 ********* 2025-08-29 17:30:50.741976 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.741987 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.742003 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis:7.0.15.20250711', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.742057 | orchestrator | changed: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.742079 | orchestrator | changed: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.742098 | orchestrator | changed: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}}) 2025-08-29 17:30:50.742115 | orchestrator | 2025-08-29 17:30:50.742135 | orchestrator | TASK [redis : Flush handlers] ************************************************** 2025-08-29 17:30:50.742153 | orchestrator | Friday 29 August 2025 17:30:34 +0000 (0:00:01.898) 0:00:11.065 ********* 2025-08-29 17:30:50.742171 | orchestrator | 2025-08-29 17:30:50.742189 | orchestrator | TASK [redis : Flush handlers] ************************************************** 2025-08-29 17:30:50.742207 | orchestrator | Friday 29 August 2025 17:30:34 +0000 (0:00:00.200) 0:00:11.265 ********* 2025-08-29 17:30:50.742226 | orchestrator | 2025-08-29 17:30:50.742248 | orchestrator | TASK [redis : Flush handlers] ************************************************** 2025-08-29 17:30:50.742271 | orchestrator | Friday 29 August 2025 17:30:35 +0000 (0:00:00.275) 0:00:11.540 ********* 2025-08-29 17:30:50.742291 | orchestrator | 2025-08-29 17:30:50.742312 | orchestrator | RUNNING HANDLER [redis : Restart redis container] ****************************** 2025-08-29 17:30:50.742330 | orchestrator | Friday 29 August 2025 17:30:35 +0000 (0:00:00.354) 0:00:11.894 ********* 2025-08-29 17:30:50.742384 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:50.742406 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:50.742422 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:50.742433 | orchestrator | 2025-08-29 17:30:50.742444 | orchestrator | RUNNING HANDLER [redis : Restart redis-sentinel container] ********************* 2025-08-29 17:30:50.742455 | orchestrator | Friday 29 August 2025 17:30:40 +0000 (0:00:04.893) 0:00:16.788 ********* 2025-08-29 17:30:50.742465 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:30:50.742476 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:30:50.742486 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:30:50.742497 | orchestrator | 2025-08-29 17:30:50.742507 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:30:50.742518 | orchestrator | testbed-node-0 : ok=9  changed=6  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:30:50.742529 | orchestrator | testbed-node-1 : ok=9  changed=6  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:30:50.742540 | orchestrator | testbed-node-2 : ok=9  changed=6  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:30:50.742550 | orchestrator | 2025-08-29 17:30:50.742570 | orchestrator | 2025-08-29 17:30:50.742581 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:30:50.742592 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:07.830) 0:00:24.618 ********* 2025-08-29 17:30:50.742608 | orchestrator | =============================================================================== 2025-08-29 17:30:50.742619 | orchestrator | redis : Restart redis-sentinel container -------------------------------- 7.83s 2025-08-29 17:30:50.742630 | orchestrator | redis : Restart redis container ----------------------------------------- 4.89s 2025-08-29 17:30:50.742640 | orchestrator | redis : Copying over default config.json files -------------------------- 3.04s 2025-08-29 17:30:50.742651 | orchestrator | redis : Copying over redis config files --------------------------------- 2.66s 2025-08-29 17:30:50.742662 | orchestrator | redis : Check redis containers ------------------------------------------ 1.90s 2025-08-29 17:30:50.742672 | orchestrator | redis : Ensuring config directories exist ------------------------------- 1.47s 2025-08-29 17:30:50.742683 | orchestrator | redis : Flush handlers -------------------------------------------------- 0.83s 2025-08-29 17:30:50.742693 | orchestrator | redis : include_tasks --------------------------------------------------- 0.73s 2025-08-29 17:30:50.742704 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.50s 2025-08-29 17:30:50.742715 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.34s 2025-08-29 17:30:50.742732 | orchestrator | 2025-08-29 17:30:50 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:50.742743 | orchestrator | 2025-08-29 17:30:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:53.763695 | orchestrator | 2025-08-29 17:30:53 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:53.764893 | orchestrator | 2025-08-29 17:30:53 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:53.765299 | orchestrator | 2025-08-29 17:30:53 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:53.766849 | orchestrator | 2025-08-29 17:30:53 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:53.767987 | orchestrator | 2025-08-29 17:30:53 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:53.768008 | orchestrator | 2025-08-29 17:30:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:56.799849 | orchestrator | 2025-08-29 17:30:56 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:56.800191 | orchestrator | 2025-08-29 17:30:56 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:56.801155 | orchestrator | 2025-08-29 17:30:56 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:56.801677 | orchestrator | 2025-08-29 17:30:56 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:56.803188 | orchestrator | 2025-08-29 17:30:56 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:56.803214 | orchestrator | 2025-08-29 17:30:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:30:59.828671 | orchestrator | 2025-08-29 17:30:59 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:30:59.828761 | orchestrator | 2025-08-29 17:30:59 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:30:59.829111 | orchestrator | 2025-08-29 17:30:59 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:30:59.829682 | orchestrator | 2025-08-29 17:30:59 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:30:59.830261 | orchestrator | 2025-08-29 17:30:59 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:30:59.830305 | orchestrator | 2025-08-29 17:30:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:02.933305 | orchestrator | 2025-08-29 17:31:02 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:02.933741 | orchestrator | 2025-08-29 17:31:02 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:31:02.934144 | orchestrator | 2025-08-29 17:31:02 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:02.934662 | orchestrator | 2025-08-29 17:31:02 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:02.935297 | orchestrator | 2025-08-29 17:31:02 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:02.935318 | orchestrator | 2025-08-29 17:31:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:06.032069 | orchestrator | 2025-08-29 17:31:06 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:06.032162 | orchestrator | 2025-08-29 17:31:06 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:31:06.032193 | orchestrator | 2025-08-29 17:31:06 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:06.032205 | orchestrator | 2025-08-29 17:31:06 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:06.032217 | orchestrator | 2025-08-29 17:31:06 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:06.032228 | orchestrator | 2025-08-29 17:31:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:09.236595 | orchestrator | 2025-08-29 17:31:09 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:09.236685 | orchestrator | 2025-08-29 17:31:09 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:31:09.236700 | orchestrator | 2025-08-29 17:31:09 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:09.236711 | orchestrator | 2025-08-29 17:31:09 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:09.236722 | orchestrator | 2025-08-29 17:31:09 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:09.236733 | orchestrator | 2025-08-29 17:31:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:12.309093 | orchestrator | 2025-08-29 17:31:12 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:12.309179 | orchestrator | 2025-08-29 17:31:12 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state STARTED 2025-08-29 17:31:12.309194 | orchestrator | 2025-08-29 17:31:12 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:12.309205 | orchestrator | 2025-08-29 17:31:12 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:12.309217 | orchestrator | 2025-08-29 17:31:12 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:12.309289 | orchestrator | 2025-08-29 17:31:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:15.425397 | orchestrator | 2025-08-29 17:31:15 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:15.426176 | orchestrator | 2025-08-29 17:31:15 | INFO  | Task dd96029d-f6d4-4f19-b191-93ba27dbc4fd is in state SUCCESS 2025-08-29 17:31:15.427819 | orchestrator | 2025-08-29 17:31:15.427847 | orchestrator | 2025-08-29 17:31:15.427899 | orchestrator | PLAY [Prepare all k3s nodes] *************************************************** 2025-08-29 17:31:15.427909 | orchestrator | 2025-08-29 17:31:15.427916 | orchestrator | TASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] *** 2025-08-29 17:31:15.427927 | orchestrator | Friday 29 August 2025 17:27:41 +0000 (0:00:00.202) 0:00:00.202 ********* 2025-08-29 17:31:15.427935 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.427944 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.427951 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.427958 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.427965 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.427972 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.427979 | orchestrator | 2025-08-29 17:31:15.427987 | orchestrator | TASK [k3s_prereq : Set same timezone on every Server] ************************** 2025-08-29 17:31:15.427994 | orchestrator | Friday 29 August 2025 17:27:41 +0000 (0:00:00.629) 0:00:00.832 ********* 2025-08-29 17:31:15.428001 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428009 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428016 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428023 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428030 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428037 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428045 | orchestrator | 2025-08-29 17:31:15.428052 | orchestrator | TASK [k3s_prereq : Set SELinux to disabled state] ****************************** 2025-08-29 17:31:15.428059 | orchestrator | Friday 29 August 2025 17:27:42 +0000 (0:00:00.532) 0:00:01.365 ********* 2025-08-29 17:31:15.428066 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428073 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428080 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428087 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428094 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428101 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428108 | orchestrator | 2025-08-29 17:31:15.428115 | orchestrator | TASK [k3s_prereq : Enable IPv4 forwarding] ************************************* 2025-08-29 17:31:15.428122 | orchestrator | Friday 29 August 2025 17:27:42 +0000 (0:00:00.642) 0:00:02.007 ********* 2025-08-29 17:31:15.428129 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.428136 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.428143 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.428150 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.428157 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.428165 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.428172 | orchestrator | 2025-08-29 17:31:15.428179 | orchestrator | TASK [k3s_prereq : Enable IPv6 forwarding] ************************************* 2025-08-29 17:31:15.428187 | orchestrator | Friday 29 August 2025 17:27:44 +0000 (0:00:01.993) 0:00:04.001 ********* 2025-08-29 17:31:15.428194 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.428201 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.428208 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.428215 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.428222 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.428229 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.428236 | orchestrator | 2025-08-29 17:31:15.428248 | orchestrator | TASK [k3s_prereq : Enable IPv6 router advertisements] ************************** 2025-08-29 17:31:15.428255 | orchestrator | Friday 29 August 2025 17:27:46 +0000 (0:00:01.083) 0:00:05.084 ********* 2025-08-29 17:31:15.428262 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.428270 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.428277 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.428284 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.428291 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.428298 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.428305 | orchestrator | 2025-08-29 17:31:15.428312 | orchestrator | TASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] ******************* 2025-08-29 17:31:15.428324 | orchestrator | Friday 29 August 2025 17:27:47 +0000 (0:00:00.990) 0:00:06.074 ********* 2025-08-29 17:31:15.428331 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428338 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428345 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428381 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428388 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428395 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428402 | orchestrator | 2025-08-29 17:31:15.428409 | orchestrator | TASK [k3s_prereq : Load br_netfilter] ****************************************** 2025-08-29 17:31:15.428417 | orchestrator | Friday 29 August 2025 17:27:47 +0000 (0:00:00.809) 0:00:06.884 ********* 2025-08-29 17:31:15.428425 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428434 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428442 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428450 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428458 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428466 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428474 | orchestrator | 2025-08-29 17:31:15.428482 | orchestrator | TASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] ************** 2025-08-29 17:31:15.428490 | orchestrator | Friday 29 August 2025 17:27:48 +0000 (0:00:00.836) 0:00:07.720 ********* 2025-08-29 17:31:15.428499 | orchestrator | skipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:31:15.428507 | orchestrator | skipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:31:15.428515 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428523 | orchestrator | skipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:31:15.428531 | orchestrator | skipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:31:15.428539 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428547 | orchestrator | skipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:31:15.428556 | orchestrator | skipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:31:15.428564 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428572 | orchestrator | skipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:31:15.428589 | orchestrator | skipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:31:15.428598 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428606 | orchestrator | skipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:31:15.428614 | orchestrator | skipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:31:15.428622 | orchestrator | skipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:31:15.428630 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428638 | orchestrator | skipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:31:15.428646 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428654 | orchestrator | 2025-08-29 17:31:15.428662 | orchestrator | TASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] ********************* 2025-08-29 17:31:15.428670 | orchestrator | Friday 29 August 2025 17:27:49 +0000 (0:00:00.905) 0:00:08.625 ********* 2025-08-29 17:31:15.428678 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428686 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428695 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428703 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428711 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428719 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428727 | orchestrator | 2025-08-29 17:31:15.428734 | orchestrator | TASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] *** 2025-08-29 17:31:15.428742 | orchestrator | Friday 29 August 2025 17:27:50 +0000 (0:00:01.110) 0:00:09.736 ********* 2025-08-29 17:31:15.428754 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.428762 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.428769 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.428776 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.428783 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.428790 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.428797 | orchestrator | 2025-08-29 17:31:15.428804 | orchestrator | TASK [k3s_download : Download k3s binary x64] ********************************** 2025-08-29 17:31:15.428812 | orchestrator | Friday 29 August 2025 17:27:51 +0000 (0:00:00.949) 0:00:10.686 ********* 2025-08-29 17:31:15.428819 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.428826 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.428833 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.428840 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.428847 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.428854 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.428862 | orchestrator | 2025-08-29 17:31:15.428869 | orchestrator | TASK [k3s_download : Download k3s binary arm64] ******************************** 2025-08-29 17:31:15.428876 | orchestrator | Friday 29 August 2025 17:27:57 +0000 (0:00:05.981) 0:00:16.667 ********* 2025-08-29 17:31:15.428883 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428890 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428897 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428904 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428912 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428919 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428926 | orchestrator | 2025-08-29 17:31:15.428936 | orchestrator | TASK [k3s_download : Download k3s binary armhf] ******************************** 2025-08-29 17:31:15.428944 | orchestrator | Friday 29 August 2025 17:27:59 +0000 (0:00:01.631) 0:00:18.298 ********* 2025-08-29 17:31:15.428951 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.428958 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.428965 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.428972 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.428979 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.428987 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.428994 | orchestrator | 2025-08-29 17:31:15.429001 | orchestrator | TASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] *** 2025-08-29 17:31:15.429010 | orchestrator | Friday 29 August 2025 17:28:01 +0000 (0:00:02.372) 0:00:20.671 ********* 2025-08-29 17:31:15.429017 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.429024 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.429031 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.429038 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.429045 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.429052 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.429059 | orchestrator | 2025-08-29 17:31:15.429067 | orchestrator | TASK [k3s_custom_registries : Create directory /etc/rancher/k3s] *************** 2025-08-29 17:31:15.429074 | orchestrator | Friday 29 August 2025 17:28:02 +0000 (0:00:01.117) 0:00:21.788 ********* 2025-08-29 17:31:15.429081 | orchestrator | changed: [testbed-node-3] => (item=rancher) 2025-08-29 17:31:15.429088 | orchestrator | changed: [testbed-node-4] => (item=rancher) 2025-08-29 17:31:15.429096 | orchestrator | changed: [testbed-node-3] => (item=rancher/k3s) 2025-08-29 17:31:15.429103 | orchestrator | changed: [testbed-node-5] => (item=rancher) 2025-08-29 17:31:15.429110 | orchestrator | changed: [testbed-node-4] => (item=rancher/k3s) 2025-08-29 17:31:15.429117 | orchestrator | changed: [testbed-node-1] => (item=rancher) 2025-08-29 17:31:15.429124 | orchestrator | changed: [testbed-node-0] => (item=rancher) 2025-08-29 17:31:15.429131 | orchestrator | changed: [testbed-node-5] => (item=rancher/k3s) 2025-08-29 17:31:15.429138 | orchestrator | changed: [testbed-node-2] => (item=rancher) 2025-08-29 17:31:15.429146 | orchestrator | changed: [testbed-node-1] => (item=rancher/k3s) 2025-08-29 17:31:15.429157 | orchestrator | changed: [testbed-node-0] => (item=rancher/k3s) 2025-08-29 17:31:15.429164 | orchestrator | changed: [testbed-node-2] => (item=rancher/k3s) 2025-08-29 17:31:15.429172 | orchestrator | 2025-08-29 17:31:15.429179 | orchestrator | TASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] *** 2025-08-29 17:31:15.429186 | orchestrator | Friday 29 August 2025 17:28:05 +0000 (0:00:02.652) 0:00:24.441 ********* 2025-08-29 17:31:15.429193 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.429200 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.429208 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.429215 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.429222 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.429229 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.429236 | orchestrator | 2025-08-29 17:31:15.429247 | orchestrator | PLAY [Deploy k3s master nodes] ************************************************* 2025-08-29 17:31:15.429255 | orchestrator | 2025-08-29 17:31:15.429262 | orchestrator | TASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] *** 2025-08-29 17:31:15.429269 | orchestrator | Friday 29 August 2025 17:28:07 +0000 (0:00:01.931) 0:00:26.373 ********* 2025-08-29 17:31:15.429276 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.429284 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.429291 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.429298 | orchestrator | 2025-08-29 17:31:15.429305 | orchestrator | TASK [k3s_server : Stop k3s-init] ********************************************** 2025-08-29 17:31:15.429312 | orchestrator | Friday 29 August 2025 17:28:08 +0000 (0:00:01.162) 0:00:27.535 ********* 2025-08-29 17:31:15.429319 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.429327 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.429334 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.429341 | orchestrator | 2025-08-29 17:31:15.429361 | orchestrator | TASK [k3s_server : Stop k3s] *************************************************** 2025-08-29 17:31:15.429369 | orchestrator | Friday 29 August 2025 17:28:10 +0000 (0:00:01.608) 0:00:29.144 ********* 2025-08-29 17:31:15.429376 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.429383 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.429390 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.429398 | orchestrator | 2025-08-29 17:31:15.429405 | orchestrator | TASK [k3s_server : Clean previous runs of k3s-init] **************************** 2025-08-29 17:31:15.429412 | orchestrator | Friday 29 August 2025 17:28:11 +0000 (0:00:01.359) 0:00:30.503 ********* 2025-08-29 17:31:15.429419 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.429426 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.429433 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.429441 | orchestrator | 2025-08-29 17:31:15.429448 | orchestrator | TASK [k3s_server : Deploy K3s http_proxy conf] ********************************* 2025-08-29 17:31:15.429455 | orchestrator | Friday 29 August 2025 17:28:12 +0000 (0:00:00.999) 0:00:31.502 ********* 2025-08-29 17:31:15.429462 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.429470 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.429477 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.429484 | orchestrator | 2025-08-29 17:31:15.429491 | orchestrator | TASK [k3s_server : Create /etc/rancher/k3s directory] ************************** 2025-08-29 17:31:15.429498 | orchestrator | Friday 29 August 2025 17:28:12 +0000 (0:00:00.475) 0:00:31.978 ********* 2025-08-29 17:31:15.429506 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.429513 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.429520 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.429527 | orchestrator | 2025-08-29 17:31:15.429535 | orchestrator | TASK [k3s_server : Create custom resolv.conf for k3s] ************************** 2025-08-29 17:31:15.429542 | orchestrator | Friday 29 August 2025 17:28:13 +0000 (0:00:00.729) 0:00:32.708 ********* 2025-08-29 17:31:15.429549 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.429556 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.429563 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.429570 | orchestrator | 2025-08-29 17:31:15.429582 | orchestrator | TASK [k3s_server : Deploy vip manifest] **************************************** 2025-08-29 17:31:15.429592 | orchestrator | Friday 29 August 2025 17:28:15 +0000 (0:00:01.415) 0:00:34.123 ********* 2025-08-29 17:31:15.429600 | orchestrator | included: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:31:15.429607 | orchestrator | 2025-08-29 17:31:15.429614 | orchestrator | TASK [k3s_server : Set _kube_vip_bgp_peers fact] ******************************* 2025-08-29 17:31:15.429621 | orchestrator | Friday 29 August 2025 17:28:15 +0000 (0:00:00.690) 0:00:34.814 ********* 2025-08-29 17:31:15.429628 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.429636 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.429643 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.429650 | orchestrator | 2025-08-29 17:31:15.429657 | orchestrator | TASK [k3s_server : Create manifests directory on first master] ***************** 2025-08-29 17:31:15.429665 | orchestrator | Friday 29 August 2025 17:28:16 +0000 (0:00:01.159) 0:00:35.973 ********* 2025-08-29 17:31:15.429672 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.429679 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.429686 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.429693 | orchestrator | 2025-08-29 17:31:15.429700 | orchestrator | TASK [k3s_server : Download vip rbac manifest to first master] ***************** 2025-08-29 17:31:15.429707 | orchestrator | Friday 29 August 2025 17:28:17 +0000 (0:00:00.836) 0:00:36.809 ********* 2025-08-29 17:31:15.429714 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.429722 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.429729 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.429736 | orchestrator | 2025-08-29 17:31:15.429743 | orchestrator | TASK [k3s_server : Copy vip manifest to first master] ************************** 2025-08-29 17:31:15.429750 | orchestrator | Friday 29 August 2025 17:28:18 +0000 (0:00:00.806) 0:00:37.615 ********* 2025-08-29 17:31:15.429758 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.429765 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.429772 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.429779 | orchestrator | 2025-08-29 17:31:15.429786 | orchestrator | TASK [k3s_server : Deploy metallb manifest] ************************************ 2025-08-29 17:31:15.429793 | orchestrator | Friday 29 August 2025 17:28:20 +0000 (0:00:01.877) 0:00:39.493 ********* 2025-08-29 17:31:15.429800 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.429816 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.429830 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.429838 | orchestrator | 2025-08-29 17:31:15.429845 | orchestrator | TASK [k3s_server : Deploy kube-vip manifest] *********************************** 2025-08-29 17:31:15.429852 | orchestrator | Friday 29 August 2025 17:28:21 +0000 (0:00:00.576) 0:00:40.069 ********* 2025-08-29 17:31:15.429859 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.429866 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.429874 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.429881 | orchestrator | 2025-08-29 17:31:15.429888 | orchestrator | TASK [k3s_server : Init cluster inside the transient k3s-init service] ********* 2025-08-29 17:31:15.429895 | orchestrator | Friday 29 August 2025 17:28:21 +0000 (0:00:00.555) 0:00:40.625 ********* 2025-08-29 17:31:15.429927 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.429935 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.429942 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.429949 | orchestrator | 2025-08-29 17:31:15.429961 | orchestrator | TASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] *** 2025-08-29 17:31:15.429969 | orchestrator | Friday 29 August 2025 17:28:24 +0000 (0:00:02.518) 0:00:43.144 ********* 2025-08-29 17:31:15.429976 | orchestrator | FAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left). 2025-08-29 17:31:15.429984 | orchestrator | FAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left). 2025-08-29 17:31:15.429996 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left). 2025-08-29 17:31:15.430003 | orchestrator | FAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left). 2025-08-29 17:31:15.430010 | orchestrator | FAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left). 2025-08-29 17:31:15.430057 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left). 2025-08-29 17:31:15.430065 | orchestrator | FAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left). 2025-08-29 17:31:15.430073 | orchestrator | FAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left). 2025-08-29 17:31:15.430080 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left). 2025-08-29 17:31:15.430087 | orchestrator | FAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left). 2025-08-29 17:31:15.430095 | orchestrator | FAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left). 2025-08-29 17:31:15.430102 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left). 2025-08-29 17:31:15.430109 | orchestrator | FAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left). 2025-08-29 17:31:15.430116 | orchestrator | FAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left). 2025-08-29 17:31:15.430124 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left). 2025-08-29 17:31:15.430131 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.430139 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.430146 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.430153 | orchestrator | 2025-08-29 17:31:15.430160 | orchestrator | TASK [k3s_server : Save logs of k3s-init.service] ****************************** 2025-08-29 17:31:15.430168 | orchestrator | Friday 29 August 2025 17:29:19 +0000 (0:00:55.728) 0:01:38.872 ********* 2025-08-29 17:31:15.430175 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.430182 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.430190 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.430197 | orchestrator | 2025-08-29 17:31:15.430204 | orchestrator | TASK [k3s_server : Kill the temporary service used for initialization] ********* 2025-08-29 17:31:15.430217 | orchestrator | Friday 29 August 2025 17:29:20 +0000 (0:00:00.270) 0:01:39.142 ********* 2025-08-29 17:31:15.430224 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430232 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430239 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430246 | orchestrator | 2025-08-29 17:31:15.430253 | orchestrator | TASK [k3s_server : Copy K3s service file] ************************************** 2025-08-29 17:31:15.430261 | orchestrator | Friday 29 August 2025 17:29:21 +0000 (0:00:01.273) 0:01:40.416 ********* 2025-08-29 17:31:15.430268 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430275 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430282 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430290 | orchestrator | 2025-08-29 17:31:15.430297 | orchestrator | TASK [k3s_server : Enable and check K3s service] ******************************* 2025-08-29 17:31:15.430304 | orchestrator | Friday 29 August 2025 17:29:22 +0000 (0:00:01.288) 0:01:41.705 ********* 2025-08-29 17:31:15.430311 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430323 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430330 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430337 | orchestrator | 2025-08-29 17:31:15.430345 | orchestrator | TASK [k3s_server : Wait for node-token] **************************************** 2025-08-29 17:31:15.430389 | orchestrator | Friday 29 August 2025 17:29:47 +0000 (0:00:24.591) 0:02:06.296 ********* 2025-08-29 17:31:15.430397 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.430404 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.430411 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.430418 | orchestrator | 2025-08-29 17:31:15.430425 | orchestrator | TASK [k3s_server : Register node-token file access mode] *********************** 2025-08-29 17:31:15.430433 | orchestrator | Friday 29 August 2025 17:29:47 +0000 (0:00:00.704) 0:02:07.000 ********* 2025-08-29 17:31:15.430440 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.430447 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.430454 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.430461 | orchestrator | 2025-08-29 17:31:15.430474 | orchestrator | TASK [k3s_server : Change file access node-token] ****************************** 2025-08-29 17:31:15.430481 | orchestrator | Friday 29 August 2025 17:29:48 +0000 (0:00:00.973) 0:02:07.974 ********* 2025-08-29 17:31:15.430489 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430496 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430503 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430510 | orchestrator | 2025-08-29 17:31:15.430517 | orchestrator | TASK [k3s_server : Read node-token from master] ******************************** 2025-08-29 17:31:15.430525 | orchestrator | Friday 29 August 2025 17:29:49 +0000 (0:00:00.732) 0:02:08.706 ********* 2025-08-29 17:31:15.430532 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.430539 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.430546 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.430554 | orchestrator | 2025-08-29 17:31:15.430561 | orchestrator | TASK [k3s_server : Store Master node-token] ************************************ 2025-08-29 17:31:15.430568 | orchestrator | Friday 29 August 2025 17:29:50 +0000 (0:00:00.719) 0:02:09.425 ********* 2025-08-29 17:31:15.430575 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.430583 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.430590 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.430597 | orchestrator | 2025-08-29 17:31:15.430604 | orchestrator | TASK [k3s_server : Restore node-token file access] ***************************** 2025-08-29 17:31:15.430611 | orchestrator | Friday 29 August 2025 17:29:50 +0000 (0:00:00.303) 0:02:09.728 ********* 2025-08-29 17:31:15.430619 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430626 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430633 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430640 | orchestrator | 2025-08-29 17:31:15.430648 | orchestrator | TASK [k3s_server : Create directory .kube] ************************************* 2025-08-29 17:31:15.430655 | orchestrator | Friday 29 August 2025 17:29:51 +0000 (0:00:00.954) 0:02:10.683 ********* 2025-08-29 17:31:15.430662 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430669 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430677 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430684 | orchestrator | 2025-08-29 17:31:15.430691 | orchestrator | TASK [k3s_server : Copy config file to user home directory] ******************** 2025-08-29 17:31:15.430698 | orchestrator | Friday 29 August 2025 17:29:52 +0000 (0:00:00.684) 0:02:11.367 ********* 2025-08-29 17:31:15.430706 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430713 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430720 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430727 | orchestrator | 2025-08-29 17:31:15.430734 | orchestrator | TASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] ***** 2025-08-29 17:31:15.430741 | orchestrator | Friday 29 August 2025 17:29:53 +0000 (0:00:00.887) 0:02:12.255 ********* 2025-08-29 17:31:15.430748 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:15.430756 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:15.430763 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:15.430777 | orchestrator | 2025-08-29 17:31:15.430784 | orchestrator | TASK [k3s_server : Create kubectl symlink] ************************************* 2025-08-29 17:31:15.430791 | orchestrator | Friday 29 August 2025 17:29:54 +0000 (0:00:00.843) 0:02:13.099 ********* 2025-08-29 17:31:15.430799 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.430809 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.430816 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.430824 | orchestrator | 2025-08-29 17:31:15.430831 | orchestrator | TASK [k3s_server : Create crictl symlink] ************************************** 2025-08-29 17:31:15.430838 | orchestrator | Friday 29 August 2025 17:29:54 +0000 (0:00:00.533) 0:02:13.633 ********* 2025-08-29 17:31:15.430845 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.430852 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.430860 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.430867 | orchestrator | 2025-08-29 17:31:15.430874 | orchestrator | TASK [k3s_server : Get contents of manifests folder] *************************** 2025-08-29 17:31:15.430881 | orchestrator | Friday 29 August 2025 17:29:54 +0000 (0:00:00.290) 0:02:13.923 ********* 2025-08-29 17:31:15.430889 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.430896 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.430903 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.430910 | orchestrator | 2025-08-29 17:31:15.430918 | orchestrator | TASK [k3s_server : Get sub dirs of manifests folder] *************************** 2025-08-29 17:31:15.430925 | orchestrator | Friday 29 August 2025 17:29:55 +0000 (0:00:00.652) 0:02:14.575 ********* 2025-08-29 17:31:15.430932 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.430940 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.430947 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.430954 | orchestrator | 2025-08-29 17:31:15.430961 | orchestrator | TASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] *** 2025-08-29 17:31:15.430969 | orchestrator | Friday 29 August 2025 17:29:56 +0000 (0:00:00.662) 0:02:15.238 ********* 2025-08-29 17:31:15.430976 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml) 2025-08-29 17:31:15.430983 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml) 2025-08-29 17:31:15.430991 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml) 2025-08-29 17:31:15.430998 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml) 2025-08-29 17:31:15.431005 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml) 2025-08-29 17:31:15.431012 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml) 2025-08-29 17:31:15.431020 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml) 2025-08-29 17:31:15.431027 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml) 2025-08-29 17:31:15.431034 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml) 2025-08-29 17:31:15.431045 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml) 2025-08-29 17:31:15.431053 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml) 2025-08-29 17:31:15.431060 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml) 2025-08-29 17:31:15.431068 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml) 2025-08-29 17:31:15.431075 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml) 2025-08-29 17:31:15.431082 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml) 2025-08-29 17:31:15.431089 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml) 2025-08-29 17:31:15.431100 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server) 2025-08-29 17:31:15.431107 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server) 2025-08-29 17:31:15.431115 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml) 2025-08-29 17:31:15.431122 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server) 2025-08-29 17:31:15.431129 | orchestrator | 2025-08-29 17:31:15.431136 | orchestrator | PLAY [Deploy k3s worker nodes] ************************************************* 2025-08-29 17:31:15.431144 | orchestrator | 2025-08-29 17:31:15.431151 | orchestrator | TASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] *** 2025-08-29 17:31:15.431158 | orchestrator | Friday 29 August 2025 17:29:59 +0000 (0:00:03.325) 0:02:18.563 ********* 2025-08-29 17:31:15.431165 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.431173 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.431180 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.431187 | orchestrator | 2025-08-29 17:31:15.431194 | orchestrator | TASK [k3s_agent : Check if system is PXE-booted] ******************************* 2025-08-29 17:31:15.431201 | orchestrator | Friday 29 August 2025 17:29:59 +0000 (0:00:00.313) 0:02:18.877 ********* 2025-08-29 17:31:15.431208 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.431216 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.431223 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.431230 | orchestrator | 2025-08-29 17:31:15.431237 | orchestrator | TASK [k3s_agent : Set fact for PXE-booted system] ****************************** 2025-08-29 17:31:15.431244 | orchestrator | Friday 29 August 2025 17:30:00 +0000 (0:00:00.616) 0:02:19.494 ********* 2025-08-29 17:31:15.431251 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.431258 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.431265 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.431273 | orchestrator | 2025-08-29 17:31:15.431280 | orchestrator | TASK [k3s_agent : Include http_proxy configuration tasks] ********************** 2025-08-29 17:31:15.431287 | orchestrator | Friday 29 August 2025 17:30:00 +0000 (0:00:00.475) 0:02:19.970 ********* 2025-08-29 17:31:15.431300 | orchestrator | included: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:31:15.431308 | orchestrator | 2025-08-29 17:31:15.431315 | orchestrator | TASK [k3s_agent : Create k3s-node.service.d directory] ************************* 2025-08-29 17:31:15.431322 | orchestrator | Friday 29 August 2025 17:30:01 +0000 (0:00:00.447) 0:02:20.417 ********* 2025-08-29 17:31:15.431329 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.431337 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.431344 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.431380 | orchestrator | 2025-08-29 17:31:15.431388 | orchestrator | TASK [k3s_agent : Copy K3s http_proxy conf file] ******************************* 2025-08-29 17:31:15.431395 | orchestrator | Friday 29 August 2025 17:30:01 +0000 (0:00:00.293) 0:02:20.711 ********* 2025-08-29 17:31:15.431402 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.431410 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.431417 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.431424 | orchestrator | 2025-08-29 17:31:15.431431 | orchestrator | TASK [k3s_agent : Deploy K3s http_proxy conf] ********************************** 2025-08-29 17:31:15.431438 | orchestrator | Friday 29 August 2025 17:30:02 +0000 (0:00:00.497) 0:02:21.209 ********* 2025-08-29 17:31:15.431446 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.431453 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.431460 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.431467 | orchestrator | 2025-08-29 17:31:15.431474 | orchestrator | TASK [k3s_agent : Create /etc/rancher/k3s directory] *************************** 2025-08-29 17:31:15.431482 | orchestrator | Friday 29 August 2025 17:30:02 +0000 (0:00:00.278) 0:02:21.488 ********* 2025-08-29 17:31:15.431489 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.431496 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.431508 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.431515 | orchestrator | 2025-08-29 17:31:15.431523 | orchestrator | TASK [k3s_agent : Create custom resolv.conf for k3s] *************************** 2025-08-29 17:31:15.431530 | orchestrator | Friday 29 August 2025 17:30:03 +0000 (0:00:00.623) 0:02:22.112 ********* 2025-08-29 17:31:15.431537 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.431545 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.431552 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.431559 | orchestrator | 2025-08-29 17:31:15.431566 | orchestrator | TASK [k3s_agent : Configure the k3s service] *********************************** 2025-08-29 17:31:15.431574 | orchestrator | Friday 29 August 2025 17:30:04 +0000 (0:00:01.022) 0:02:23.135 ********* 2025-08-29 17:31:15.431581 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.431588 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.431595 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.431602 | orchestrator | 2025-08-29 17:31:15.431609 | orchestrator | TASK [k3s_agent : Manage k3s service] ****************************************** 2025-08-29 17:31:15.431616 | orchestrator | Friday 29 August 2025 17:30:05 +0000 (0:00:01.321) 0:02:24.456 ********* 2025-08-29 17:31:15.431624 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:15.431631 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:15.431638 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:15.431645 | orchestrator | 2025-08-29 17:31:15.431656 | orchestrator | PLAY [Prepare kubeconfig file] ************************************************* 2025-08-29 17:31:15.431664 | orchestrator | 2025-08-29 17:31:15.431671 | orchestrator | TASK [Get home directory of operator user] ************************************* 2025-08-29 17:31:15.431678 | orchestrator | Friday 29 August 2025 17:30:17 +0000 (0:00:11.970) 0:02:36.427 ********* 2025-08-29 17:31:15.431685 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:15.431693 | orchestrator | 2025-08-29 17:31:15.431700 | orchestrator | TASK [Create .kube directory] ************************************************** 2025-08-29 17:31:15.431707 | orchestrator | Friday 29 August 2025 17:30:18 +0000 (0:00:00.770) 0:02:37.197 ********* 2025-08-29 17:31:15.431714 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.431722 | orchestrator | 2025-08-29 17:31:15.431729 | orchestrator | TASK [Get kubeconfig file] ***************************************************** 2025-08-29 17:31:15.431736 | orchestrator | Friday 29 August 2025 17:30:18 +0000 (0:00:00.436) 0:02:37.633 ********* 2025-08-29 17:31:15.431743 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] 2025-08-29 17:31:15.431750 | orchestrator | 2025-08-29 17:31:15.431757 | orchestrator | TASK [Write kubeconfig file] *************************************************** 2025-08-29 17:31:15.431765 | orchestrator | Friday 29 August 2025 17:30:19 +0000 (0:00:00.564) 0:02:38.198 ********* 2025-08-29 17:31:15.431772 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.431779 | orchestrator | 2025-08-29 17:31:15.431786 | orchestrator | TASK [Change server address in the kubeconfig] ********************************* 2025-08-29 17:31:15.431793 | orchestrator | Friday 29 August 2025 17:30:20 +0000 (0:00:00.920) 0:02:39.118 ********* 2025-08-29 17:31:15.431801 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.431808 | orchestrator | 2025-08-29 17:31:15.431815 | orchestrator | TASK [Make kubeconfig available for use inside the manager service] ************ 2025-08-29 17:31:15.431822 | orchestrator | Friday 29 August 2025 17:30:21 +0000 (0:00:01.055) 0:02:40.173 ********* 2025-08-29 17:31:15.431829 | orchestrator | changed: [testbed-manager -> localhost] 2025-08-29 17:31:15.431837 | orchestrator | 2025-08-29 17:31:15.431844 | orchestrator | TASK [Change server address in the kubeconfig inside the manager service] ****** 2025-08-29 17:31:15.431851 | orchestrator | Friday 29 August 2025 17:30:22 +0000 (0:00:01.571) 0:02:41.745 ********* 2025-08-29 17:31:15.431858 | orchestrator | changed: [testbed-manager -> localhost] 2025-08-29 17:31:15.431865 | orchestrator | 2025-08-29 17:31:15.431873 | orchestrator | TASK [Set KUBECONFIG environment variable] ************************************* 2025-08-29 17:31:15.431880 | orchestrator | Friday 29 August 2025 17:30:23 +0000 (0:00:00.754) 0:02:42.500 ********* 2025-08-29 17:31:15.431891 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.431898 | orchestrator | 2025-08-29 17:31:15.431905 | orchestrator | TASK [Enable kubectl command line completion] ********************************** 2025-08-29 17:31:15.431912 | orchestrator | Friday 29 August 2025 17:30:23 +0000 (0:00:00.342) 0:02:42.842 ********* 2025-08-29 17:31:15.431920 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.431927 | orchestrator | 2025-08-29 17:31:15.431934 | orchestrator | PLAY [Apply role kubectl] ****************************************************** 2025-08-29 17:31:15.431942 | orchestrator | 2025-08-29 17:31:15.431952 | orchestrator | TASK [kubectl : Gather variables for each operating system] ******************** 2025-08-29 17:31:15.431960 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.371) 0:02:43.214 ********* 2025-08-29 17:31:15.431967 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:15.431974 | orchestrator | 2025-08-29 17:31:15.431981 | orchestrator | TASK [kubectl : Include distribution specific install tasks] ******************* 2025-08-29 17:31:15.431989 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.128) 0:02:43.342 ********* 2025-08-29 17:31:15.431996 | orchestrator | included: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager 2025-08-29 17:31:15.432003 | orchestrator | 2025-08-29 17:31:15.432010 | orchestrator | TASK [kubectl : Remove old architecture-dependent repository] ****************** 2025-08-29 17:31:15.432017 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.199) 0:02:43.541 ********* 2025-08-29 17:31:15.432024 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:15.432032 | orchestrator | 2025-08-29 17:31:15.432039 | orchestrator | TASK [kubectl : Install apt-transport-https package] *************************** 2025-08-29 17:31:15.432046 | orchestrator | Friday 29 August 2025 17:30:25 +0000 (0:00:00.873) 0:02:44.415 ********* 2025-08-29 17:31:15.432053 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:15.432060 | orchestrator | 2025-08-29 17:31:15.432068 | orchestrator | TASK [kubectl : Add repository gpg key] **************************************** 2025-08-29 17:31:15.432075 | orchestrator | Friday 29 August 2025 17:30:26 +0000 (0:00:01.196) 0:02:45.611 ********* 2025-08-29 17:31:15.432082 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.432089 | orchestrator | 2025-08-29 17:31:15.432096 | orchestrator | TASK [kubectl : Set permissions of gpg key] ************************************ 2025-08-29 17:31:15.432104 | orchestrator | Friday 29 August 2025 17:30:27 +0000 (0:00:00.638) 0:02:46.250 ********* 2025-08-29 17:31:15.432111 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:15.432118 | orchestrator | 2025-08-29 17:31:15.432125 | orchestrator | TASK [kubectl : Add repository Debian] ***************************************** 2025-08-29 17:31:15.432132 | orchestrator | Friday 29 August 2025 17:30:27 +0000 (0:00:00.574) 0:02:46.825 ********* 2025-08-29 17:31:15.432140 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.432147 | orchestrator | 2025-08-29 17:31:15.432154 | orchestrator | TASK [kubectl : Install required packages] ************************************* 2025-08-29 17:31:15.432161 | orchestrator | Friday 29 August 2025 17:30:34 +0000 (0:00:06.404) 0:02:53.229 ********* 2025-08-29 17:31:15.432169 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.432176 | orchestrator | 2025-08-29 17:31:15.432183 | orchestrator | TASK [kubectl : Remove kubectl symlink] **************************************** 2025-08-29 17:31:15.432190 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:10.777) 0:03:04.007 ********* 2025-08-29 17:31:15.432197 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:15.432205 | orchestrator | 2025-08-29 17:31:15.432212 | orchestrator | PLAY [Run post actions on master nodes] **************************************** 2025-08-29 17:31:15.432219 | orchestrator | 2025-08-29 17:31:15.432226 | orchestrator | TASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] *** 2025-08-29 17:31:15.432237 | orchestrator | Friday 29 August 2025 17:30:45 +0000 (0:00:00.394) 0:03:04.401 ********* 2025-08-29 17:31:15.432244 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.432252 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.432259 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.432266 | orchestrator | 2025-08-29 17:31:15.432273 | orchestrator | TASK [k3s_server_post : Deploy calico] ***************************************** 2025-08-29 17:31:15.432285 | orchestrator | Friday 29 August 2025 17:30:45 +0000 (0:00:00.479) 0:03:04.881 ********* 2025-08-29 17:31:15.432292 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432299 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.432306 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.432314 | orchestrator | 2025-08-29 17:31:15.432321 | orchestrator | TASK [k3s_server_post : Deploy cilium] ***************************************** 2025-08-29 17:31:15.432328 | orchestrator | Friday 29 August 2025 17:30:46 +0000 (0:00:00.298) 0:03:05.179 ********* 2025-08-29 17:31:15.432335 | orchestrator | included: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:31:15.432342 | orchestrator | 2025-08-29 17:31:15.432360 | orchestrator | TASK [k3s_server_post : Create tmp directory on first master] ****************** 2025-08-29 17:31:15.432368 | orchestrator | Friday 29 August 2025 17:30:46 +0000 (0:00:00.476) 0:03:05.655 ********* 2025-08-29 17:31:15.432375 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432382 | orchestrator | 2025-08-29 17:31:15.432389 | orchestrator | TASK [k3s_server_post : Check if Cilium CLI is installed] ********************** 2025-08-29 17:31:15.432396 | orchestrator | Friday 29 August 2025 17:30:46 +0000 (0:00:00.154) 0:03:05.810 ********* 2025-08-29 17:31:15.432404 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432411 | orchestrator | 2025-08-29 17:31:15.432418 | orchestrator | TASK [k3s_server_post : Check for Cilium CLI version in command output] ******** 2025-08-29 17:31:15.432425 | orchestrator | Friday 29 August 2025 17:30:46 +0000 (0:00:00.157) 0:03:05.968 ********* 2025-08-29 17:31:15.432433 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432440 | orchestrator | 2025-08-29 17:31:15.432447 | orchestrator | TASK [k3s_server_post : Get latest stable Cilium CLI version file] ************* 2025-08-29 17:31:15.432454 | orchestrator | Friday 29 August 2025 17:30:47 +0000 (0:00:00.476) 0:03:06.445 ********* 2025-08-29 17:31:15.432461 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432469 | orchestrator | 2025-08-29 17:31:15.432476 | orchestrator | TASK [k3s_server_post : Read Cilium CLI stable version from file] ************** 2025-08-29 17:31:15.432483 | orchestrator | Friday 29 August 2025 17:30:47 +0000 (0:00:00.178) 0:03:06.623 ********* 2025-08-29 17:31:15.432490 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432497 | orchestrator | 2025-08-29 17:31:15.432505 | orchestrator | TASK [k3s_server_post : Log installed Cilium CLI version] ********************** 2025-08-29 17:31:15.432512 | orchestrator | Friday 29 August 2025 17:30:47 +0000 (0:00:00.180) 0:03:06.803 ********* 2025-08-29 17:31:15.432519 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432526 | orchestrator | 2025-08-29 17:31:15.432534 | orchestrator | TASK [k3s_server_post : Log latest stable Cilium CLI version] ****************** 2025-08-29 17:31:15.432541 | orchestrator | Friday 29 August 2025 17:30:47 +0000 (0:00:00.173) 0:03:06.976 ********* 2025-08-29 17:31:15.432551 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432559 | orchestrator | 2025-08-29 17:31:15.432566 | orchestrator | TASK [k3s_server_post : Determine if Cilium CLI needs installation or update] *** 2025-08-29 17:31:15.432573 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:00.246) 0:03:07.223 ********* 2025-08-29 17:31:15.432580 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432587 | orchestrator | 2025-08-29 17:31:15.432595 | orchestrator | TASK [k3s_server_post : Set architecture variable] ***************************** 2025-08-29 17:31:15.432602 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:00.236) 0:03:07.459 ********* 2025-08-29 17:31:15.432609 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432616 | orchestrator | 2025-08-29 17:31:15.432623 | orchestrator | TASK [k3s_server_post : Download Cilium CLI and checksum] ********************** 2025-08-29 17:31:15.432631 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:00.190) 0:03:07.650 ********* 2025-08-29 17:31:15.432638 | orchestrator | skipping: [testbed-node-0] => (item=.tar.gz)  2025-08-29 17:31:15.432645 | orchestrator | skipping: [testbed-node-0] => (item=.tar.gz.sha256sum)  2025-08-29 17:31:15.432652 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432663 | orchestrator | 2025-08-29 17:31:15.432671 | orchestrator | TASK [k3s_server_post : Verify the downloaded tarball] ************************* 2025-08-29 17:31:15.432678 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:00.263) 0:03:07.914 ********* 2025-08-29 17:31:15.432685 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432693 | orchestrator | 2025-08-29 17:31:15.432700 | orchestrator | TASK [k3s_server_post : Extract Cilium CLI to /usr/local/bin] ****************** 2025-08-29 17:31:15.432707 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:00.211) 0:03:08.125 ********* 2025-08-29 17:31:15.432714 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432721 | orchestrator | 2025-08-29 17:31:15.432729 | orchestrator | TASK [k3s_server_post : Remove downloaded tarball and checksum file] *********** 2025-08-29 17:31:15.432736 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:00.185) 0:03:08.311 ********* 2025-08-29 17:31:15.432743 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432750 | orchestrator | 2025-08-29 17:31:15.432758 | orchestrator | TASK [k3s_server_post : Wait for connectivity to kube VIP] ********************* 2025-08-29 17:31:15.432765 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:00.162) 0:03:08.474 ********* 2025-08-29 17:31:15.432772 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432779 | orchestrator | 2025-08-29 17:31:15.432787 | orchestrator | TASK [k3s_server_post : Fail if kube VIP not reachable] ************************ 2025-08-29 17:31:15.432794 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:00.146) 0:03:08.620 ********* 2025-08-29 17:31:15.432801 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432808 | orchestrator | 2025-08-29 17:31:15.432816 | orchestrator | TASK [k3s_server_post : Test for existing Cilium install] ********************** 2025-08-29 17:31:15.432823 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:00.170) 0:03:08.790 ********* 2025-08-29 17:31:15.432830 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432837 | orchestrator | 2025-08-29 17:31:15.432845 | orchestrator | TASK [k3s_server_post : Check Cilium version] ********************************** 2025-08-29 17:31:15.432856 | orchestrator | Friday 29 August 2025 17:30:50 +0000 (0:00:00.471) 0:03:09.262 ********* 2025-08-29 17:31:15.432863 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432871 | orchestrator | 2025-08-29 17:31:15.432878 | orchestrator | TASK [k3s_server_post : Parse installed Cilium version] ************************ 2025-08-29 17:31:15.432885 | orchestrator | Friday 29 August 2025 17:30:50 +0000 (0:00:00.164) 0:03:09.426 ********* 2025-08-29 17:31:15.432892 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432900 | orchestrator | 2025-08-29 17:31:15.432907 | orchestrator | TASK [k3s_server_post : Determine if Cilium needs update] ********************** 2025-08-29 17:31:15.432914 | orchestrator | Friday 29 August 2025 17:30:50 +0000 (0:00:00.219) 0:03:09.646 ********* 2025-08-29 17:31:15.432921 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432928 | orchestrator | 2025-08-29 17:31:15.432936 | orchestrator | TASK [k3s_server_post : Log result] ******************************************** 2025-08-29 17:31:15.432943 | orchestrator | Friday 29 August 2025 17:30:50 +0000 (0:00:00.188) 0:03:09.834 ********* 2025-08-29 17:31:15.432950 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432957 | orchestrator | 2025-08-29 17:31:15.432965 | orchestrator | TASK [k3s_server_post : Install Cilium] **************************************** 2025-08-29 17:31:15.432972 | orchestrator | Friday 29 August 2025 17:30:50 +0000 (0:00:00.144) 0:03:09.978 ********* 2025-08-29 17:31:15.432979 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.432986 | orchestrator | 2025-08-29 17:31:15.432994 | orchestrator | TASK [k3s_server_post : Wait for Cilium resources] ***************************** 2025-08-29 17:31:15.433001 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.178) 0:03:10.157 ********* 2025-08-29 17:31:15.433008 | orchestrator | skipping: [testbed-node-0] => (item=deployment/cilium-operator)  2025-08-29 17:31:15.433015 | orchestrator | skipping: [testbed-node-0] => (item=daemonset/cilium)  2025-08-29 17:31:15.433023 | orchestrator | skipping: [testbed-node-0] => (item=deployment/hubble-relay)  2025-08-29 17:31:15.433030 | orchestrator | skipping: [testbed-node-0] => (item=deployment/hubble-ui)  2025-08-29 17:31:15.433043 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433050 | orchestrator | 2025-08-29 17:31:15.433058 | orchestrator | TASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************ 2025-08-29 17:31:15.433065 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.406) 0:03:10.563 ********* 2025-08-29 17:31:15.433072 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433079 | orchestrator | 2025-08-29 17:31:15.433086 | orchestrator | TASK [k3s_server_post : Copy BGP manifests to first master] ******************** 2025-08-29 17:31:15.433094 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.193) 0:03:10.757 ********* 2025-08-29 17:31:15.433101 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433108 | orchestrator | 2025-08-29 17:31:15.433115 | orchestrator | TASK [k3s_server_post : Apply BGP manifests] *********************************** 2025-08-29 17:31:15.433123 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.186) 0:03:10.943 ********* 2025-08-29 17:31:15.433130 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433137 | orchestrator | 2025-08-29 17:31:15.433147 | orchestrator | TASK [k3s_server_post : Print error message if BGP manifests application fails] *** 2025-08-29 17:31:15.433155 | orchestrator | Friday 29 August 2025 17:30:52 +0000 (0:00:00.181) 0:03:11.125 ********* 2025-08-29 17:31:15.433162 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433169 | orchestrator | 2025-08-29 17:31:15.433176 | orchestrator | TASK [k3s_server_post : Test for BGP config resources] ************************* 2025-08-29 17:31:15.433184 | orchestrator | Friday 29 August 2025 17:30:52 +0000 (0:00:00.265) 0:03:11.391 ********* 2025-08-29 17:31:15.433191 | orchestrator | skipping: [testbed-node-0] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)  2025-08-29 17:31:15.433198 | orchestrator | skipping: [testbed-node-0] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)  2025-08-29 17:31:15.433205 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433212 | orchestrator | 2025-08-29 17:31:15.433219 | orchestrator | TASK [k3s_server_post : Deploy metallb pool] *********************************** 2025-08-29 17:31:15.433227 | orchestrator | Friday 29 August 2025 17:30:52 +0000 (0:00:00.409) 0:03:11.801 ********* 2025-08-29 17:31:15.433234 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433241 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.433248 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.433255 | orchestrator | 2025-08-29 17:31:15.433263 | orchestrator | TASK [k3s_server_post : Remove tmp directory used for manifests] *************** 2025-08-29 17:31:15.433270 | orchestrator | Friday 29 August 2025 17:30:53 +0000 (0:00:00.487) 0:03:12.288 ********* 2025-08-29 17:31:15.433277 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.433284 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.433292 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.433299 | orchestrator | 2025-08-29 17:31:15.433306 | orchestrator | PLAY [Apply role k9s] ********************************************************** 2025-08-29 17:31:15.433313 | orchestrator | 2025-08-29 17:31:15.433320 | orchestrator | TASK [k9s : Gather variables for each operating system] ************************ 2025-08-29 17:31:15.433328 | orchestrator | Friday 29 August 2025 17:30:54 +0000 (0:00:01.061) 0:03:13.349 ********* 2025-08-29 17:31:15.433335 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:15.433342 | orchestrator | 2025-08-29 17:31:15.433359 | orchestrator | TASK [k9s : Include distribution specific install tasks] *********************** 2025-08-29 17:31:15.433366 | orchestrator | Friday 29 August 2025 17:30:54 +0000 (0:00:00.105) 0:03:13.455 ********* 2025-08-29 17:31:15.433374 | orchestrator | included: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager 2025-08-29 17:31:15.433381 | orchestrator | 2025-08-29 17:31:15.433388 | orchestrator | TASK [k9s : Install k9s packages] ********************************************** 2025-08-29 17:31:15.433396 | orchestrator | Friday 29 August 2025 17:30:54 +0000 (0:00:00.283) 0:03:13.738 ********* 2025-08-29 17:31:15.433403 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:15.433410 | orchestrator | 2025-08-29 17:31:15.433418 | orchestrator | PLAY [Manage labels, annotations, and taints on all k3s nodes] ***************** 2025-08-29 17:31:15.433429 | orchestrator | 2025-08-29 17:31:15.433437 | orchestrator | TASK [Merge labels, annotations, and taints] *********************************** 2025-08-29 17:31:15.433447 | orchestrator | Friday 29 August 2025 17:30:59 +0000 (0:00:05.194) 0:03:18.933 ********* 2025-08-29 17:31:15.433455 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:15.433462 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:15.433469 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:15.433477 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:15.433484 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:15.433491 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:15.433498 | orchestrator | 2025-08-29 17:31:15.433506 | orchestrator | TASK [Manage labels] *********************************************************** 2025-08-29 17:31:15.433513 | orchestrator | Friday 29 August 2025 17:31:00 +0000 (0:00:00.437) 0:03:19.370 ********* 2025-08-29 17:31:15.433520 | orchestrator | ok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true) 2025-08-29 17:31:15.433528 | orchestrator | ok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true) 2025-08-29 17:31:15.433535 | orchestrator | ok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true) 2025-08-29 17:31:15.433542 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true) 2025-08-29 17:31:15.433549 | orchestrator | ok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true) 2025-08-29 17:31:15.433556 | orchestrator | ok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true) 2025-08-29 17:31:15.433564 | orchestrator | ok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker) 2025-08-29 17:31:15.433571 | orchestrator | ok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker) 2025-08-29 17:31:15.433578 | orchestrator | ok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker) 2025-08-29 17:31:15.433585 | orchestrator | ok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled) 2025-08-29 17:31:15.433592 | orchestrator | ok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled) 2025-08-29 17:31:15.433599 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled) 2025-08-29 17:31:15.433607 | orchestrator | ok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true) 2025-08-29 17:31:15.433614 | orchestrator | ok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true) 2025-08-29 17:31:15.433621 | orchestrator | ok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true) 2025-08-29 17:31:15.433628 | orchestrator | ok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true) 2025-08-29 17:31:15.433635 | orchestrator | ok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true) 2025-08-29 17:31:15.433643 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true) 2025-08-29 17:31:15.433653 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true) 2025-08-29 17:31:15.433661 | orchestrator | ok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true) 2025-08-29 17:31:15.433668 | orchestrator | ok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true) 2025-08-29 17:31:15.433675 | orchestrator | ok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true) 2025-08-29 17:31:15.433682 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true) 2025-08-29 17:31:15.433689 | orchestrator | ok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true) 2025-08-29 17:31:15.433696 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true) 2025-08-29 17:31:15.433704 | orchestrator | ok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true) 2025-08-29 17:31:15.433711 | orchestrator | ok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true) 2025-08-29 17:31:15.433722 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true) 2025-08-29 17:31:15.433729 | orchestrator | ok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true) 2025-08-29 17:31:15.433736 | orchestrator | ok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true) 2025-08-29 17:31:15.433744 | orchestrator | 2025-08-29 17:31:15.433751 | orchestrator | TASK [Manage annotations] ****************************************************** 2025-08-29 17:31:15.433786 | orchestrator | Friday 29 August 2025 17:31:12 +0000 (0:00:12.324) 0:03:31.695 ********* 2025-08-29 17:31:15.433795 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.433802 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.433809 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.433817 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433824 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.433831 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.433838 | orchestrator | 2025-08-29 17:31:15.433846 | orchestrator | TASK [Manage taints] *********************************************************** 2025-08-29 17:31:15.433853 | orchestrator | Friday 29 August 2025 17:31:13 +0000 (0:00:00.386) 0:03:32.082 ********* 2025-08-29 17:31:15.433860 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:15.433867 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:15.433874 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:15.433882 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:15.433889 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:15.433896 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:15.433903 | orchestrator | 2025-08-29 17:31:15.433910 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:31:15.433922 | orchestrator | testbed-manager : ok=21  changed=11  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:31:15.433930 | orchestrator | testbed-node-0 : ok=42  changed=20  unreachable=0 failed=0 skipped=45  rescued=0 ignored=0 2025-08-29 17:31:15.433938 | orchestrator | testbed-node-1 : ok=39  changed=17  unreachable=0 failed=0 skipped=21  rescued=0 ignored=0 2025-08-29 17:31:15.433945 | orchestrator | testbed-node-2 : ok=39  changed=17  unreachable=0 failed=0 skipped=21  rescued=0 ignored=0 2025-08-29 17:31:15.433953 | orchestrator | testbed-node-3 : ok=19  changed=9  unreachable=0 failed=0 skipped=13  rescued=0 ignored=0 2025-08-29 17:31:15.433960 | orchestrator | testbed-node-4 : ok=19  changed=9  unreachable=0 failed=0 skipped=13  rescued=0 ignored=0 2025-08-29 17:31:15.433967 | orchestrator | testbed-node-5 : ok=19  changed=9  unreachable=0 failed=0 skipped=13  rescued=0 ignored=0 2025-08-29 17:31:15.433974 | orchestrator | 2025-08-29 17:31:15.433982 | orchestrator | 2025-08-29 17:31:15.433989 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:31:15.433996 | orchestrator | Friday 29 August 2025 17:31:13 +0000 (0:00:00.452) 0:03:32.535 ********* 2025-08-29 17:31:15.434003 | orchestrator | =============================================================================== 2025-08-29 17:31:15.434010 | orchestrator | k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 55.73s 2025-08-29 17:31:15.434040 | orchestrator | k3s_server : Enable and check K3s service ------------------------------ 24.59s 2025-08-29 17:31:15.434048 | orchestrator | Manage labels ---------------------------------------------------------- 12.32s 2025-08-29 17:31:15.434055 | orchestrator | k3s_agent : Manage k3s service ----------------------------------------- 11.97s 2025-08-29 17:31:15.434062 | orchestrator | kubectl : Install required packages ------------------------------------ 10.78s 2025-08-29 17:31:15.434074 | orchestrator | kubectl : Add repository Debian ----------------------------------------- 6.40s 2025-08-29 17:31:15.434082 | orchestrator | k3s_download : Download k3s binary x64 ---------------------------------- 5.98s 2025-08-29 17:31:15.434089 | orchestrator | k9s : Install k9s packages ---------------------------------------------- 5.19s 2025-08-29 17:31:15.434096 | orchestrator | k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 3.33s 2025-08-29 17:31:15.434107 | orchestrator | k3s_custom_registries : Create directory /etc/rancher/k3s --------------- 2.65s 2025-08-29 17:31:15.434114 | orchestrator | k3s_server : Init cluster inside the transient k3s-init service --------- 2.52s 2025-08-29 17:31:15.434122 | orchestrator | k3s_download : Download k3s binary armhf -------------------------------- 2.37s 2025-08-29 17:31:15.434129 | orchestrator | k3s_prereq : Enable IPv4 forwarding ------------------------------------- 1.99s 2025-08-29 17:31:15.434136 | orchestrator | k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml --- 1.93s 2025-08-29 17:31:15.434143 | orchestrator | k3s_server : Copy vip manifest to first master -------------------------- 1.88s 2025-08-29 17:31:15.434150 | orchestrator | k3s_download : Download k3s binary arm64 -------------------------------- 1.63s 2025-08-29 17:31:15.434157 | orchestrator | k3s_server : Stop k3s-init ---------------------------------------------- 1.61s 2025-08-29 17:31:15.434164 | orchestrator | Make kubeconfig available for use inside the manager service ------------ 1.57s 2025-08-29 17:31:15.434172 | orchestrator | k3s_server : Create custom resolv.conf for k3s -------------------------- 1.42s 2025-08-29 17:31:15.434179 | orchestrator | k3s_server : Stop k3s --------------------------------------------------- 1.36s 2025-08-29 17:31:15.434186 | orchestrator | 2025-08-29 17:31:15 | INFO  | Task d07c9d88-ed7b-4446-acc6-2ed4b9726847 is in state STARTED 2025-08-29 17:31:15.434194 | orchestrator | 2025-08-29 17:31:15 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:15.434201 | orchestrator | 2025-08-29 17:31:15 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:15.434265 | orchestrator | 2025-08-29 17:31:15 | INFO  | Task 381c333b-e0ad-4796-88d0-51a262ec5073 is in state STARTED 2025-08-29 17:31:15.434511 | orchestrator | 2025-08-29 17:31:15 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:15.434669 | orchestrator | 2025-08-29 17:31:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:18.596333 | orchestrator | 2025-08-29 17:31:18 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:18.596507 | orchestrator | 2025-08-29 17:31:18 | INFO  | Task d07c9d88-ed7b-4446-acc6-2ed4b9726847 is in state STARTED 2025-08-29 17:31:18.596534 | orchestrator | 2025-08-29 17:31:18 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:18.596554 | orchestrator | 2025-08-29 17:31:18 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:18.596574 | orchestrator | 2025-08-29 17:31:18 | INFO  | Task 381c333b-e0ad-4796-88d0-51a262ec5073 is in state STARTED 2025-08-29 17:31:18.596592 | orchestrator | 2025-08-29 17:31:18 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:18.596611 | orchestrator | 2025-08-29 17:31:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:21.619502 | orchestrator | 2025-08-29 17:31:21 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:21.619589 | orchestrator | 2025-08-29 17:31:21 | INFO  | Task d07c9d88-ed7b-4446-acc6-2ed4b9726847 is in state SUCCESS 2025-08-29 17:31:21.621729 | orchestrator | 2025-08-29 17:31:21 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:21.622155 | orchestrator | 2025-08-29 17:31:21 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:21.622747 | orchestrator | 2025-08-29 17:31:21 | INFO  | Task 381c333b-e0ad-4796-88d0-51a262ec5073 is in state STARTED 2025-08-29 17:31:21.624688 | orchestrator | 2025-08-29 17:31:21 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:21.624718 | orchestrator | 2025-08-29 17:31:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:24.654146 | orchestrator | 2025-08-29 17:31:24 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:24.654431 | orchestrator | 2025-08-29 17:31:24 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:24.654816 | orchestrator | 2025-08-29 17:31:24 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:24.655123 | orchestrator | 2025-08-29 17:31:24 | INFO  | Task 381c333b-e0ad-4796-88d0-51a262ec5073 is in state SUCCESS 2025-08-29 17:31:24.655810 | orchestrator | 2025-08-29 17:31:24 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:24.655838 | orchestrator | 2025-08-29 17:31:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:27.682232 | orchestrator | 2025-08-29 17:31:27 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state STARTED 2025-08-29 17:31:27.683716 | orchestrator | 2025-08-29 17:31:27 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:27.685246 | orchestrator | 2025-08-29 17:31:27 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:27.686575 | orchestrator | 2025-08-29 17:31:27 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:27.686916 | orchestrator | 2025-08-29 17:31:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:30.794553 | orchestrator | 2025-08-29 17:31:30.794720 | orchestrator | 2025-08-29 17:31:30.794738 | orchestrator | PLAY [Copy kubeconfig to the configuration repository] ************************* 2025-08-29 17:31:30.794751 | orchestrator | 2025-08-29 17:31:30.794762 | orchestrator | TASK [Get kubeconfig file] ***************************************************** 2025-08-29 17:31:30.794773 | orchestrator | Friday 29 August 2025 17:31:17 +0000 (0:00:00.160) 0:00:00.160 ********* 2025-08-29 17:31:30.794785 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] 2025-08-29 17:31:30.794796 | orchestrator | 2025-08-29 17:31:30.794808 | orchestrator | TASK [Write kubeconfig file] *************************************************** 2025-08-29 17:31:30.794818 | orchestrator | Friday 29 August 2025 17:31:18 +0000 (0:00:00.625) 0:00:00.786 ********* 2025-08-29 17:31:30.794829 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:30.794840 | orchestrator | 2025-08-29 17:31:30.794851 | orchestrator | TASK [Change server address in the kubeconfig file] **************************** 2025-08-29 17:31:30.794862 | orchestrator | Friday 29 August 2025 17:31:19 +0000 (0:00:01.077) 0:00:01.864 ********* 2025-08-29 17:31:30.794873 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:30.794883 | orchestrator | 2025-08-29 17:31:30.794894 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:31:30.794905 | orchestrator | testbed-manager : ok=3  changed=2  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:31:30.794917 | orchestrator | 2025-08-29 17:31:30.794928 | orchestrator | 2025-08-29 17:31:30.794939 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:31:30.794950 | orchestrator | Friday 29 August 2025 17:31:19 +0000 (0:00:00.329) 0:00:02.193 ********* 2025-08-29 17:31:30.794961 | orchestrator | =============================================================================== 2025-08-29 17:31:30.794971 | orchestrator | Write kubeconfig file --------------------------------------------------- 1.08s 2025-08-29 17:31:30.795003 | orchestrator | Get kubeconfig file ----------------------------------------------------- 0.63s 2025-08-29 17:31:30.795015 | orchestrator | Change server address in the kubeconfig file ---------------------------- 0.33s 2025-08-29 17:31:30.795025 | orchestrator | 2025-08-29 17:31:30.795036 | orchestrator | 2025-08-29 17:31:30.795047 | orchestrator | PLAY [Prepare kubeconfig file] ************************************************* 2025-08-29 17:31:30.795057 | orchestrator | 2025-08-29 17:31:30.795068 | orchestrator | TASK [Get home directory of operator user] ************************************* 2025-08-29 17:31:30.795079 | orchestrator | Friday 29 August 2025 17:31:17 +0000 (0:00:00.194) 0:00:00.194 ********* 2025-08-29 17:31:30.795090 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:30.795101 | orchestrator | 2025-08-29 17:31:30.795114 | orchestrator | TASK [Create .kube directory] ************************************************** 2025-08-29 17:31:30.795126 | orchestrator | Friday 29 August 2025 17:31:18 +0000 (0:00:00.584) 0:00:00.778 ********* 2025-08-29 17:31:30.795138 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:30.795150 | orchestrator | 2025-08-29 17:31:30.795162 | orchestrator | TASK [Get kubeconfig file] ***************************************************** 2025-08-29 17:31:30.795175 | orchestrator | Friday 29 August 2025 17:31:18 +0000 (0:00:00.445) 0:00:01.224 ********* 2025-08-29 17:31:30.795186 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] 2025-08-29 17:31:30.795199 | orchestrator | 2025-08-29 17:31:30.795211 | orchestrator | TASK [Write kubeconfig file] *************************************************** 2025-08-29 17:31:30.795223 | orchestrator | Friday 29 August 2025 17:31:19 +0000 (0:00:00.615) 0:00:01.839 ********* 2025-08-29 17:31:30.795235 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:30.795245 | orchestrator | 2025-08-29 17:31:30.795256 | orchestrator | TASK [Change server address in the kubeconfig] ********************************* 2025-08-29 17:31:30.795267 | orchestrator | Friday 29 August 2025 17:31:20 +0000 (0:00:00.999) 0:00:02.839 ********* 2025-08-29 17:31:30.795277 | orchestrator | changed: [testbed-manager] 2025-08-29 17:31:30.795288 | orchestrator | 2025-08-29 17:31:30.795299 | orchestrator | TASK [Make kubeconfig available for use inside the manager service] ************ 2025-08-29 17:31:30.795309 | orchestrator | Friday 29 August 2025 17:31:21 +0000 (0:00:00.640) 0:00:03.479 ********* 2025-08-29 17:31:30.795320 | orchestrator | changed: [testbed-manager -> localhost] 2025-08-29 17:31:30.795331 | orchestrator | 2025-08-29 17:31:30.795341 | orchestrator | TASK [Change server address in the kubeconfig inside the manager service] ****** 2025-08-29 17:31:30.795380 | orchestrator | Friday 29 August 2025 17:31:22 +0000 (0:00:01.240) 0:00:04.720 ********* 2025-08-29 17:31:30.795392 | orchestrator | changed: [testbed-manager -> localhost] 2025-08-29 17:31:30.795402 | orchestrator | 2025-08-29 17:31:30.795413 | orchestrator | TASK [Set KUBECONFIG environment variable] ************************************* 2025-08-29 17:31:30.795424 | orchestrator | Friday 29 August 2025 17:31:23 +0000 (0:00:00.994) 0:00:05.715 ********* 2025-08-29 17:31:30.795435 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:30.795445 | orchestrator | 2025-08-29 17:31:30.795456 | orchestrator | TASK [Enable kubectl command line completion] ********************************** 2025-08-29 17:31:30.795467 | orchestrator | Friday 29 August 2025 17:31:23 +0000 (0:00:00.295) 0:00:06.010 ********* 2025-08-29 17:31:30.795477 | orchestrator | ok: [testbed-manager] 2025-08-29 17:31:30.795488 | orchestrator | 2025-08-29 17:31:30.795499 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:31:30.795510 | orchestrator | testbed-manager : ok=9  changed=4  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:31:30.795521 | orchestrator | 2025-08-29 17:31:30.795531 | orchestrator | 2025-08-29 17:31:30.795542 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:31:30.795565 | orchestrator | Friday 29 August 2025 17:31:24 +0000 (0:00:00.277) 0:00:06.287 ********* 2025-08-29 17:31:30.795576 | orchestrator | =============================================================================== 2025-08-29 17:31:30.795587 | orchestrator | Make kubeconfig available for use inside the manager service ------------ 1.24s 2025-08-29 17:31:30.795605 | orchestrator | Write kubeconfig file --------------------------------------------------- 1.00s 2025-08-29 17:31:30.795617 | orchestrator | Change server address in the kubeconfig inside the manager service ------ 0.99s 2025-08-29 17:31:30.795644 | orchestrator | Change server address in the kubeconfig --------------------------------- 0.64s 2025-08-29 17:31:30.795656 | orchestrator | Get kubeconfig file ----------------------------------------------------- 0.62s 2025-08-29 17:31:30.795667 | orchestrator | Get home directory of operator user ------------------------------------- 0.58s 2025-08-29 17:31:30.795677 | orchestrator | Create .kube directory -------------------------------------------------- 0.45s 2025-08-29 17:31:30.795688 | orchestrator | Set KUBECONFIG environment variable ------------------------------------- 0.30s 2025-08-29 17:31:30.795699 | orchestrator | Enable kubectl command line completion ---------------------------------- 0.28s 2025-08-29 17:31:30.795709 | orchestrator | 2025-08-29 17:31:30.795720 | orchestrator | 2025-08-29 17:31:30.795731 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:31:30.795741 | orchestrator | 2025-08-29 17:31:30.795752 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:31:30.795763 | orchestrator | Friday 29 August 2025 17:30:23 +0000 (0:00:00.219) 0:00:00.219 ********* 2025-08-29 17:31:30.795773 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:30.795784 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:30.795795 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:30.795806 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:30.795816 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:30.795827 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:30.795838 | orchestrator | 2025-08-29 17:31:30.795848 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:31:30.795859 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.932) 0:00:01.151 ********* 2025-08-29 17:31:30.795870 | orchestrator | ok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False) 2025-08-29 17:31:30.795882 | orchestrator | ok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False) 2025-08-29 17:31:30.795892 | orchestrator | ok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False) 2025-08-29 17:31:30.795903 | orchestrator | ok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False) 2025-08-29 17:31:30.795914 | orchestrator | ok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False) 2025-08-29 17:31:30.795924 | orchestrator | ok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False) 2025-08-29 17:31:30.795935 | orchestrator | 2025-08-29 17:31:30.795946 | orchestrator | PLAY [Apply role openvswitch] ************************************************** 2025-08-29 17:31:30.795957 | orchestrator | 2025-08-29 17:31:30.795967 | orchestrator | TASK [openvswitch : include_tasks] ********************************************* 2025-08-29 17:31:30.795978 | orchestrator | Friday 29 August 2025 17:30:25 +0000 (0:00:00.882) 0:00:02.034 ********* 2025-08-29 17:31:30.795989 | orchestrator | included: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:31:30.796001 | orchestrator | 2025-08-29 17:31:30.796012 | orchestrator | TASK [module-load : Load modules] ********************************************** 2025-08-29 17:31:30.796022 | orchestrator | Friday 29 August 2025 17:30:26 +0000 (0:00:01.418) 0:00:03.453 ********* 2025-08-29 17:31:30.796033 | orchestrator | changed: [testbed-node-3] => (item=openvswitch) 2025-08-29 17:31:30.796044 | orchestrator | changed: [testbed-node-4] => (item=openvswitch) 2025-08-29 17:31:30.796055 | orchestrator | changed: [testbed-node-5] => (item=openvswitch) 2025-08-29 17:31:30.796065 | orchestrator | changed: [testbed-node-0] => (item=openvswitch) 2025-08-29 17:31:30.796076 | orchestrator | changed: [testbed-node-1] => (item=openvswitch) 2025-08-29 17:31:30.796087 | orchestrator | changed: [testbed-node-2] => (item=openvswitch) 2025-08-29 17:31:30.796097 | orchestrator | 2025-08-29 17:31:30.796108 | orchestrator | TASK [module-load : Persist modules via modules-load.d] ************************ 2025-08-29 17:31:30.796125 | orchestrator | Friday 29 August 2025 17:30:28 +0000 (0:00:01.254) 0:00:04.708 ********* 2025-08-29 17:31:30.796136 | orchestrator | changed: [testbed-node-5] => (item=openvswitch) 2025-08-29 17:31:30.796147 | orchestrator | changed: [testbed-node-3] => (item=openvswitch) 2025-08-29 17:31:30.796157 | orchestrator | changed: [testbed-node-4] => (item=openvswitch) 2025-08-29 17:31:30.796169 | orchestrator | changed: [testbed-node-0] => (item=openvswitch) 2025-08-29 17:31:30.796180 | orchestrator | changed: [testbed-node-1] => (item=openvswitch) 2025-08-29 17:31:30.796190 | orchestrator | changed: [testbed-node-2] => (item=openvswitch) 2025-08-29 17:31:30.796201 | orchestrator | 2025-08-29 17:31:30.796218 | orchestrator | TASK [module-load : Drop module persistence] *********************************** 2025-08-29 17:31:30.796237 | orchestrator | Friday 29 August 2025 17:30:30 +0000 (0:00:02.625) 0:00:07.334 ********* 2025-08-29 17:31:30.796256 | orchestrator | skipping: [testbed-node-3] => (item=openvswitch)  2025-08-29 17:31:30.796276 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:30.796295 | orchestrator | skipping: [testbed-node-4] => (item=openvswitch)  2025-08-29 17:31:30.796314 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:30.796328 | orchestrator | skipping: [testbed-node-5] => (item=openvswitch)  2025-08-29 17:31:30.796339 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:30.796365 | orchestrator | skipping: [testbed-node-0] => (item=openvswitch)  2025-08-29 17:31:30.796377 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:30.796387 | orchestrator | skipping: [testbed-node-1] => (item=openvswitch)  2025-08-29 17:31:30.796398 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:30.796414 | orchestrator | skipping: [testbed-node-2] => (item=openvswitch)  2025-08-29 17:31:30.796425 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:30.796436 | orchestrator | 2025-08-29 17:31:30.796447 | orchestrator | TASK [openvswitch : Create /run/openvswitch directory on host] ***************** 2025-08-29 17:31:30.796457 | orchestrator | Friday 29 August 2025 17:30:32 +0000 (0:00:01.355) 0:00:08.689 ********* 2025-08-29 17:31:30.796468 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:30.796479 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:30.796489 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:30.796507 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:30.796518 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:30.796528 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:30.796539 | orchestrator | 2025-08-29 17:31:30.796549 | orchestrator | TASK [openvswitch : Ensuring config directories exist] ************************* 2025-08-29 17:31:30.796560 | orchestrator | Friday 29 August 2025 17:30:32 +0000 (0:00:00.702) 0:00:09.392 ********* 2025-08-29 17:31:30.796574 | orchestrator | changed: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.796590 | orchestrator | changed: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.796610 | orchestrator | changed: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.796622 | orchestrator | changed: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.796633 | orchestrator | changed: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.796652 | orchestrator | changed: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.796664 | orchestrator | changed: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797258 | orchestrator | changed: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797282 | orchestrator | changed: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797294 | orchestrator | changed: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797305 | orchestrator | changed: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797326 | orchestrator | changed: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797338 | orchestrator | 2025-08-29 17:31:30.797401 | orchestrator | TASK [openvswitch : Copying over config.json files for services] *************** 2025-08-29 17:31:30.797417 | orchestrator | Friday 29 August 2025 17:30:34 +0000 (0:00:01.997) 0:00:11.390 ********* 2025-08-29 17:31:30.797434 | orchestrator | changed: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797453 | orchestrator | changed: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797465 | orchestrator | changed: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797477 | orchestrator | changed: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797496 | orchestrator | changed: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797508 | orchestrator | changed: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797530 | orchestrator | changed: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797542 | orchestrator | changed: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797553 | orchestrator | changed: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797565 | orchestrator | changed: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797583 | orchestrator | changed: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797595 | orchestrator | changed: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797612 | orchestrator | 2025-08-29 17:31:30.797628 | orchestrator | TASK [openvswitch : Copying over ovs-vsctl wrapper] **************************** 2025-08-29 17:31:30.797639 | orchestrator | Friday 29 August 2025 17:30:38 +0000 (0:00:04.085) 0:00:15.476 ********* 2025-08-29 17:31:30.797650 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:30.797661 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:30.797672 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:30.797683 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:31:30.797694 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:31:30.797704 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:31:30.797715 | orchestrator | 2025-08-29 17:31:30.797726 | orchestrator | TASK [openvswitch : Check openvswitch containers] ****************************** 2025-08-29 17:31:30.797737 | orchestrator | Friday 29 August 2025 17:30:40 +0000 (0:00:01.252) 0:00:16.728 ********* 2025-08-29 17:31:30.797748 | orchestrator | changed: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797759 | orchestrator | changed: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797771 | orchestrator | changed: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797789 | orchestrator | changed: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797807 | orchestrator | changed: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797822 | orchestrator | changed: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797834 | orchestrator | changed: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797845 | orchestrator | changed: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797862 | orchestrator | changed: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797874 | orchestrator | changed: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797900 | orchestrator | changed: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797912 | orchestrator | changed: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}}) 2025-08-29 17:31:30.797923 | orchestrator | 2025-08-29 17:31:30.797934 | orchestrator | TASK [openvswitch : Flush Handlers] ******************************************** 2025-08-29 17:31:30.797945 | orchestrator | Friday 29 August 2025 17:30:43 +0000 (0:00:03.155) 0:00:19.883 ********* 2025-08-29 17:31:30.797956 | orchestrator | 2025-08-29 17:31:30.797967 | orchestrator | TASK [openvswitch : Flush Handlers] ******************************************** 2025-08-29 17:31:30.797978 | orchestrator | Friday 29 August 2025 17:30:43 +0000 (0:00:00.337) 0:00:20.220 ********* 2025-08-29 17:31:30.797988 | orchestrator | 2025-08-29 17:31:30.797999 | orchestrator | TASK [openvswitch : Flush Handlers] ******************************************** 2025-08-29 17:31:30.798010 | orchestrator | Friday 29 August 2025 17:30:43 +0000 (0:00:00.277) 0:00:20.498 ********* 2025-08-29 17:31:30.798070 | orchestrator | 2025-08-29 17:31:30.798082 | orchestrator | TASK [openvswitch : Flush Handlers] ******************************************** 2025-08-29 17:31:30.798093 | orchestrator | Friday 29 August 2025 17:30:43 +0000 (0:00:00.134) 0:00:20.632 ********* 2025-08-29 17:31:30.798104 | orchestrator | 2025-08-29 17:31:30.798114 | orchestrator | TASK [openvswitch : Flush Handlers] ******************************************** 2025-08-29 17:31:30.798125 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:00.109) 0:00:20.742 ********* 2025-08-29 17:31:30.798136 | orchestrator | 2025-08-29 17:31:30.798146 | orchestrator | TASK [openvswitch : Flush Handlers] ******************************************** 2025-08-29 17:31:30.798157 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:00.123) 0:00:20.866 ********* 2025-08-29 17:31:30.798168 | orchestrator | 2025-08-29 17:31:30.798178 | orchestrator | RUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ******** 2025-08-29 17:31:30.798189 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:00.297) 0:00:21.163 ********* 2025-08-29 17:31:30.798200 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:30.798210 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:30.798221 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:30.798231 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:30.798249 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:30.798260 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:30.798270 | orchestrator | 2025-08-29 17:31:30.798281 | orchestrator | RUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] *** 2025-08-29 17:31:30.798292 | orchestrator | Friday 29 August 2025 17:30:55 +0000 (0:00:10.545) 0:00:31.709 ********* 2025-08-29 17:31:30.798302 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:31:30.798313 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:31:30.798324 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:31:30.798334 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:31:30.798345 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:31:30.798372 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:31:30.798383 | orchestrator | 2025-08-29 17:31:30.798394 | orchestrator | RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] ********* 2025-08-29 17:31:30.798412 | orchestrator | Friday 29 August 2025 17:30:56 +0000 (0:00:01.874) 0:00:33.583 ********* 2025-08-29 17:31:30.798424 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:30.798435 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:30.798446 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:30.798456 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:30.798467 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:30.798478 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:30.798489 | orchestrator | 2025-08-29 17:31:30.798499 | orchestrator | TASK [openvswitch : Set system-id, hostname and hw-offload] ******************** 2025-08-29 17:31:30.798510 | orchestrator | Friday 29 August 2025 17:31:06 +0000 (0:00:09.584) 0:00:43.167 ********* 2025-08-29 17:31:30.798521 | orchestrator | changed: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'}) 2025-08-29 17:31:30.798532 | orchestrator | changed: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'}) 2025-08-29 17:31:30.798543 | orchestrator | changed: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'}) 2025-08-29 17:31:30.798554 | orchestrator | changed: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'}) 2025-08-29 17:31:30.798564 | orchestrator | changed: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'}) 2025-08-29 17:31:30.798575 | orchestrator | changed: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'}) 2025-08-29 17:31:30.798586 | orchestrator | changed: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'}) 2025-08-29 17:31:30.798602 | orchestrator | changed: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'}) 2025-08-29 17:31:30.798613 | orchestrator | changed: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'}) 2025-08-29 17:31:30.798624 | orchestrator | changed: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'}) 2025-08-29 17:31:30.798634 | orchestrator | changed: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'}) 2025-08-29 17:31:30.798645 | orchestrator | changed: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'}) 2025-08-29 17:31:30.798656 | orchestrator | ok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'}) 2025-08-29 17:31:30.798667 | orchestrator | ok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'}) 2025-08-29 17:31:30.798677 | orchestrator | ok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'}) 2025-08-29 17:31:30.798688 | orchestrator | ok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'}) 2025-08-29 17:31:30.798699 | orchestrator | ok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'}) 2025-08-29 17:31:30.798716 | orchestrator | ok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'}) 2025-08-29 17:31:30.798727 | orchestrator | 2025-08-29 17:31:30.798738 | orchestrator | TASK [openvswitch : Ensuring OVS bridge is properly setup] ********************* 2025-08-29 17:31:30.798749 | orchestrator | Friday 29 August 2025 17:31:13 +0000 (0:00:07.177) 0:00:50.344 ********* 2025-08-29 17:31:30.798760 | orchestrator | skipping: [testbed-node-3] => (item=br-ex)  2025-08-29 17:31:30.798771 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:30.798781 | orchestrator | skipping: [testbed-node-4] => (item=br-ex)  2025-08-29 17:31:30.798792 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:30.798803 | orchestrator | skipping: [testbed-node-5] => (item=br-ex)  2025-08-29 17:31:30.798814 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:30.798824 | orchestrator | changed: [testbed-node-0] => (item=br-ex) 2025-08-29 17:31:30.798835 | orchestrator | changed: [testbed-node-1] => (item=br-ex) 2025-08-29 17:31:30.798846 | orchestrator | changed: [testbed-node-2] => (item=br-ex) 2025-08-29 17:31:30.798857 | orchestrator | 2025-08-29 17:31:30.798868 | orchestrator | TASK [openvswitch : Ensuring OVS ports are properly setup] ********************* 2025-08-29 17:31:30.798878 | orchestrator | Friday 29 August 2025 17:31:16 +0000 (0:00:02.944) 0:00:53.289 ********* 2025-08-29 17:31:30.798889 | orchestrator | skipping: [testbed-node-3] => (item=['br-ex', 'vxlan0'])  2025-08-29 17:31:30.798900 | orchestrator | skipping: [testbed-node-4] => (item=['br-ex', 'vxlan0'])  2025-08-29 17:31:30.798911 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:31:30.798922 | orchestrator | skipping: [testbed-node-5] => (item=['br-ex', 'vxlan0'])  2025-08-29 17:31:30.798932 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:31:30.798943 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:31:30.798954 | orchestrator | changed: [testbed-node-0] => (item=['br-ex', 'vxlan0']) 2025-08-29 17:31:30.798965 | orchestrator | changed: [testbed-node-1] => (item=['br-ex', 'vxlan0']) 2025-08-29 17:31:30.798976 | orchestrator | changed: [testbed-node-2] => (item=['br-ex', 'vxlan0']) 2025-08-29 17:31:30.798987 | orchestrator | 2025-08-29 17:31:30.798997 | orchestrator | RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] ********* 2025-08-29 17:31:30.799008 | orchestrator | Friday 29 August 2025 17:31:20 +0000 (0:00:04.298) 0:00:57.587 ********* 2025-08-29 17:31:30.799019 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:31:30.799030 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:31:30.799046 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:31:30.799057 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:31:30.799068 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:31:30.799079 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:31:30.799090 | orchestrator | 2025-08-29 17:31:30.799100 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:31:30.799111 | orchestrator | testbed-node-0 : ok=15  changed=11  unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 17:31:30.799123 | orchestrator | testbed-node-1 : ok=15  changed=11  unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 17:31:30.799134 | orchestrator | testbed-node-2 : ok=15  changed=11  unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 17:31:30.799145 | orchestrator | testbed-node-3 : ok=13  changed=9  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:31:30.799156 | orchestrator | testbed-node-4 : ok=13  changed=9  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:31:30.799167 | orchestrator | testbed-node-5 : ok=13  changed=9  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:31:30.799183 | orchestrator | 2025-08-29 17:31:30.799194 | orchestrator | 2025-08-29 17:31:30.799205 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:31:30.799220 | orchestrator | Friday 29 August 2025 17:31:28 +0000 (0:00:07.963) 0:01:05.551 ********* 2025-08-29 17:31:30.799232 | orchestrator | =============================================================================== 2025-08-29 17:31:30.799243 | orchestrator | openvswitch : Restart openvswitch-vswitchd container ------------------- 17.55s 2025-08-29 17:31:30.799253 | orchestrator | openvswitch : Restart openvswitch-db-server container ------------------ 10.55s 2025-08-29 17:31:30.799264 | orchestrator | openvswitch : Set system-id, hostname and hw-offload -------------------- 7.18s 2025-08-29 17:31:30.799275 | orchestrator | openvswitch : Ensuring OVS ports are properly setup --------------------- 4.30s 2025-08-29 17:31:30.799285 | orchestrator | openvswitch : Copying over config.json files for services --------------- 4.09s 2025-08-29 17:31:30.799296 | orchestrator | openvswitch : Check openvswitch containers ------------------------------ 3.16s 2025-08-29 17:31:30.799307 | orchestrator | openvswitch : Ensuring OVS bridge is properly setup --------------------- 2.94s 2025-08-29 17:31:30.799318 | orchestrator | module-load : Persist modules via modules-load.d ------------------------ 2.63s 2025-08-29 17:31:30.799328 | orchestrator | openvswitch : Ensuring config directories exist ------------------------- 2.00s 2025-08-29 17:31:30.799339 | orchestrator | openvswitch : Waiting for openvswitch_db service to be ready ------------ 1.87s 2025-08-29 17:31:30.799361 | orchestrator | openvswitch : include_tasks --------------------------------------------- 1.42s 2025-08-29 17:31:30.799373 | orchestrator | module-load : Drop module persistence ----------------------------------- 1.36s 2025-08-29 17:31:30.799384 | orchestrator | openvswitch : Flush Handlers -------------------------------------------- 1.28s 2025-08-29 17:31:30.799394 | orchestrator | module-load : Load modules ---------------------------------------------- 1.25s 2025-08-29 17:31:30.799405 | orchestrator | openvswitch : Copying over ovs-vsctl wrapper ---------------------------- 1.25s 2025-08-29 17:31:30.799416 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.93s 2025-08-29 17:31:30.799426 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.88s 2025-08-29 17:31:30.799437 | orchestrator | openvswitch : Create /run/openvswitch directory on host ----------------- 0.70s 2025-08-29 17:31:30.799448 | orchestrator | 2025-08-29 17:31:30 | INFO  | Task e197a489-cf57-46f3-9c72-f0b3041807d8 is in state SUCCESS 2025-08-29 17:31:30.799458 | orchestrator | 2025-08-29 17:31:30 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:30.799469 | orchestrator | 2025-08-29 17:31:30 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:30.799480 | orchestrator | 2025-08-29 17:31:30 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:30.799491 | orchestrator | 2025-08-29 17:31:30 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:30.799502 | orchestrator | 2025-08-29 17:31:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:33.775584 | orchestrator | 2025-08-29 17:31:33 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:33.775689 | orchestrator | 2025-08-29 17:31:33 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:33.776059 | orchestrator | 2025-08-29 17:31:33 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:33.776684 | orchestrator | 2025-08-29 17:31:33 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:33.776705 | orchestrator | 2025-08-29 17:31:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:36.914589 | orchestrator | 2025-08-29 17:31:36 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:36.914653 | orchestrator | 2025-08-29 17:31:36 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:36.915456 | orchestrator | 2025-08-29 17:31:36 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:36.916547 | orchestrator | 2025-08-29 17:31:36 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:36.916654 | orchestrator | 2025-08-29 17:31:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:39.938624 | orchestrator | 2025-08-29 17:31:39 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:39.941174 | orchestrator | 2025-08-29 17:31:39 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:39.941194 | orchestrator | 2025-08-29 17:31:39 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:39.941202 | orchestrator | 2025-08-29 17:31:39 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:39.941208 | orchestrator | 2025-08-29 17:31:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:42.970972 | orchestrator | 2025-08-29 17:31:42 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:42.974187 | orchestrator | 2025-08-29 17:31:42 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:42.975447 | orchestrator | 2025-08-29 17:31:42 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:42.976903 | orchestrator | 2025-08-29 17:31:42 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:42.977251 | orchestrator | 2025-08-29 17:31:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:46.007435 | orchestrator | 2025-08-29 17:31:46 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:46.009882 | orchestrator | 2025-08-29 17:31:46 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:46.011899 | orchestrator | 2025-08-29 17:31:46 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:46.014493 | orchestrator | 2025-08-29 17:31:46 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:46.014536 | orchestrator | 2025-08-29 17:31:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:49.059062 | orchestrator | 2025-08-29 17:31:49 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:49.059141 | orchestrator | 2025-08-29 17:31:49 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:49.059702 | orchestrator | 2025-08-29 17:31:49 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:49.060319 | orchestrator | 2025-08-29 17:31:49 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:49.061668 | orchestrator | 2025-08-29 17:31:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:52.094687 | orchestrator | 2025-08-29 17:31:52 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:52.096982 | orchestrator | 2025-08-29 17:31:52 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:52.102721 | orchestrator | 2025-08-29 17:31:52 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:52.103670 | orchestrator | 2025-08-29 17:31:52 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:52.104114 | orchestrator | 2025-08-29 17:31:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:55.133593 | orchestrator | 2025-08-29 17:31:55 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:55.134533 | orchestrator | 2025-08-29 17:31:55 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:55.135928 | orchestrator | 2025-08-29 17:31:55 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:55.137584 | orchestrator | 2025-08-29 17:31:55 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:55.137618 | orchestrator | 2025-08-29 17:31:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:31:58.180157 | orchestrator | 2025-08-29 17:31:58 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:31:58.180401 | orchestrator | 2025-08-29 17:31:58 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:31:58.181272 | orchestrator | 2025-08-29 17:31:58 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:31:58.182191 | orchestrator | 2025-08-29 17:31:58 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:31:58.182232 | orchestrator | 2025-08-29 17:31:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:01.209967 | orchestrator | 2025-08-29 17:32:01 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:01.210103 | orchestrator | 2025-08-29 17:32:01 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:01.210905 | orchestrator | 2025-08-29 17:32:01 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:01.211827 | orchestrator | 2025-08-29 17:32:01 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:01.211848 | orchestrator | 2025-08-29 17:32:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:04.253042 | orchestrator | 2025-08-29 17:32:04 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:04.253166 | orchestrator | 2025-08-29 17:32:04 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:04.255290 | orchestrator | 2025-08-29 17:32:04 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:04.256008 | orchestrator | 2025-08-29 17:32:04 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:04.256037 | orchestrator | 2025-08-29 17:32:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:07.311801 | orchestrator | 2025-08-29 17:32:07 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:07.317078 | orchestrator | 2025-08-29 17:32:07 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:07.321265 | orchestrator | 2025-08-29 17:32:07 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:07.322908 | orchestrator | 2025-08-29 17:32:07 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:07.322933 | orchestrator | 2025-08-29 17:32:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:10.354868 | orchestrator | 2025-08-29 17:32:10 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:10.354955 | orchestrator | 2025-08-29 17:32:10 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:10.354973 | orchestrator | 2025-08-29 17:32:10 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:10.355956 | orchestrator | 2025-08-29 17:32:10 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:10.356070 | orchestrator | 2025-08-29 17:32:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:13.398203 | orchestrator | 2025-08-29 17:32:13 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:13.399524 | orchestrator | 2025-08-29 17:32:13 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:13.402621 | orchestrator | 2025-08-29 17:32:13 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:13.403786 | orchestrator | 2025-08-29 17:32:13 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:13.403815 | orchestrator | 2025-08-29 17:32:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:16.446915 | orchestrator | 2025-08-29 17:32:16 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:16.447716 | orchestrator | 2025-08-29 17:32:16 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:16.449703 | orchestrator | 2025-08-29 17:32:16 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:16.451745 | orchestrator | 2025-08-29 17:32:16 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:16.451798 | orchestrator | 2025-08-29 17:32:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:19.494297 | orchestrator | 2025-08-29 17:32:19 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:19.495010 | orchestrator | 2025-08-29 17:32:19 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:19.497133 | orchestrator | 2025-08-29 17:32:19 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:19.497540 | orchestrator | 2025-08-29 17:32:19 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:19.498117 | orchestrator | 2025-08-29 17:32:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:22.538211 | orchestrator | 2025-08-29 17:32:22 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:22.539163 | orchestrator | 2025-08-29 17:32:22 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:22.540848 | orchestrator | 2025-08-29 17:32:22 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:22.541824 | orchestrator | 2025-08-29 17:32:22 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:22.541848 | orchestrator | 2025-08-29 17:32:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:25.598014 | orchestrator | 2025-08-29 17:32:25 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:25.600405 | orchestrator | 2025-08-29 17:32:25 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:25.602657 | orchestrator | 2025-08-29 17:32:25 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:25.604698 | orchestrator | 2025-08-29 17:32:25 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:25.605231 | orchestrator | 2025-08-29 17:32:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:28.646640 | orchestrator | 2025-08-29 17:32:28 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:28.649641 | orchestrator | 2025-08-29 17:32:28 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:28.649679 | orchestrator | 2025-08-29 17:32:28 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:28.650475 | orchestrator | 2025-08-29 17:32:28 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:28.650567 | orchestrator | 2025-08-29 17:32:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:31.688241 | orchestrator | 2025-08-29 17:32:31 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:31.690503 | orchestrator | 2025-08-29 17:32:31 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:31.692643 | orchestrator | 2025-08-29 17:32:31 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:31.694782 | orchestrator | 2025-08-29 17:32:31 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:31.694805 | orchestrator | 2025-08-29 17:32:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:34.740949 | orchestrator | 2025-08-29 17:32:34 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:34.743920 | orchestrator | 2025-08-29 17:32:34 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:34.747420 | orchestrator | 2025-08-29 17:32:34 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:34.749783 | orchestrator | 2025-08-29 17:32:34 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:34.750546 | orchestrator | 2025-08-29 17:32:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:37.785049 | orchestrator | 2025-08-29 17:32:37 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:37.785154 | orchestrator | 2025-08-29 17:32:37 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:37.785170 | orchestrator | 2025-08-29 17:32:37 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:37.785773 | orchestrator | 2025-08-29 17:32:37 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:37.785796 | orchestrator | 2025-08-29 17:32:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:40.820760 | orchestrator | 2025-08-29 17:32:40 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:40.822592 | orchestrator | 2025-08-29 17:32:40 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:40.825458 | orchestrator | 2025-08-29 17:32:40 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:40.827670 | orchestrator | 2025-08-29 17:32:40 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:40.827692 | orchestrator | 2025-08-29 17:32:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:43.869947 | orchestrator | 2025-08-29 17:32:43 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:43.870922 | orchestrator | 2025-08-29 17:32:43 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:43.872019 | orchestrator | 2025-08-29 17:32:43 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:43.873247 | orchestrator | 2025-08-29 17:32:43 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:43.873318 | orchestrator | 2025-08-29 17:32:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:46.915198 | orchestrator | 2025-08-29 17:32:46 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:46.915279 | orchestrator | 2025-08-29 17:32:46 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:46.915542 | orchestrator | 2025-08-29 17:32:46 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:46.919322 | orchestrator | 2025-08-29 17:32:46 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:46.919379 | orchestrator | 2025-08-29 17:32:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:49.956195 | orchestrator | 2025-08-29 17:32:49 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:49.958255 | orchestrator | 2025-08-29 17:32:49 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:49.959706 | orchestrator | 2025-08-29 17:32:49 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:49.961322 | orchestrator | 2025-08-29 17:32:49 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:49.961796 | orchestrator | 2025-08-29 17:32:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:53.008833 | orchestrator | 2025-08-29 17:32:53 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:53.011851 | orchestrator | 2025-08-29 17:32:53 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:53.013814 | orchestrator | 2025-08-29 17:32:53 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:53.016009 | orchestrator | 2025-08-29 17:32:53 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:53.016039 | orchestrator | 2025-08-29 17:32:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:56.045294 | orchestrator | 2025-08-29 17:32:56 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:56.046248 | orchestrator | 2025-08-29 17:32:56 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:56.047097 | orchestrator | 2025-08-29 17:32:56 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:56.047996 | orchestrator | 2025-08-29 17:32:56 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:56.048187 | orchestrator | 2025-08-29 17:32:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:32:59.080018 | orchestrator | 2025-08-29 17:32:59 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state STARTED 2025-08-29 17:32:59.082641 | orchestrator | 2025-08-29 17:32:59 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:32:59.085214 | orchestrator | 2025-08-29 17:32:59 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:32:59.087030 | orchestrator | 2025-08-29 17:32:59 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:32:59.087407 | orchestrator | 2025-08-29 17:32:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:02.121943 | orchestrator | 2025-08-29 17:33:02 | INFO  | Task c7b838d6-dbf1-4627-9e95-bb02f9afaf88 is in state SUCCESS 2025-08-29 17:33:02.123134 | orchestrator | 2025-08-29 17:33:02.123179 | orchestrator | 2025-08-29 17:33:02.123193 | orchestrator | PLAY [Set kolla_action_rabbitmq] *********************************************** 2025-08-29 17:33:02.123208 | orchestrator | 2025-08-29 17:33:02.123262 | orchestrator | TASK [Inform the user about the following task] ******************************** 2025-08-29 17:33:02.123289 | orchestrator | Friday 29 August 2025 17:30:40 +0000 (0:00:00.249) 0:00:00.249 ********* 2025-08-29 17:33:02.123306 | orchestrator | ok: [localhost] => { 2025-08-29 17:33:02.123326 | orchestrator |  "msg": "The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine." 2025-08-29 17:33:02.123343 | orchestrator | } 2025-08-29 17:33:02.123468 | orchestrator | 2025-08-29 17:33:02.123484 | orchestrator | TASK [Check RabbitMQ service] ************************************************** 2025-08-29 17:33:02.123495 | orchestrator | Friday 29 August 2025 17:30:40 +0000 (0:00:00.266) 0:00:00.515 ********* 2025-08-29 17:33:02.123507 | orchestrator | fatal: [localhost]: FAILED! => {"changed": false, "elapsed": 2, "msg": "Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672"} 2025-08-29 17:33:02.123519 | orchestrator | ...ignoring 2025-08-29 17:33:02.123530 | orchestrator | 2025-08-29 17:33:02.123541 | orchestrator | TASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ****** 2025-08-29 17:33:02.123552 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:03.820) 0:00:04.335 ********* 2025-08-29 17:33:02.123563 | orchestrator | skipping: [localhost] 2025-08-29 17:33:02.123573 | orchestrator | 2025-08-29 17:33:02.123584 | orchestrator | TASK [Set kolla_action_rabbitmq = kolla_action_ng] ***************************** 2025-08-29 17:33:02.123595 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:00.116) 0:00:04.452 ********* 2025-08-29 17:33:02.123605 | orchestrator | ok: [localhost] 2025-08-29 17:33:02.123616 | orchestrator | 2025-08-29 17:33:02.123627 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:33:02.123638 | orchestrator | 2025-08-29 17:33:02.123649 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:33:02.123659 | orchestrator | Friday 29 August 2025 17:30:45 +0000 (0:00:00.565) 0:00:05.018 ********* 2025-08-29 17:33:02.123670 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:33:02.123681 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:33:02.123691 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:33:02.123702 | orchestrator | 2025-08-29 17:33:02.123713 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:33:02.123723 | orchestrator | Friday 29 August 2025 17:30:45 +0000 (0:00:00.791) 0:00:05.810 ********* 2025-08-29 17:33:02.123734 | orchestrator | ok: [testbed-node-0] => (item=enable_rabbitmq_True) 2025-08-29 17:33:02.123745 | orchestrator | ok: [testbed-node-1] => (item=enable_rabbitmq_True) 2025-08-29 17:33:02.123756 | orchestrator | ok: [testbed-node-2] => (item=enable_rabbitmq_True) 2025-08-29 17:33:02.123766 | orchestrator | 2025-08-29 17:33:02.123777 | orchestrator | PLAY [Apply role rabbitmq] ***************************************************** 2025-08-29 17:33:02.123787 | orchestrator | 2025-08-29 17:33:02.123798 | orchestrator | TASK [rabbitmq : include_tasks] ************************************************ 2025-08-29 17:33:02.123809 | orchestrator | Friday 29 August 2025 17:30:46 +0000 (0:00:00.806) 0:00:06.617 ********* 2025-08-29 17:33:02.123820 | orchestrator | included: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:33:02.123831 | orchestrator | 2025-08-29 17:33:02.123842 | orchestrator | TASK [rabbitmq : Get container facts] ****************************************** 2025-08-29 17:33:02.123852 | orchestrator | Friday 29 August 2025 17:30:47 +0000 (0:00:00.543) 0:00:07.160 ********* 2025-08-29 17:33:02.123863 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:33:02.123873 | orchestrator | 2025-08-29 17:33:02.123884 | orchestrator | TASK [rabbitmq : Get current RabbitMQ version] ********************************* 2025-08-29 17:33:02.123895 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:00.999) 0:00:08.160 ********* 2025-08-29 17:33:02.123905 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.123916 | orchestrator | 2025-08-29 17:33:02.123927 | orchestrator | TASK [rabbitmq : Get new RabbitMQ version] ************************************* 2025-08-29 17:33:02.123938 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:00.328) 0:00:08.488 ********* 2025-08-29 17:33:02.123962 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.123973 | orchestrator | 2025-08-29 17:33:02.123983 | orchestrator | TASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ****** 2025-08-29 17:33:02.123995 | orchestrator | Friday 29 August 2025 17:30:48 +0000 (0:00:00.332) 0:00:08.821 ********* 2025-08-29 17:33:02.124006 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.124016 | orchestrator | 2025-08-29 17:33:02.124027 | orchestrator | TASK [rabbitmq : Catch when RabbitMQ is being downgraded] ********************** 2025-08-29 17:33:02.124038 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:00.294) 0:00:09.116 ********* 2025-08-29 17:33:02.124049 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.124059 | orchestrator | 2025-08-29 17:33:02.124070 | orchestrator | TASK [rabbitmq : include_tasks] ************************************************ 2025-08-29 17:33:02.124080 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:00.505) 0:00:09.621 ********* 2025-08-29 17:33:02.124091 | orchestrator | included: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:33:02.124102 | orchestrator | 2025-08-29 17:33:02.124113 | orchestrator | TASK [rabbitmq : Get container facts] ****************************************** 2025-08-29 17:33:02.124123 | orchestrator | Friday 29 August 2025 17:30:50 +0000 (0:00:00.696) 0:00:10.317 ********* 2025-08-29 17:33:02.124134 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:33:02.124145 | orchestrator | 2025-08-29 17:33:02.124155 | orchestrator | TASK [rabbitmq : List RabbitMQ policies] *************************************** 2025-08-29 17:33:02.124166 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.888) 0:00:11.206 ********* 2025-08-29 17:33:02.124177 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.124187 | orchestrator | 2025-08-29 17:33:02.124198 | orchestrator | TASK [rabbitmq : Remove ha-all policy from RabbitMQ] *************************** 2025-08-29 17:33:02.124209 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.282) 0:00:11.489 ********* 2025-08-29 17:33:02.124220 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.124230 | orchestrator | 2025-08-29 17:33:02.124254 | orchestrator | TASK [rabbitmq : Ensuring config directories exist] **************************** 2025-08-29 17:33:02.124266 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.325) 0:00:11.815 ********* 2025-08-29 17:33:02.124407 | orchestrator | changed: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.124446 | orchestrator | changed: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.124469 | orchestrator | changed: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.124482 | orchestrator | 2025-08-29 17:33:02.124494 | orchestrator | TASK [rabbitmq : Copying over config.json files for services] ****************** 2025-08-29 17:33:02.124505 | orchestrator | Friday 29 August 2025 17:30:52 +0000 (0:00:00.823) 0:00:12.638 ********* 2025-08-29 17:33:02.124530 | orchestrator | changed: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.124548 | orchestrator | changed: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.124561 | orchestrator | changed: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.124579 | orchestrator | 2025-08-29 17:33:02.124590 | orchestrator | TASK [rabbitmq : Copying over rabbitmq-env.conf] ******************************* 2025-08-29 17:33:02.124601 | orchestrator | Friday 29 August 2025 17:30:54 +0000 (0:00:01.961) 0:00:14.600 ********* 2025-08-29 17:33:02.124611 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2) 2025-08-29 17:33:02.124622 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2) 2025-08-29 17:33:02.124633 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2) 2025-08-29 17:33:02.124644 | orchestrator | 2025-08-29 17:33:02.124655 | orchestrator | TASK [rabbitmq : Copying over rabbitmq.conf] *********************************** 2025-08-29 17:33:02.124666 | orchestrator | Friday 29 August 2025 17:30:56 +0000 (0:00:02.108) 0:00:16.708 ********* 2025-08-29 17:33:02.124677 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2) 2025-08-29 17:33:02.124687 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2) 2025-08-29 17:33:02.124698 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2) 2025-08-29 17:33:02.124709 | orchestrator | 2025-08-29 17:33:02.124720 | orchestrator | TASK [rabbitmq : Copying over erl_inetrc] ************************************** 2025-08-29 17:33:02.124730 | orchestrator | Friday 29 August 2025 17:31:00 +0000 (0:00:03.435) 0:00:20.144 ********* 2025-08-29 17:33:02.124741 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2) 2025-08-29 17:33:02.124752 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2) 2025-08-29 17:33:02.124763 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2) 2025-08-29 17:33:02.124774 | orchestrator | 2025-08-29 17:33:02.124792 | orchestrator | TASK [rabbitmq : Copying over advanced.config] ********************************* 2025-08-29 17:33:02.124803 | orchestrator | Friday 29 August 2025 17:31:01 +0000 (0:00:01.497) 0:00:21.642 ********* 2025-08-29 17:33:02.124814 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2) 2025-08-29 17:33:02.124825 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2) 2025-08-29 17:33:02.124836 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2) 2025-08-29 17:33:02.124847 | orchestrator | 2025-08-29 17:33:02.124858 | orchestrator | TASK [rabbitmq : Copying over definitions.json] ******************************** 2025-08-29 17:33:02.124869 | orchestrator | Friday 29 August 2025 17:31:03 +0000 (0:00:01.843) 0:00:23.486 ********* 2025-08-29 17:33:02.124879 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2) 2025-08-29 17:33:02.124890 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2) 2025-08-29 17:33:02.124901 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2) 2025-08-29 17:33:02.124918 | orchestrator | 2025-08-29 17:33:02.124929 | orchestrator | TASK [rabbitmq : Copying over enabled_plugins] ********************************* 2025-08-29 17:33:02.124940 | orchestrator | Friday 29 August 2025 17:31:04 +0000 (0:00:01.352) 0:00:24.838 ********* 2025-08-29 17:33:02.124951 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2) 2025-08-29 17:33:02.124962 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2) 2025-08-29 17:33:02.124973 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2) 2025-08-29 17:33:02.124984 | orchestrator | 2025-08-29 17:33:02.124995 | orchestrator | TASK [rabbitmq : include_tasks] ************************************************ 2025-08-29 17:33:02.125010 | orchestrator | Friday 29 August 2025 17:31:07 +0000 (0:00:02.311) 0:00:27.150 ********* 2025-08-29 17:33:02.125021 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.125032 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:33:02.125043 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:33:02.125054 | orchestrator | 2025-08-29 17:33:02.125064 | orchestrator | TASK [rabbitmq : Check rabbitmq containers] ************************************ 2025-08-29 17:33:02.125115 | orchestrator | Friday 29 August 2025 17:31:08 +0000 (0:00:00.950) 0:00:28.100 ********* 2025-08-29 17:33:02.125129 | orchestrator | changed: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.125142 | orchestrator | changed: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.125164 | orchestrator | changed: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:33:02.125185 | orchestrator | 2025-08-29 17:33:02.125196 | orchestrator | TASK [rabbitmq : Creating rabbitmq volume] ************************************* 2025-08-29 17:33:02.125207 | orchestrator | Friday 29 August 2025 17:31:09 +0000 (0:00:01.441) 0:00:29.542 ********* 2025-08-29 17:33:02.125218 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:33:02.125229 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:33:02.125240 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:33:02.125251 | orchestrator | 2025-08-29 17:33:02.125262 | orchestrator | TASK [rabbitmq : Running RabbitMQ bootstrap container] ************************* 2025-08-29 17:33:02.125273 | orchestrator | Friday 29 August 2025 17:31:10 +0000 (0:00:01.154) 0:00:30.696 ********* 2025-08-29 17:33:02.125283 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:33:02.125294 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:33:02.125310 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:33:02.125321 | orchestrator | 2025-08-29 17:33:02.125332 | orchestrator | RUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************ 2025-08-29 17:33:02.125343 | orchestrator | Friday 29 August 2025 17:31:18 +0000 (0:00:07.696) 0:00:38.393 ********* 2025-08-29 17:33:02.125385 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:33:02.125399 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:33:02.125410 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:33:02.125420 | orchestrator | 2025-08-29 17:33:02.125431 | orchestrator | PLAY [Restart rabbitmq services] *********************************************** 2025-08-29 17:33:02.125442 | orchestrator | 2025-08-29 17:33:02.125453 | orchestrator | TASK [rabbitmq : Get info on RabbitMQ container] ******************************* 2025-08-29 17:33:02.125463 | orchestrator | Friday 29 August 2025 17:31:18 +0000 (0:00:00.274) 0:00:38.667 ********* 2025-08-29 17:33:02.125474 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:33:02.125485 | orchestrator | 2025-08-29 17:33:02.125496 | orchestrator | TASK [rabbitmq : Put RabbitMQ node into maintenance mode] ********************** 2025-08-29 17:33:02.125507 | orchestrator | Friday 29 August 2025 17:31:19 +0000 (0:00:00.567) 0:00:39.234 ********* 2025-08-29 17:33:02.125517 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:33:02.125528 | orchestrator | 2025-08-29 17:33:02.125539 | orchestrator | TASK [rabbitmq : Restart rabbitmq container] *********************************** 2025-08-29 17:33:02.125550 | orchestrator | Friday 29 August 2025 17:31:19 +0000 (0:00:00.319) 0:00:39.554 ********* 2025-08-29 17:33:02.125560 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:33:02.125571 | orchestrator | 2025-08-29 17:33:02.125582 | orchestrator | TASK [rabbitmq : Waiting for rabbitmq to start] ******************************** 2025-08-29 17:33:02.125593 | orchestrator | Friday 29 August 2025 17:31:26 +0000 (0:00:07.024) 0:00:46.579 ********* 2025-08-29 17:33:02.125603 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:33:02.125614 | orchestrator | 2025-08-29 17:33:02.125625 | orchestrator | PLAY [Restart rabbitmq services] *********************************************** 2025-08-29 17:33:02.125635 | orchestrator | 2025-08-29 17:33:02.125646 | orchestrator | TASK [rabbitmq : Get info on RabbitMQ container] ******************************* 2025-08-29 17:33:02.125657 | orchestrator | Friday 29 August 2025 17:32:18 +0000 (0:00:51.873) 0:01:38.452 ********* 2025-08-29 17:33:02.125668 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:33:02.125678 | orchestrator | 2025-08-29 17:33:02.125689 | orchestrator | TASK [rabbitmq : Put RabbitMQ node into maintenance mode] ********************** 2025-08-29 17:33:02.125700 | orchestrator | Friday 29 August 2025 17:32:19 +0000 (0:00:00.596) 0:01:39.049 ********* 2025-08-29 17:33:02.125711 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:33:02.125728 | orchestrator | 2025-08-29 17:33:02.125739 | orchestrator | TASK [rabbitmq : Restart rabbitmq container] *********************************** 2025-08-29 17:33:02.125750 | orchestrator | Friday 29 August 2025 17:32:19 +0000 (0:00:00.399) 0:01:39.448 ********* 2025-08-29 17:33:02.125761 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:33:02.125771 | orchestrator | 2025-08-29 17:33:02.125782 | orchestrator | TASK [rabbitmq : Waiting for rabbitmq to start] ******************************** 2025-08-29 17:33:02.125793 | orchestrator | Friday 29 August 2025 17:32:21 +0000 (0:00:01.809) 0:01:41.258 ********* 2025-08-29 17:33:02.125804 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:33:02.125814 | orchestrator | 2025-08-29 17:33:02.125825 | orchestrator | PLAY [Restart rabbitmq services] *********************************************** 2025-08-29 17:33:02.125836 | orchestrator | 2025-08-29 17:33:02.125847 | orchestrator | TASK [rabbitmq : Get info on RabbitMQ container] ******************************* 2025-08-29 17:33:02.125857 | orchestrator | Friday 29 August 2025 17:32:37 +0000 (0:00:16.090) 0:01:57.348 ********* 2025-08-29 17:33:02.125868 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:33:02.125879 | orchestrator | 2025-08-29 17:33:02.125889 | orchestrator | TASK [rabbitmq : Put RabbitMQ node into maintenance mode] ********************** 2025-08-29 17:33:02.125900 | orchestrator | Friday 29 August 2025 17:32:38 +0000 (0:00:00.617) 0:01:57.966 ********* 2025-08-29 17:33:02.125911 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:33:02.125922 | orchestrator | 2025-08-29 17:33:02.125933 | orchestrator | TASK [rabbitmq : Restart rabbitmq container] *********************************** 2025-08-29 17:33:02.125949 | orchestrator | Friday 29 August 2025 17:32:38 +0000 (0:00:00.249) 0:01:58.216 ********* 2025-08-29 17:33:02.125961 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:33:02.125972 | orchestrator | 2025-08-29 17:33:02.125983 | orchestrator | TASK [rabbitmq : Waiting for rabbitmq to start] ******************************** 2025-08-29 17:33:02.125993 | orchestrator | Friday 29 August 2025 17:32:44 +0000 (0:00:06.603) 0:02:04.820 ********* 2025-08-29 17:33:02.126004 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:33:02.126072 | orchestrator | 2025-08-29 17:33:02.126087 | orchestrator | PLAY [Apply rabbitmq post-configuration] *************************************** 2025-08-29 17:33:02.126099 | orchestrator | 2025-08-29 17:33:02.126109 | orchestrator | TASK [Include rabbitmq post-deploy.yml] **************************************** 2025-08-29 17:33:02.126120 | orchestrator | Friday 29 August 2025 17:32:57 +0000 (0:00:12.249) 0:02:17.069 ********* 2025-08-29 17:33:02.126131 | orchestrator | included: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:33:02.126142 | orchestrator | 2025-08-29 17:33:02.126153 | orchestrator | TASK [rabbitmq : Enable all stable feature flags] ****************************** 2025-08-29 17:33:02.126164 | orchestrator | Friday 29 August 2025 17:32:57 +0000 (0:00:00.736) 0:02:17.805 ********* 2025-08-29 17:33:02.126175 | orchestrator | [WARNING]: Could not match supplied host pattern, ignoring: 2025-08-29 17:33:02.126185 | orchestrator | enable_outward_rabbitmq_True 2025-08-29 17:33:02.126196 | orchestrator | [WARNING]: Could not match supplied host pattern, ignoring: 2025-08-29 17:33:02.126207 | orchestrator | outward_rabbitmq_restart 2025-08-29 17:33:02.126218 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:33:02.126229 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:33:02.126239 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:33:02.126250 | orchestrator | 2025-08-29 17:33:02.126262 | orchestrator | PLAY [Apply role rabbitmq (outward)] ******************************************* 2025-08-29 17:33:02.126272 | orchestrator | skipping: no hosts matched 2025-08-29 17:33:02.126283 | orchestrator | 2025-08-29 17:33:02.126294 | orchestrator | PLAY [Restart rabbitmq (outward) services] ************************************* 2025-08-29 17:33:02.126305 | orchestrator | skipping: no hosts matched 2025-08-29 17:33:02.126316 | orchestrator | 2025-08-29 17:33:02.126327 | orchestrator | PLAY [Apply rabbitmq (outward) post-configuration] ***************************** 2025-08-29 17:33:02.126337 | orchestrator | skipping: no hosts matched 2025-08-29 17:33:02.126348 | orchestrator | 2025-08-29 17:33:02.126425 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:33:02.126438 | orchestrator | localhost : ok=3  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=1  2025-08-29 17:33:02.126457 | orchestrator | testbed-node-0 : ok=23  changed=14  unreachable=0 failed=0 skipped=8  rescued=0 ignored=0 2025-08-29 17:33:02.126468 | orchestrator | testbed-node-1 : ok=21  changed=14  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:33:02.126479 | orchestrator | testbed-node-2 : ok=21  changed=14  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:33:02.126490 | orchestrator | 2025-08-29 17:33:02.126501 | orchestrator | 2025-08-29 17:33:02.126512 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:33:02.126521 | orchestrator | Friday 29 August 2025 17:33:00 +0000 (0:00:02.581) 0:02:20.387 ********* 2025-08-29 17:33:02.126531 | orchestrator | =============================================================================== 2025-08-29 17:33:02.126541 | orchestrator | rabbitmq : Waiting for rabbitmq to start ------------------------------- 80.21s 2025-08-29 17:33:02.126550 | orchestrator | rabbitmq : Restart rabbitmq container ---------------------------------- 15.44s 2025-08-29 17:33:02.126560 | orchestrator | rabbitmq : Running RabbitMQ bootstrap container ------------------------- 7.70s 2025-08-29 17:33:02.126569 | orchestrator | Check RabbitMQ service -------------------------------------------------- 3.82s 2025-08-29 17:33:02.126579 | orchestrator | rabbitmq : Copying over rabbitmq.conf ----------------------------------- 3.44s 2025-08-29 17:33:02.126588 | orchestrator | rabbitmq : Enable all stable feature flags ------------------------------ 2.58s 2025-08-29 17:33:02.126598 | orchestrator | rabbitmq : Copying over enabled_plugins --------------------------------- 2.31s 2025-08-29 17:33:02.126608 | orchestrator | rabbitmq : Copying over rabbitmq-env.conf ------------------------------- 2.11s 2025-08-29 17:33:02.126617 | orchestrator | rabbitmq : Copying over config.json files for services ------------------ 1.96s 2025-08-29 17:33:02.126627 | orchestrator | rabbitmq : Copying over advanced.config --------------------------------- 1.84s 2025-08-29 17:33:02.126636 | orchestrator | rabbitmq : Get info on RabbitMQ container ------------------------------- 1.78s 2025-08-29 17:33:02.126645 | orchestrator | rabbitmq : Copying over erl_inetrc -------------------------------------- 1.50s 2025-08-29 17:33:02.126655 | orchestrator | rabbitmq : Check rabbitmq containers ------------------------------------ 1.44s 2025-08-29 17:33:02.126664 | orchestrator | rabbitmq : Copying over definitions.json -------------------------------- 1.35s 2025-08-29 17:33:02.126674 | orchestrator | rabbitmq : Creating rabbitmq volume ------------------------------------- 1.15s 2025-08-29 17:33:02.126683 | orchestrator | rabbitmq : Get container facts ------------------------------------------ 1.00s 2025-08-29 17:33:02.126693 | orchestrator | rabbitmq : Put RabbitMQ node into maintenance mode ---------------------- 0.97s 2025-08-29 17:33:02.126702 | orchestrator | rabbitmq : include_tasks ------------------------------------------------ 0.95s 2025-08-29 17:33:02.126712 | orchestrator | rabbitmq : Get container facts ------------------------------------------ 0.89s 2025-08-29 17:33:02.126721 | orchestrator | rabbitmq : Ensuring config directories exist ---------------------------- 0.82s 2025-08-29 17:33:02.126736 | orchestrator | 2025-08-29 17:33:02 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:02.126746 | orchestrator | 2025-08-29 17:33:02 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:02.127051 | orchestrator | 2025-08-29 17:33:02 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:02.127149 | orchestrator | 2025-08-29 17:33:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:05.164645 | orchestrator | 2025-08-29 17:33:05 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:05.165025 | orchestrator | 2025-08-29 17:33:05 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:05.166853 | orchestrator | 2025-08-29 17:33:05 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:05.167135 | orchestrator | 2025-08-29 17:33:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:08.204640 | orchestrator | 2025-08-29 17:33:08 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:08.204834 | orchestrator | 2025-08-29 17:33:08 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:08.205598 | orchestrator | 2025-08-29 17:33:08 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:08.205624 | orchestrator | 2025-08-29 17:33:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:11.241442 | orchestrator | 2025-08-29 17:33:11 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:11.242599 | orchestrator | 2025-08-29 17:33:11 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:11.244580 | orchestrator | 2025-08-29 17:33:11 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:11.244874 | orchestrator | 2025-08-29 17:33:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:14.291926 | orchestrator | 2025-08-29 17:33:14 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:14.292036 | orchestrator | 2025-08-29 17:33:14 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:14.296539 | orchestrator | 2025-08-29 17:33:14 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:14.296579 | orchestrator | 2025-08-29 17:33:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:17.324699 | orchestrator | 2025-08-29 17:33:17 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:17.326607 | orchestrator | 2025-08-29 17:33:17 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:17.328996 | orchestrator | 2025-08-29 17:33:17 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:17.330739 | orchestrator | 2025-08-29 17:33:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:20.366574 | orchestrator | 2025-08-29 17:33:20 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:20.366993 | orchestrator | 2025-08-29 17:33:20 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:20.367950 | orchestrator | 2025-08-29 17:33:20 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:20.368181 | orchestrator | 2025-08-29 17:33:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:23.413280 | orchestrator | 2025-08-29 17:33:23 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:23.414518 | orchestrator | 2025-08-29 17:33:23 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:23.416133 | orchestrator | 2025-08-29 17:33:23 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:23.416158 | orchestrator | 2025-08-29 17:33:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:26.460923 | orchestrator | 2025-08-29 17:33:26 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:26.463280 | orchestrator | 2025-08-29 17:33:26 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:26.465846 | orchestrator | 2025-08-29 17:33:26 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:26.465907 | orchestrator | 2025-08-29 17:33:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:29.497034 | orchestrator | 2025-08-29 17:33:29 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:29.498423 | orchestrator | 2025-08-29 17:33:29 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:29.500072 | orchestrator | 2025-08-29 17:33:29 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:29.500110 | orchestrator | 2025-08-29 17:33:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:32.533040 | orchestrator | 2025-08-29 17:33:32 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:32.535187 | orchestrator | 2025-08-29 17:33:32 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:32.536004 | orchestrator | 2025-08-29 17:33:32 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:32.536033 | orchestrator | 2025-08-29 17:33:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:35.562216 | orchestrator | 2025-08-29 17:33:35 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:35.562487 | orchestrator | 2025-08-29 17:33:35 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:35.563339 | orchestrator | 2025-08-29 17:33:35 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:35.563389 | orchestrator | 2025-08-29 17:33:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:38.594331 | orchestrator | 2025-08-29 17:33:38 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:38.595816 | orchestrator | 2025-08-29 17:33:38 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:38.600335 | orchestrator | 2025-08-29 17:33:38 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:38.600495 | orchestrator | 2025-08-29 17:33:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:41.637976 | orchestrator | 2025-08-29 17:33:41 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:41.641475 | orchestrator | 2025-08-29 17:33:41 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:41.643449 | orchestrator | 2025-08-29 17:33:41 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:41.643504 | orchestrator | 2025-08-29 17:33:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:44.682709 | orchestrator | 2025-08-29 17:33:44 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:44.684606 | orchestrator | 2025-08-29 17:33:44 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:44.686209 | orchestrator | 2025-08-29 17:33:44 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:44.686536 | orchestrator | 2025-08-29 17:33:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:47.729318 | orchestrator | 2025-08-29 17:33:47 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:47.731796 | orchestrator | 2025-08-29 17:33:47 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:47.733971 | orchestrator | 2025-08-29 17:33:47 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:47.734441 | orchestrator | 2025-08-29 17:33:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:50.787853 | orchestrator | 2025-08-29 17:33:50 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:50.788444 | orchestrator | 2025-08-29 17:33:50 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:50.789200 | orchestrator | 2025-08-29 17:33:50 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:50.789282 | orchestrator | 2025-08-29 17:33:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:53.831896 | orchestrator | 2025-08-29 17:33:53 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:53.832559 | orchestrator | 2025-08-29 17:33:53 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:53.835381 | orchestrator | 2025-08-29 17:33:53 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:53.835793 | orchestrator | 2025-08-29 17:33:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:56.886350 | orchestrator | 2025-08-29 17:33:56 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:56.887123 | orchestrator | 2025-08-29 17:33:56 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:56.888668 | orchestrator | 2025-08-29 17:33:56 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:56.888717 | orchestrator | 2025-08-29 17:33:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:33:59.944952 | orchestrator | 2025-08-29 17:33:59 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:33:59.947413 | orchestrator | 2025-08-29 17:33:59 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:33:59.949869 | orchestrator | 2025-08-29 17:33:59 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:33:59.950250 | orchestrator | 2025-08-29 17:33:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:03.005056 | orchestrator | 2025-08-29 17:34:03 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:34:03.008542 | orchestrator | 2025-08-29 17:34:03 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:03.011913 | orchestrator | 2025-08-29 17:34:03 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:03.011950 | orchestrator | 2025-08-29 17:34:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:06.049448 | orchestrator | 2025-08-29 17:34:06 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state STARTED 2025-08-29 17:34:06.049567 | orchestrator | 2025-08-29 17:34:06 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:06.050449 | orchestrator | 2025-08-29 17:34:06 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:06.050548 | orchestrator | 2025-08-29 17:34:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:09.101955 | orchestrator | 2025-08-29 17:34:09 | INFO  | Task c70a8217-d6f7-4ec8-92fc-38b297a3e06f is in state SUCCESS 2025-08-29 17:34:09.103397 | orchestrator | 2025-08-29 17:34:09.103481 | orchestrator | 2025-08-29 17:34:09.103494 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:34:09.103503 | orchestrator | 2025-08-29 17:34:09.103511 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:34:09.103520 | orchestrator | Friday 29 August 2025 17:31:33 +0000 (0:00:00.169) 0:00:00.169 ********* 2025-08-29 17:34:09.103550 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:34:09.103559 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:34:09.103566 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:34:09.103574 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.103581 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.103674 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.103682 | orchestrator | 2025-08-29 17:34:09.103689 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:34:09.103696 | orchestrator | Friday 29 August 2025 17:31:34 +0000 (0:00:00.797) 0:00:00.967 ********* 2025-08-29 17:34:09.103704 | orchestrator | ok: [testbed-node-3] => (item=enable_ovn_True) 2025-08-29 17:34:09.103712 | orchestrator | ok: [testbed-node-4] => (item=enable_ovn_True) 2025-08-29 17:34:09.103719 | orchestrator | ok: [testbed-node-5] => (item=enable_ovn_True) 2025-08-29 17:34:09.103726 | orchestrator | ok: [testbed-node-0] => (item=enable_ovn_True) 2025-08-29 17:34:09.103734 | orchestrator | ok: [testbed-node-1] => (item=enable_ovn_True) 2025-08-29 17:34:09.103741 | orchestrator | ok: [testbed-node-2] => (item=enable_ovn_True) 2025-08-29 17:34:09.103748 | orchestrator | 2025-08-29 17:34:09.103755 | orchestrator | PLAY [Apply role ovn-controller] *********************************************** 2025-08-29 17:34:09.103762 | orchestrator | 2025-08-29 17:34:09.103770 | orchestrator | TASK [ovn-controller : include_tasks] ****************************************** 2025-08-29 17:34:09.103777 | orchestrator | Friday 29 August 2025 17:31:35 +0000 (0:00:01.366) 0:00:02.333 ********* 2025-08-29 17:34:09.103785 | orchestrator | included: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:34:09.103794 | orchestrator | 2025-08-29 17:34:09.103801 | orchestrator | TASK [ovn-controller : Ensuring config directories exist] ********************** 2025-08-29 17:34:09.103808 | orchestrator | Friday 29 August 2025 17:31:36 +0000 (0:00:01.342) 0:00:03.675 ********* 2025-08-29 17:34:09.103817 | orchestrator | changed: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103826 | orchestrator | changed: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103834 | orchestrator | changed: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103842 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103862 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103897 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103906 | orchestrator | 2025-08-29 17:34:09.103914 | orchestrator | TASK [ovn-controller : Copying over config.json files for services] ************ 2025-08-29 17:34:09.103921 | orchestrator | Friday 29 August 2025 17:31:37 +0000 (0:00:00.962) 0:00:04.638 ********* 2025-08-29 17:34:09.103929 | orchestrator | changed: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103936 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103944 | orchestrator | changed: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103952 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103959 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103967 | orchestrator | changed: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.103974 | orchestrator | 2025-08-29 17:34:09.103982 | orchestrator | TASK [ovn-controller : Ensuring systemd override directory exists] ************* 2025-08-29 17:34:09.103989 | orchestrator | Friday 29 August 2025 17:31:39 +0000 (0:00:01.418) 0:00:06.057 ********* 2025-08-29 17:34:09.104001 | orchestrator | changed: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104013 | orchestrator | changed: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104026 | orchestrator | changed: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104034 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104041 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104049 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104056 | orchestrator | 2025-08-29 17:34:09.104063 | orchestrator | TASK [ovn-controller : Copying over systemd override] ************************** 2025-08-29 17:34:09.104071 | orchestrator | Friday 29 August 2025 17:31:40 +0000 (0:00:01.454) 0:00:07.511 ********* 2025-08-29 17:34:09.104078 | orchestrator | changed: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104086 | orchestrator | changed: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104093 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104106 | orchestrator | changed: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104117 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104129 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104137 | orchestrator | 2025-08-29 17:34:09.104144 | orchestrator | TASK [ovn-controller : Check ovn-controller containers] ************************ 2025-08-29 17:34:09.104151 | orchestrator | Friday 29 August 2025 17:31:42 +0000 (0:00:01.526) 0:00:09.037 ********* 2025-08-29 17:34:09.104159 | orchestrator | changed: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104166 | orchestrator | changed: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104174 | orchestrator | changed: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104181 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104189 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104201 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.104208 | orchestrator | 2025-08-29 17:34:09.104216 | orchestrator | TASK [ovn-controller : Create br-int bridge on OpenvSwitch] ******************** 2025-08-29 17:34:09.104223 | orchestrator | Friday 29 August 2025 17:31:43 +0000 (0:00:01.338) 0:00:10.376 ********* 2025-08-29 17:34:09.104230 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:34:09.104238 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:34:09.104246 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.104253 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:34:09.104260 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.104268 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.104275 | orchestrator | 2025-08-29 17:34:09.104285 | orchestrator | TASK [ovn-controller : Configure OVN in OVSDB] ********************************* 2025-08-29 17:34:09.104293 | orchestrator | Friday 29 August 2025 17:31:46 +0000 (0:00:02.460) 0:00:12.837 ********* 2025-08-29 17:34:09.104300 | orchestrator | changed: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'}) 2025-08-29 17:34:09.104308 | orchestrator | changed: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'}) 2025-08-29 17:34:09.104315 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'}) 2025-08-29 17:34:09.104326 | orchestrator | changed: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'}) 2025-08-29 17:34:09.104334 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'}) 2025-08-29 17:34:09.104341 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'}) 2025-08-29 17:34:09.104348 | orchestrator | changed: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'}) 2025-08-29 17:34:09.104356 | orchestrator | changed: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'}) 2025-08-29 17:34:09.104391 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'}) 2025-08-29 17:34:09.104399 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'}) 2025-08-29 17:34:09.104406 | orchestrator | changed: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'}) 2025-08-29 17:34:09.104414 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'}) 2025-08-29 17:34:09.104421 | orchestrator | changed: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}) 2025-08-29 17:34:09.104430 | orchestrator | changed: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}) 2025-08-29 17:34:09.104438 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}) 2025-08-29 17:34:09.104446 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}) 2025-08-29 17:34:09.104453 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}) 2025-08-29 17:34:09.104471 | orchestrator | changed: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'}) 2025-08-29 17:34:09.104479 | orchestrator | changed: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'}) 2025-08-29 17:34:09.104487 | orchestrator | changed: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'}) 2025-08-29 17:34:09.104494 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'}) 2025-08-29 17:34:09.104502 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'}) 2025-08-29 17:34:09.104509 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'}) 2025-08-29 17:34:09.104516 | orchestrator | changed: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'}) 2025-08-29 17:34:09.104524 | orchestrator | changed: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'}) 2025-08-29 17:34:09.104531 | orchestrator | changed: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'}) 2025-08-29 17:34:09.104538 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'}) 2025-08-29 17:34:09.104545 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'}) 2025-08-29 17:34:09.104553 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'}) 2025-08-29 17:34:09.104560 | orchestrator | changed: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'}) 2025-08-29 17:34:09.104567 | orchestrator | changed: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False}) 2025-08-29 17:34:09.104575 | orchestrator | changed: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False}) 2025-08-29 17:34:09.104582 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False}) 2025-08-29 17:34:09.104589 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False}) 2025-08-29 17:34:09.104597 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False}) 2025-08-29 17:34:09.104604 | orchestrator | changed: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False}) 2025-08-29 17:34:09.104611 | orchestrator | ok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'}) 2025-08-29 17:34:09.104623 | orchestrator | ok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'}) 2025-08-29 17:34:09.104630 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'}) 2025-08-29 17:34:09.104638 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'}) 2025-08-29 17:34:09.104649 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'}) 2025-08-29 17:34:09.104657 | orchestrator | ok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'}) 2025-08-29 17:34:09.104665 | orchestrator | changed: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'}) 2025-08-29 17:34:09.104673 | orchestrator | changed: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'}) 2025-08-29 17:34:09.104680 | orchestrator | ok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'}) 2025-08-29 17:34:09.104687 | orchestrator | ok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'}) 2025-08-29 17:34:09.104699 | orchestrator | ok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'}) 2025-08-29 17:34:09.104707 | orchestrator | ok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'}) 2025-08-29 17:34:09.104714 | orchestrator | changed: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'}) 2025-08-29 17:34:09.104721 | orchestrator | changed: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'}) 2025-08-29 17:34:09.104729 | orchestrator | ok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'}) 2025-08-29 17:34:09.104736 | orchestrator | changed: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'}) 2025-08-29 17:34:09.104744 | orchestrator | changed: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'}) 2025-08-29 17:34:09.104751 | orchestrator | ok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'}) 2025-08-29 17:34:09.104758 | orchestrator | 2025-08-29 17:34:09.104765 | orchestrator | TASK [ovn-controller : Flush handlers] ***************************************** 2025-08-29 17:34:09.104773 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:19.549) 0:00:32.386 ********* 2025-08-29 17:34:09.104780 | orchestrator | 2025-08-29 17:34:09.104788 | orchestrator | TASK [ovn-controller : Flush handlers] ***************************************** 2025-08-29 17:34:09.104795 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.235) 0:00:32.621 ********* 2025-08-29 17:34:09.104802 | orchestrator | 2025-08-29 17:34:09.104810 | orchestrator | TASK [ovn-controller : Flush handlers] ***************************************** 2025-08-29 17:34:09.104817 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.064) 0:00:32.686 ********* 2025-08-29 17:34:09.104824 | orchestrator | 2025-08-29 17:34:09.104831 | orchestrator | TASK [ovn-controller : Flush handlers] ***************************************** 2025-08-29 17:34:09.104839 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.063) 0:00:32.749 ********* 2025-08-29 17:34:09.104846 | orchestrator | 2025-08-29 17:34:09.104853 | orchestrator | TASK [ovn-controller : Flush handlers] ***************************************** 2025-08-29 17:34:09.104861 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.060) 0:00:32.809 ********* 2025-08-29 17:34:09.104868 | orchestrator | 2025-08-29 17:34:09.104875 | orchestrator | TASK [ovn-controller : Flush handlers] ***************************************** 2025-08-29 17:34:09.104882 | orchestrator | Friday 29 August 2025 17:32:06 +0000 (0:00:00.063) 0:00:32.873 ********* 2025-08-29 17:34:09.104890 | orchestrator | 2025-08-29 17:34:09.104897 | orchestrator | RUNNING HANDLER [ovn-controller : Reload systemd config] *********************** 2025-08-29 17:34:09.104904 | orchestrator | Friday 29 August 2025 17:32:06 +0000 (0:00:00.061) 0:00:32.935 ********* 2025-08-29 17:34:09.104928 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.104936 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:34:09.104943 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:34:09.104950 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:34:09.104958 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.104965 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.104972 | orchestrator | 2025-08-29 17:34:09.104979 | orchestrator | RUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************ 2025-08-29 17:34:09.104987 | orchestrator | Friday 29 August 2025 17:32:07 +0000 (0:00:01.779) 0:00:34.715 ********* 2025-08-29 17:34:09.104994 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.105001 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:34:09.105009 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:34:09.105016 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.105023 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.105030 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:34:09.105038 | orchestrator | 2025-08-29 17:34:09.105045 | orchestrator | PLAY [Apply role ovn-db] ******************************************************* 2025-08-29 17:34:09.105057 | orchestrator | 2025-08-29 17:34:09.105064 | orchestrator | TASK [ovn-db : include_tasks] ************************************************** 2025-08-29 17:34:09.105075 | orchestrator | Friday 29 August 2025 17:32:45 +0000 (0:00:38.067) 0:01:12.782 ********* 2025-08-29 17:34:09.105083 | orchestrator | included: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:34:09.105090 | orchestrator | 2025-08-29 17:34:09.105097 | orchestrator | TASK [ovn-db : include_tasks] ************************************************** 2025-08-29 17:34:09.105105 | orchestrator | Friday 29 August 2025 17:32:46 +0000 (0:00:00.691) 0:01:13.473 ********* 2025-08-29 17:34:09.105112 | orchestrator | included: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:34:09.105119 | orchestrator | 2025-08-29 17:34:09.105131 | orchestrator | TASK [ovn-db : Checking for any existing OVN DB container volumes] ************* 2025-08-29 17:34:09.105139 | orchestrator | Friday 29 August 2025 17:32:47 +0000 (0:00:00.670) 0:01:14.144 ********* 2025-08-29 17:34:09.105146 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.105154 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.105161 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.105168 | orchestrator | 2025-08-29 17:34:09.105175 | orchestrator | TASK [ovn-db : Divide hosts by their OVN NB volume availability] *************** 2025-08-29 17:34:09.105183 | orchestrator | Friday 29 August 2025 17:32:48 +0000 (0:00:00.911) 0:01:15.055 ********* 2025-08-29 17:34:09.105190 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.105197 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.105204 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.105212 | orchestrator | 2025-08-29 17:34:09.105219 | orchestrator | TASK [ovn-db : Divide hosts by their OVN SB volume availability] *************** 2025-08-29 17:34:09.105226 | orchestrator | Friday 29 August 2025 17:32:48 +0000 (0:00:00.274) 0:01:15.329 ********* 2025-08-29 17:34:09.105233 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.105241 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.105248 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.105255 | orchestrator | 2025-08-29 17:34:09.105269 | orchestrator | TASK [ovn-db : Establish whether the OVN NB cluster has already existed] ******* 2025-08-29 17:34:09.105277 | orchestrator | Friday 29 August 2025 17:32:48 +0000 (0:00:00.281) 0:01:15.611 ********* 2025-08-29 17:34:09.105284 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.105292 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.105299 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.105306 | orchestrator | 2025-08-29 17:34:09.105314 | orchestrator | TASK [ovn-db : Establish whether the OVN SB cluster has already existed] ******* 2025-08-29 17:34:09.105321 | orchestrator | Friday 29 August 2025 17:32:49 +0000 (0:00:00.298) 0:01:15.909 ********* 2025-08-29 17:34:09.105328 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.105335 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.105342 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.105350 | orchestrator | 2025-08-29 17:34:09.105357 | orchestrator | TASK [ovn-db : Check if running on all OVN NB DB hosts] ************************ 2025-08-29 17:34:09.105377 | orchestrator | Friday 29 August 2025 17:32:49 +0000 (0:00:00.447) 0:01:16.357 ********* 2025-08-29 17:34:09.105384 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105392 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105399 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105406 | orchestrator | 2025-08-29 17:34:09.105413 | orchestrator | TASK [ovn-db : Check OVN NB service port liveness] ***************************** 2025-08-29 17:34:09.105421 | orchestrator | Friday 29 August 2025 17:32:49 +0000 (0:00:00.294) 0:01:16.651 ********* 2025-08-29 17:34:09.105428 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105435 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105442 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105449 | orchestrator | 2025-08-29 17:34:09.105456 | orchestrator | TASK [ovn-db : Divide hosts by their OVN NB service port liveness] ************* 2025-08-29 17:34:09.105464 | orchestrator | Friday 29 August 2025 17:32:50 +0000 (0:00:00.263) 0:01:16.915 ********* 2025-08-29 17:34:09.105476 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105483 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105490 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105497 | orchestrator | 2025-08-29 17:34:09.105505 | orchestrator | TASK [ovn-db : Get OVN NB database information] ******************************** 2025-08-29 17:34:09.105512 | orchestrator | Friday 29 August 2025 17:32:50 +0000 (0:00:00.302) 0:01:17.218 ********* 2025-08-29 17:34:09.105519 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105526 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105533 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105540 | orchestrator | 2025-08-29 17:34:09.105548 | orchestrator | TASK [ovn-db : Divide hosts by their OVN NB leader/follower role] ************** 2025-08-29 17:34:09.105555 | orchestrator | Friday 29 August 2025 17:32:50 +0000 (0:00:00.537) 0:01:17.755 ********* 2025-08-29 17:34:09.105562 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105569 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105576 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105583 | orchestrator | 2025-08-29 17:34:09.105591 | orchestrator | TASK [ovn-db : Fail on existing OVN NB cluster with no leader] ***************** 2025-08-29 17:34:09.105598 | orchestrator | Friday 29 August 2025 17:32:51 +0000 (0:00:00.305) 0:01:18.061 ********* 2025-08-29 17:34:09.105605 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105613 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105620 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105627 | orchestrator | 2025-08-29 17:34:09.105634 | orchestrator | TASK [ovn-db : Check if running on all OVN SB DB hosts] ************************ 2025-08-29 17:34:09.105641 | orchestrator | Friday 29 August 2025 17:32:51 +0000 (0:00:00.292) 0:01:18.354 ********* 2025-08-29 17:34:09.105648 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105656 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105663 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105670 | orchestrator | 2025-08-29 17:34:09.105677 | orchestrator | TASK [ovn-db : Check OVN SB service port liveness] ***************************** 2025-08-29 17:34:09.105684 | orchestrator | Friday 29 August 2025 17:32:51 +0000 (0:00:00.312) 0:01:18.667 ********* 2025-08-29 17:34:09.105691 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105698 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105706 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105713 | orchestrator | 2025-08-29 17:34:09.105720 | orchestrator | TASK [ovn-db : Divide hosts by their OVN SB service port liveness] ************* 2025-08-29 17:34:09.105727 | orchestrator | Friday 29 August 2025 17:32:52 +0000 (0:00:00.493) 0:01:19.160 ********* 2025-08-29 17:34:09.105735 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105742 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105749 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105756 | orchestrator | 2025-08-29 17:34:09.105764 | orchestrator | TASK [ovn-db : Get OVN SB database information] ******************************** 2025-08-29 17:34:09.105771 | orchestrator | Friday 29 August 2025 17:32:52 +0000 (0:00:00.291) 0:01:19.451 ********* 2025-08-29 17:34:09.105778 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105785 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105792 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105799 | orchestrator | 2025-08-29 17:34:09.105811 | orchestrator | TASK [ovn-db : Divide hosts by their OVN SB leader/follower role] ************** 2025-08-29 17:34:09.105818 | orchestrator | Friday 29 August 2025 17:32:52 +0000 (0:00:00.303) 0:01:19.755 ********* 2025-08-29 17:34:09.105826 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105833 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105840 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105847 | orchestrator | 2025-08-29 17:34:09.105854 | orchestrator | TASK [ovn-db : Fail on existing OVN SB cluster with no leader] ***************** 2025-08-29 17:34:09.105915 | orchestrator | Friday 29 August 2025 17:32:53 +0000 (0:00:00.288) 0:01:20.044 ********* 2025-08-29 17:34:09.105934 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.105941 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.105949 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.105956 | orchestrator | 2025-08-29 17:34:09.105963 | orchestrator | TASK [ovn-db : include_tasks] ************************************************** 2025-08-29 17:34:09.105970 | orchestrator | Friday 29 August 2025 17:32:53 +0000 (0:00:00.459) 0:01:20.503 ********* 2025-08-29 17:34:09.105978 | orchestrator | included: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:34:09.105985 | orchestrator | 2025-08-29 17:34:09.105992 | orchestrator | TASK [ovn-db : Set bootstrap args fact for NB (new cluster)] ******************* 2025-08-29 17:34:09.105999 | orchestrator | Friday 29 August 2025 17:32:54 +0000 (0:00:00.604) 0:01:21.108 ********* 2025-08-29 17:34:09.106006 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.106065 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.106077 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.106084 | orchestrator | 2025-08-29 17:34:09.106092 | orchestrator | TASK [ovn-db : Set bootstrap args fact for SB (new cluster)] ******************* 2025-08-29 17:34:09.106099 | orchestrator | Friday 29 August 2025 17:32:54 +0000 (0:00:00.588) 0:01:21.697 ********* 2025-08-29 17:34:09.106106 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.106114 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.106121 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.106128 | orchestrator | 2025-08-29 17:34:09.106136 | orchestrator | TASK [ovn-db : Check NB cluster status] **************************************** 2025-08-29 17:34:09.106143 | orchestrator | Friday 29 August 2025 17:32:55 +0000 (0:00:00.801) 0:01:22.499 ********* 2025-08-29 17:34:09.106150 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.106157 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.106165 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.106172 | orchestrator | 2025-08-29 17:34:09.106179 | orchestrator | TASK [ovn-db : Check SB cluster status] **************************************** 2025-08-29 17:34:09.106186 | orchestrator | Friday 29 August 2025 17:32:56 +0000 (0:00:00.383) 0:01:22.882 ********* 2025-08-29 17:34:09.106193 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.106200 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.106208 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.106215 | orchestrator | 2025-08-29 17:34:09.106222 | orchestrator | TASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] *** 2025-08-29 17:34:09.106229 | orchestrator | Friday 29 August 2025 17:32:56 +0000 (0:00:00.300) 0:01:23.183 ********* 2025-08-29 17:34:09.106237 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.106244 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.106251 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.106258 | orchestrator | 2025-08-29 17:34:09.106265 | orchestrator | TASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] *** 2025-08-29 17:34:09.106273 | orchestrator | Friday 29 August 2025 17:32:56 +0000 (0:00:00.318) 0:01:23.501 ********* 2025-08-29 17:34:09.106280 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.106287 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.106294 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.106301 | orchestrator | 2025-08-29 17:34:09.106308 | orchestrator | TASK [ovn-db : Set bootstrap args fact for NB (new member)] ******************** 2025-08-29 17:34:09.106316 | orchestrator | Friday 29 August 2025 17:32:57 +0000 (0:00:00.493) 0:01:23.995 ********* 2025-08-29 17:34:09.106323 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.106330 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.106337 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.106344 | orchestrator | 2025-08-29 17:34:09.106351 | orchestrator | TASK [ovn-db : Set bootstrap args fact for SB (new member)] ******************** 2025-08-29 17:34:09.106372 | orchestrator | Friday 29 August 2025 17:32:57 +0000 (0:00:00.395) 0:01:24.391 ********* 2025-08-29 17:34:09.106380 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.106392 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.106399 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.106407 | orchestrator | 2025-08-29 17:34:09.106414 | orchestrator | TASK [ovn-db : Ensuring config directories exist] ****************************** 2025-08-29 17:34:09.106421 | orchestrator | Friday 29 August 2025 17:32:57 +0000 (0:00:00.322) 0:01:24.714 ********* 2025-08-29 17:34:09.106430 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106443 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106638 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106709 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106727 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106740 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106751 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106762 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106774 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106812 | orchestrator | 2025-08-29 17:34:09.106826 | orchestrator | TASK [ovn-db : Copying over config.json files for services] ******************** 2025-08-29 17:34:09.106838 | orchestrator | Friday 29 August 2025 17:32:59 +0000 (0:00:01.556) 0:01:26.270 ********* 2025-08-29 17:34:09.106850 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106862 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106887 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106913 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106926 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106937 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106949 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106960 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106971 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.106990 | orchestrator | 2025-08-29 17:34:09.107002 | orchestrator | TASK [ovn-db : Check ovn containers] ******************************************* 2025-08-29 17:34:09.107013 | orchestrator | Friday 29 August 2025 17:33:03 +0000 (0:00:03.936) 0:01:30.206 ********* 2025-08-29 17:34:09.107024 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107036 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107052 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107071 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107083 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107094 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107105 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107116 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107134 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.107146 | orchestrator | 2025-08-29 17:34:09.107157 | orchestrator | TASK [ovn-db : Flush handlers] ************************************************* 2025-08-29 17:34:09.107168 | orchestrator | Friday 29 August 2025 17:33:05 +0000 (0:00:02.129) 0:01:32.336 ********* 2025-08-29 17:34:09.107179 | orchestrator | 2025-08-29 17:34:09.107190 | orchestrator | TASK [ovn-db : Flush handlers] ************************************************* 2025-08-29 17:34:09.107201 | orchestrator | Friday 29 August 2025 17:33:05 +0000 (0:00:00.136) 0:01:32.473 ********* 2025-08-29 17:34:09.107211 | orchestrator | 2025-08-29 17:34:09.107222 | orchestrator | TASK [ovn-db : Flush handlers] ************************************************* 2025-08-29 17:34:09.107233 | orchestrator | Friday 29 August 2025 17:33:05 +0000 (0:00:00.133) 0:01:32.607 ********* 2025-08-29 17:34:09.107244 | orchestrator | 2025-08-29 17:34:09.107254 | orchestrator | RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] ************************* 2025-08-29 17:34:09.107265 | orchestrator | Friday 29 August 2025 17:33:05 +0000 (0:00:00.077) 0:01:32.684 ********* 2025-08-29 17:34:09.107276 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.107288 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.107298 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.107309 | orchestrator | 2025-08-29 17:34:09.107319 | orchestrator | RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] ************************* 2025-08-29 17:34:09.107330 | orchestrator | Friday 29 August 2025 17:33:13 +0000 (0:00:07.799) 0:01:40.483 ********* 2025-08-29 17:34:09.107341 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.107352 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.107387 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.107399 | orchestrator | 2025-08-29 17:34:09.107410 | orchestrator | RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************ 2025-08-29 17:34:09.107421 | orchestrator | Friday 29 August 2025 17:33:21 +0000 (0:00:07.857) 0:01:48.341 ********* 2025-08-29 17:34:09.107432 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.107442 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.107453 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.107463 | orchestrator | 2025-08-29 17:34:09.107474 | orchestrator | TASK [ovn-db : Wait for leader election] *************************************** 2025-08-29 17:34:09.107490 | orchestrator | Friday 29 August 2025 17:33:28 +0000 (0:00:07.382) 0:01:55.723 ********* 2025-08-29 17:34:09.107501 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.107512 | orchestrator | 2025-08-29 17:34:09.107523 | orchestrator | TASK [ovn-db : Get OVN_Northbound cluster leader] ****************************** 2025-08-29 17:34:09.107533 | orchestrator | Friday 29 August 2025 17:33:29 +0000 (0:00:00.144) 0:01:55.868 ********* 2025-08-29 17:34:09.107544 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.107555 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.107566 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.107577 | orchestrator | 2025-08-29 17:34:09.107594 | orchestrator | TASK [ovn-db : Configure OVN NB connection settings] *************************** 2025-08-29 17:34:09.107605 | orchestrator | Friday 29 August 2025 17:33:29 +0000 (0:00:00.915) 0:01:56.783 ********* 2025-08-29 17:34:09.107616 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.107627 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.107638 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.107649 | orchestrator | 2025-08-29 17:34:09.107660 | orchestrator | TASK [ovn-db : Get OVN_Southbound cluster leader] ****************************** 2025-08-29 17:34:09.107671 | orchestrator | Friday 29 August 2025 17:33:30 +0000 (0:00:00.870) 0:01:57.653 ********* 2025-08-29 17:34:09.107681 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.107692 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.107713 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.107724 | orchestrator | 2025-08-29 17:34:09.107735 | orchestrator | TASK [ovn-db : Configure OVN SB connection settings] *************************** 2025-08-29 17:34:09.107747 | orchestrator | Friday 29 August 2025 17:33:32 +0000 (0:00:01.208) 0:01:58.862 ********* 2025-08-29 17:34:09.107758 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.107769 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.107779 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.107790 | orchestrator | 2025-08-29 17:34:09.107801 | orchestrator | TASK [ovn-db : Wait for ovn-nb-db] ********************************************* 2025-08-29 17:34:09.107812 | orchestrator | Friday 29 August 2025 17:33:32 +0000 (0:00:00.605) 0:01:59.468 ********* 2025-08-29 17:34:09.107823 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.107834 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.107844 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.107855 | orchestrator | 2025-08-29 17:34:09.107866 | orchestrator | TASK [ovn-db : Wait for ovn-sb-db] ********************************************* 2025-08-29 17:34:09.107877 | orchestrator | Friday 29 August 2025 17:33:33 +0000 (0:00:00.885) 0:02:00.353 ********* 2025-08-29 17:34:09.107888 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.107899 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.107910 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.107921 | orchestrator | 2025-08-29 17:34:09.107931 | orchestrator | TASK [ovn-db : Unset bootstrap args fact] ************************************** 2025-08-29 17:34:09.107942 | orchestrator | Friday 29 August 2025 17:33:34 +0000 (0:00:01.154) 0:02:01.508 ********* 2025-08-29 17:34:09.107953 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.107964 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.107975 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.107986 | orchestrator | 2025-08-29 17:34:09.107997 | orchestrator | TASK [ovn-db : Ensuring config directories exist] ****************************** 2025-08-29 17:34:09.108008 | orchestrator | Friday 29 August 2025 17:33:35 +0000 (0:00:00.752) 0:02:02.260 ********* 2025-08-29 17:34:09.108019 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108031 | orchestrator | ok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108042 | orchestrator | ok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108054 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108066 | orchestrator | ok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108088 | orchestrator | ok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108106 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108118 | orchestrator | ok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108129 | orchestrator | ok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108140 | orchestrator | 2025-08-29 17:34:09.108151 | orchestrator | TASK [ovn-db : Copying over config.json files for services] ******************** 2025-08-29 17:34:09.108162 | orchestrator | Friday 29 August 2025 17:33:36 +0000 (0:00:01.428) 0:02:03.689 ********* 2025-08-29 17:34:09.108174 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108185 | orchestrator | ok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108196 | orchestrator | ok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108207 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108219 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108241 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108259 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108271 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108282 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108293 | orchestrator | 2025-08-29 17:34:09.108304 | orchestrator | TASK [ovn-db : Check ovn containers] ******************************************* 2025-08-29 17:34:09.108315 | orchestrator | Friday 29 August 2025 17:33:41 +0000 (0:00:04.462) 0:02:08.151 ********* 2025-08-29 17:34:09.108326 | orchestrator | ok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108337 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108349 | orchestrator | ok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108379 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108391 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108410 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108426 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108445 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108457 | orchestrator | ok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:34:09.108468 | orchestrator | 2025-08-29 17:34:09.108479 | orchestrator | TASK [ovn-db : Flush handlers] ************************************************* 2025-08-29 17:34:09.108490 | orchestrator | Friday 29 August 2025 17:33:44 +0000 (0:00:02.998) 0:02:11.150 ********* 2025-08-29 17:34:09.108501 | orchestrator | 2025-08-29 17:34:09.108512 | orchestrator | TASK [ovn-db : Flush handlers] ************************************************* 2025-08-29 17:34:09.108523 | orchestrator | Friday 29 August 2025 17:33:44 +0000 (0:00:00.067) 0:02:11.217 ********* 2025-08-29 17:34:09.108534 | orchestrator | 2025-08-29 17:34:09.108544 | orchestrator | TASK [ovn-db : Flush handlers] ************************************************* 2025-08-29 17:34:09.108555 | orchestrator | Friday 29 August 2025 17:33:44 +0000 (0:00:00.215) 0:02:11.433 ********* 2025-08-29 17:34:09.108566 | orchestrator | 2025-08-29 17:34:09.108577 | orchestrator | RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] ************************* 2025-08-29 17:34:09.108588 | orchestrator | Friday 29 August 2025 17:33:44 +0000 (0:00:00.067) 0:02:11.500 ********* 2025-08-29 17:34:09.108598 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.108609 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.108620 | orchestrator | 2025-08-29 17:34:09.108631 | orchestrator | RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] ************************* 2025-08-29 17:34:09.108641 | orchestrator | Friday 29 August 2025 17:33:50 +0000 (0:00:06.181) 0:02:17.681 ********* 2025-08-29 17:34:09.108652 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.108663 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.108674 | orchestrator | 2025-08-29 17:34:09.108684 | orchestrator | RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************ 2025-08-29 17:34:09.108696 | orchestrator | Friday 29 August 2025 17:33:57 +0000 (0:00:06.284) 0:02:23.966 ********* 2025-08-29 17:34:09.108706 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:34:09.108717 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:34:09.108728 | orchestrator | 2025-08-29 17:34:09.108739 | orchestrator | TASK [ovn-db : Wait for leader election] *************************************** 2025-08-29 17:34:09.108750 | orchestrator | Friday 29 August 2025 17:34:03 +0000 (0:00:06.213) 0:02:30.180 ********* 2025-08-29 17:34:09.108767 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:34:09.108778 | orchestrator | 2025-08-29 17:34:09.108789 | orchestrator | TASK [ovn-db : Get OVN_Northbound cluster leader] ****************************** 2025-08-29 17:34:09.108800 | orchestrator | Friday 29 August 2025 17:34:03 +0000 (0:00:00.144) 0:02:30.325 ********* 2025-08-29 17:34:09.108811 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.108821 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.108832 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.108843 | orchestrator | 2025-08-29 17:34:09.108854 | orchestrator | TASK [ovn-db : Configure OVN NB connection settings] *************************** 2025-08-29 17:34:09.108865 | orchestrator | Friday 29 August 2025 17:34:04 +0000 (0:00:00.724) 0:02:31.049 ********* 2025-08-29 17:34:09.108875 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.108886 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.108897 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.108908 | orchestrator | 2025-08-29 17:34:09.108918 | orchestrator | TASK [ovn-db : Get OVN_Southbound cluster leader] ****************************** 2025-08-29 17:34:09.108929 | orchestrator | Friday 29 August 2025 17:34:04 +0000 (0:00:00.593) 0:02:31.643 ********* 2025-08-29 17:34:09.108940 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.108951 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.108962 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.108973 | orchestrator | 2025-08-29 17:34:09.108984 | orchestrator | TASK [ovn-db : Configure OVN SB connection settings] *************************** 2025-08-29 17:34:09.108994 | orchestrator | Friday 29 August 2025 17:34:05 +0000 (0:00:00.780) 0:02:32.424 ********* 2025-08-29 17:34:09.109005 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:34:09.109016 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:34:09.109027 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:34:09.109038 | orchestrator | 2025-08-29 17:34:09.109048 | orchestrator | TASK [ovn-db : Wait for ovn-nb-db] ********************************************* 2025-08-29 17:34:09.109059 | orchestrator | Friday 29 August 2025 17:34:06 +0000 (0:00:00.575) 0:02:32.999 ********* 2025-08-29 17:34:09.109070 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.109081 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.109091 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.109102 | orchestrator | 2025-08-29 17:34:09.109114 | orchestrator | TASK [ovn-db : Wait for ovn-sb-db] ********************************************* 2025-08-29 17:34:09.109125 | orchestrator | Friday 29 August 2025 17:34:07 +0000 (0:00:00.828) 0:02:33.827 ********* 2025-08-29 17:34:09.109135 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:34:09.109146 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:34:09.109157 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:34:09.109167 | orchestrator | 2025-08-29 17:34:09.109188 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:34:09.109200 | orchestrator | testbed-node-0 : ok=44  changed=18  unreachable=0 failed=0 skipped=20  rescued=0 ignored=0 2025-08-29 17:34:09.109212 | orchestrator | testbed-node-1 : ok=43  changed=19  unreachable=0 failed=0 skipped=22  rescued=0 ignored=0 2025-08-29 17:34:09.109229 | orchestrator | testbed-node-2 : ok=43  changed=19  unreachable=0 failed=0 skipped=22  rescued=0 ignored=0 2025-08-29 17:34:09.109241 | orchestrator | testbed-node-3 : ok=12  changed=8  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:34:09.109252 | orchestrator | testbed-node-4 : ok=12  changed=8  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:34:09.109263 | orchestrator | testbed-node-5 : ok=12  changed=8  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:34:09.109274 | orchestrator | 2025-08-29 17:34:09.109285 | orchestrator | 2025-08-29 17:34:09.109302 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:34:09.109313 | orchestrator | Friday 29 August 2025 17:34:08 +0000 (0:00:01.218) 0:02:35.046 ********* 2025-08-29 17:34:09.109324 | orchestrator | =============================================================================== 2025-08-29 17:34:09.109335 | orchestrator | ovn-controller : Restart ovn-controller container ---------------------- 38.07s 2025-08-29 17:34:09.109346 | orchestrator | ovn-controller : Configure OVN in OVSDB -------------------------------- 19.55s 2025-08-29 17:34:09.109357 | orchestrator | ovn-db : Restart ovn-sb-db container ----------------------------------- 14.14s 2025-08-29 17:34:09.109385 | orchestrator | ovn-db : Restart ovn-nb-db container ----------------------------------- 13.98s 2025-08-29 17:34:09.109396 | orchestrator | ovn-db : Restart ovn-northd container ---------------------------------- 13.60s 2025-08-29 17:34:09.109407 | orchestrator | ovn-db : Copying over config.json files for services -------------------- 4.46s 2025-08-29 17:34:09.109417 | orchestrator | ovn-db : Copying over config.json files for services -------------------- 3.94s 2025-08-29 17:34:09.109428 | orchestrator | ovn-db : Check ovn containers ------------------------------------------- 3.00s 2025-08-29 17:34:09.109439 | orchestrator | ovn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.46s 2025-08-29 17:34:09.109450 | orchestrator | ovn-db : Check ovn containers ------------------------------------------- 2.13s 2025-08-29 17:34:09.109461 | orchestrator | ovn-controller : Reload systemd config ---------------------------------- 1.78s 2025-08-29 17:34:09.109471 | orchestrator | ovn-db : Ensuring config directories exist ------------------------------ 1.56s 2025-08-29 17:34:09.109482 | orchestrator | ovn-controller : Copying over systemd override -------------------------- 1.53s 2025-08-29 17:34:09.109493 | orchestrator | ovn-controller : Ensuring systemd override directory exists ------------- 1.45s 2025-08-29 17:34:09.109504 | orchestrator | ovn-db : Ensuring config directories exist ------------------------------ 1.43s 2025-08-29 17:34:09.109515 | orchestrator | ovn-controller : Copying over config.json files for services ------------ 1.42s 2025-08-29 17:34:09.109525 | orchestrator | Group hosts based on enabled services ----------------------------------- 1.37s 2025-08-29 17:34:09.109536 | orchestrator | ovn-controller : include_tasks ------------------------------------------ 1.34s 2025-08-29 17:34:09.109547 | orchestrator | ovn-controller : Check ovn-controller containers ------------------------ 1.34s 2025-08-29 17:34:09.109558 | orchestrator | ovn-db : Wait for ovn-sb-db --------------------------------------------- 1.22s 2025-08-29 17:34:09.109569 | orchestrator | 2025-08-29 17:34:09 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:09.109580 | orchestrator | 2025-08-29 17:34:09 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:09.109591 | orchestrator | 2025-08-29 17:34:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:12.159411 | orchestrator | 2025-08-29 17:34:12 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:12.159512 | orchestrator | 2025-08-29 17:34:12 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:12.159523 | orchestrator | 2025-08-29 17:34:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:15.209868 | orchestrator | 2025-08-29 17:34:15 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:15.210611 | orchestrator | 2025-08-29 17:34:15 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:15.210649 | orchestrator | 2025-08-29 17:34:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:18.252823 | orchestrator | 2025-08-29 17:34:18 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:18.254551 | orchestrator | 2025-08-29 17:34:18 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:18.254804 | orchestrator | 2025-08-29 17:34:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:21.295773 | orchestrator | 2025-08-29 17:34:21 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:21.296671 | orchestrator | 2025-08-29 17:34:21 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:21.296699 | orchestrator | 2025-08-29 17:34:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:24.337154 | orchestrator | 2025-08-29 17:34:24 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:24.337505 | orchestrator | 2025-08-29 17:34:24 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:24.337522 | orchestrator | 2025-08-29 17:34:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:27.376185 | orchestrator | 2025-08-29 17:34:27 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:27.378513 | orchestrator | 2025-08-29 17:34:27 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:27.378560 | orchestrator | 2025-08-29 17:34:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:30.410907 | orchestrator | 2025-08-29 17:34:30 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:30.411290 | orchestrator | 2025-08-29 17:34:30 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:30.411495 | orchestrator | 2025-08-29 17:34:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:33.450094 | orchestrator | 2025-08-29 17:34:33 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:33.452246 | orchestrator | 2025-08-29 17:34:33 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:33.452300 | orchestrator | 2025-08-29 17:34:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:36.483700 | orchestrator | 2025-08-29 17:34:36 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:36.483940 | orchestrator | 2025-08-29 17:34:36 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:36.483965 | orchestrator | 2025-08-29 17:34:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:39.513122 | orchestrator | 2025-08-29 17:34:39 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:39.514158 | orchestrator | 2025-08-29 17:34:39 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:39.514193 | orchestrator | 2025-08-29 17:34:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:42.577193 | orchestrator | 2025-08-29 17:34:42 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:42.580149 | orchestrator | 2025-08-29 17:34:42 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:42.580233 | orchestrator | 2025-08-29 17:34:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:45.628281 | orchestrator | 2025-08-29 17:34:45 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:45.629760 | orchestrator | 2025-08-29 17:34:45 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:45.629804 | orchestrator | 2025-08-29 17:34:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:48.672259 | orchestrator | 2025-08-29 17:34:48 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:48.673269 | orchestrator | 2025-08-29 17:34:48 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:48.673333 | orchestrator | 2025-08-29 17:34:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:51.711717 | orchestrator | 2025-08-29 17:34:51 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:51.712442 | orchestrator | 2025-08-29 17:34:51 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:51.712475 | orchestrator | 2025-08-29 17:34:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:54.756499 | orchestrator | 2025-08-29 17:34:54 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:54.758121 | orchestrator | 2025-08-29 17:34:54 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:54.758405 | orchestrator | 2025-08-29 17:34:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:34:57.801013 | orchestrator | 2025-08-29 17:34:57 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:34:57.801596 | orchestrator | 2025-08-29 17:34:57 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:34:57.801796 | orchestrator | 2025-08-29 17:34:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:00.843361 | orchestrator | 2025-08-29 17:35:00 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:00.844310 | orchestrator | 2025-08-29 17:35:00 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:00.844391 | orchestrator | 2025-08-29 17:35:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:03.897901 | orchestrator | 2025-08-29 17:35:03 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:03.899560 | orchestrator | 2025-08-29 17:35:03 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:03.899619 | orchestrator | 2025-08-29 17:35:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:06.938239 | orchestrator | 2025-08-29 17:35:06 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:06.938855 | orchestrator | 2025-08-29 17:35:06 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:06.938939 | orchestrator | 2025-08-29 17:35:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:09.988880 | orchestrator | 2025-08-29 17:35:09 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:09.990181 | orchestrator | 2025-08-29 17:35:09 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:09.990453 | orchestrator | 2025-08-29 17:35:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:13.055674 | orchestrator | 2025-08-29 17:35:13 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:13.057402 | orchestrator | 2025-08-29 17:35:13 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:13.058127 | orchestrator | 2025-08-29 17:35:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:16.110937 | orchestrator | 2025-08-29 17:35:16 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:16.112628 | orchestrator | 2025-08-29 17:35:16 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:16.113178 | orchestrator | 2025-08-29 17:35:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:19.167825 | orchestrator | 2025-08-29 17:35:19 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:19.169609 | orchestrator | 2025-08-29 17:35:19 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:19.169656 | orchestrator | 2025-08-29 17:35:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:22.222891 | orchestrator | 2025-08-29 17:35:22 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:22.226134 | orchestrator | 2025-08-29 17:35:22 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:22.226644 | orchestrator | 2025-08-29 17:35:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:25.268465 | orchestrator | 2025-08-29 17:35:25 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:25.270270 | orchestrator | 2025-08-29 17:35:25 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:25.270432 | orchestrator | 2025-08-29 17:35:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:28.321281 | orchestrator | 2025-08-29 17:35:28 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:28.322661 | orchestrator | 2025-08-29 17:35:28 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:28.322713 | orchestrator | 2025-08-29 17:35:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:31.369800 | orchestrator | 2025-08-29 17:35:31 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:31.371526 | orchestrator | 2025-08-29 17:35:31 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:31.372506 | orchestrator | 2025-08-29 17:35:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:34.418796 | orchestrator | 2025-08-29 17:35:34 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:34.422440 | orchestrator | 2025-08-29 17:35:34 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:34.422511 | orchestrator | 2025-08-29 17:35:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:37.472718 | orchestrator | 2025-08-29 17:35:37 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:37.473279 | orchestrator | 2025-08-29 17:35:37 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:37.473450 | orchestrator | 2025-08-29 17:35:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:40.515936 | orchestrator | 2025-08-29 17:35:40 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:40.516189 | orchestrator | 2025-08-29 17:35:40 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:40.516203 | orchestrator | 2025-08-29 17:35:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:43.554871 | orchestrator | 2025-08-29 17:35:43 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:43.557675 | orchestrator | 2025-08-29 17:35:43 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:43.557720 | orchestrator | 2025-08-29 17:35:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:46.601299 | orchestrator | 2025-08-29 17:35:46 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:46.601762 | orchestrator | 2025-08-29 17:35:46 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:46.601992 | orchestrator | 2025-08-29 17:35:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:49.645956 | orchestrator | 2025-08-29 17:35:49 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:49.647337 | orchestrator | 2025-08-29 17:35:49 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:49.647429 | orchestrator | 2025-08-29 17:35:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:52.685526 | orchestrator | 2025-08-29 17:35:52 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:52.687260 | orchestrator | 2025-08-29 17:35:52 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:52.687465 | orchestrator | 2025-08-29 17:35:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:55.741159 | orchestrator | 2025-08-29 17:35:55 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:55.742848 | orchestrator | 2025-08-29 17:35:55 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:55.742897 | orchestrator | 2025-08-29 17:35:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:35:58.783083 | orchestrator | 2025-08-29 17:35:58 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:35:58.785698 | orchestrator | 2025-08-29 17:35:58 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:35:58.786128 | orchestrator | 2025-08-29 17:35:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:01.837605 | orchestrator | 2025-08-29 17:36:01 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:01.839229 | orchestrator | 2025-08-29 17:36:01 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:01.839260 | orchestrator | 2025-08-29 17:36:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:04.887922 | orchestrator | 2025-08-29 17:36:04 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:04.889875 | orchestrator | 2025-08-29 17:36:04 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:04.889926 | orchestrator | 2025-08-29 17:36:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:07.929337 | orchestrator | 2025-08-29 17:36:07 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:07.931552 | orchestrator | 2025-08-29 17:36:07 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:07.931595 | orchestrator | 2025-08-29 17:36:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:10.975971 | orchestrator | 2025-08-29 17:36:10 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:10.978211 | orchestrator | 2025-08-29 17:36:10 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:10.978253 | orchestrator | 2025-08-29 17:36:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:14.023967 | orchestrator | 2025-08-29 17:36:14 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:14.026928 | orchestrator | 2025-08-29 17:36:14 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:14.026963 | orchestrator | 2025-08-29 17:36:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:17.074821 | orchestrator | 2025-08-29 17:36:17 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:17.076267 | orchestrator | 2025-08-29 17:36:17 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:17.076337 | orchestrator | 2025-08-29 17:36:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:20.117791 | orchestrator | 2025-08-29 17:36:20 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:20.119344 | orchestrator | 2025-08-29 17:36:20 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:20.119460 | orchestrator | 2025-08-29 17:36:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:23.158752 | orchestrator | 2025-08-29 17:36:23 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:23.158977 | orchestrator | 2025-08-29 17:36:23 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:23.159000 | orchestrator | 2025-08-29 17:36:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:26.189813 | orchestrator | 2025-08-29 17:36:26 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:26.190162 | orchestrator | 2025-08-29 17:36:26 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:26.190204 | orchestrator | 2025-08-29 17:36:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:29.241011 | orchestrator | 2025-08-29 17:36:29 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:29.241434 | orchestrator | 2025-08-29 17:36:29 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:29.241467 | orchestrator | 2025-08-29 17:36:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:32.292938 | orchestrator | 2025-08-29 17:36:32 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:32.293673 | orchestrator | 2025-08-29 17:36:32 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:32.293713 | orchestrator | 2025-08-29 17:36:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:35.340705 | orchestrator | 2025-08-29 17:36:35 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:35.342775 | orchestrator | 2025-08-29 17:36:35 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state STARTED 2025-08-29 17:36:35.342836 | orchestrator | 2025-08-29 17:36:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:38.394885 | orchestrator | 2025-08-29 17:36:38 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:38.394987 | orchestrator | 2025-08-29 17:36:38 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:38.395002 | orchestrator | 2025-08-29 17:36:38 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:38.397628 | orchestrator | 2025-08-29 17:36:38 | INFO  | Task 2cd22694-141e-46d9-aced-e7d6ef26d8b7 is in state SUCCESS 2025-08-29 17:36:38.399928 | orchestrator | 2025-08-29 17:36:38.399985 | orchestrator | 2025-08-29 17:36:38.399999 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:36:38.400011 | orchestrator | 2025-08-29 17:36:38.400023 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:36:38.400034 | orchestrator | Friday 29 August 2025 17:30:23 +0000 (0:00:00.266) 0:00:00.266 ********* 2025-08-29 17:36:38.400045 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.400057 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.400068 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.400079 | orchestrator | 2025-08-29 17:36:38.400089 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:36:38.400131 | orchestrator | Friday 29 August 2025 17:30:23 +0000 (0:00:00.295) 0:00:00.562 ********* 2025-08-29 17:36:38.400143 | orchestrator | ok: [testbed-node-0] => (item=enable_loadbalancer_True) 2025-08-29 17:36:38.400154 | orchestrator | ok: [testbed-node-1] => (item=enable_loadbalancer_True) 2025-08-29 17:36:38.400179 | orchestrator | ok: [testbed-node-2] => (item=enable_loadbalancer_True) 2025-08-29 17:36:38.400190 | orchestrator | 2025-08-29 17:36:38.400201 | orchestrator | PLAY [Apply role loadbalancer] ************************************************* 2025-08-29 17:36:38.400211 | orchestrator | 2025-08-29 17:36:38.400222 | orchestrator | TASK [loadbalancer : include_tasks] ******************************************** 2025-08-29 17:36:38.400233 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.395) 0:00:00.957 ********* 2025-08-29 17:36:38.400244 | orchestrator | included: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.400255 | orchestrator | 2025-08-29 17:36:38.400265 | orchestrator | TASK [loadbalancer : Check IPv6 support] *************************************** 2025-08-29 17:36:38.400276 | orchestrator | Friday 29 August 2025 17:30:24 +0000 (0:00:00.651) 0:00:01.608 ********* 2025-08-29 17:36:38.400287 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.400298 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.400309 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.400319 | orchestrator | 2025-08-29 17:36:38.400330 | orchestrator | TASK [Setting sysctl values] *************************************************** 2025-08-29 17:36:38.400437 | orchestrator | Friday 29 August 2025 17:30:25 +0000 (0:00:00.725) 0:00:02.334 ********* 2025-08-29 17:36:38.400449 | orchestrator | included: sysctl for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.400460 | orchestrator | 2025-08-29 17:36:38.400516 | orchestrator | TASK [sysctl : Check IPv6 support] ********************************************* 2025-08-29 17:36:38.400638 | orchestrator | Friday 29 August 2025 17:30:26 +0000 (0:00:00.729) 0:00:03.064 ********* 2025-08-29 17:36:38.400653 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.400725 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.400738 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.400751 | orchestrator | 2025-08-29 17:36:38.400763 | orchestrator | TASK [sysctl : Setting sysctl values] ****************************************** 2025-08-29 17:36:38.400775 | orchestrator | Friday 29 August 2025 17:30:28 +0000 (0:00:01.712) 0:00:04.776 ********* 2025-08-29 17:36:38.400787 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1}) 2025-08-29 17:36:38.400799 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1}) 2025-08-29 17:36:38.400811 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1}) 2025-08-29 17:36:38.400821 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1}) 2025-08-29 17:36:38.400832 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1}) 2025-08-29 17:36:38.400842 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1}) 2025-08-29 17:36:38.400853 | orchestrator | ok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'}) 2025-08-29 17:36:38.400864 | orchestrator | ok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'}) 2025-08-29 17:36:38.400875 | orchestrator | ok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'}) 2025-08-29 17:36:38.400885 | orchestrator | changed: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128}) 2025-08-29 17:36:38.400896 | orchestrator | changed: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128}) 2025-08-29 17:36:38.400906 | orchestrator | changed: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128}) 2025-08-29 17:36:38.400917 | orchestrator | 2025-08-29 17:36:38.400927 | orchestrator | TASK [module-load : Load modules] ********************************************** 2025-08-29 17:36:38.400938 | orchestrator | Friday 29 August 2025 17:30:31 +0000 (0:00:03.159) 0:00:07.936 ********* 2025-08-29 17:36:38.400973 | orchestrator | changed: [testbed-node-1] => (item=ip_vs) 2025-08-29 17:36:38.400984 | orchestrator | changed: [testbed-node-0] => (item=ip_vs) 2025-08-29 17:36:38.400995 | orchestrator | changed: [testbed-node-2] => (item=ip_vs) 2025-08-29 17:36:38.401005 | orchestrator | 2025-08-29 17:36:38.401016 | orchestrator | TASK [module-load : Persist modules via modules-load.d] ************************ 2025-08-29 17:36:38.401027 | orchestrator | Friday 29 August 2025 17:30:32 +0000 (0:00:00.780) 0:00:08.716 ********* 2025-08-29 17:36:38.401037 | orchestrator | changed: [testbed-node-1] => (item=ip_vs) 2025-08-29 17:36:38.401049 | orchestrator | changed: [testbed-node-0] => (item=ip_vs) 2025-08-29 17:36:38.401060 | orchestrator | changed: [testbed-node-2] => (item=ip_vs) 2025-08-29 17:36:38.401071 | orchestrator | 2025-08-29 17:36:38.401081 | orchestrator | TASK [module-load : Drop module persistence] *********************************** 2025-08-29 17:36:38.401092 | orchestrator | Friday 29 August 2025 17:30:33 +0000 (0:00:01.361) 0:00:10.077 ********* 2025-08-29 17:36:38.401103 | orchestrator | skipping: [testbed-node-0] => (item=ip_vs)  2025-08-29 17:36:38.401114 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.401140 | orchestrator | skipping: [testbed-node-1] => (item=ip_vs)  2025-08-29 17:36:38.401152 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.401163 | orchestrator | skipping: [testbed-node-2] => (item=ip_vs)  2025-08-29 17:36:38.401173 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.401184 | orchestrator | 2025-08-29 17:36:38.401194 | orchestrator | TASK [loadbalancer : Ensuring config directories exist] ************************ 2025-08-29 17:36:38.401205 | orchestrator | Friday 29 August 2025 17:30:34 +0000 (0:00:00.759) 0:00:10.837 ********* 2025-08-29 17:36:38.401225 | orchestrator | changed: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.401243 | orchestrator | changed: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.401255 | orchestrator | changed: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.401267 | orchestrator | changed: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.401286 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.401299 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.401317 | orchestrator | changed: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.401334 | orchestrator | changed: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.401346 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.401357 | orchestrator | 2025-08-29 17:36:38.401369 | orchestrator | TASK [loadbalancer : Ensuring haproxy service config subdir exists] ************ 2025-08-29 17:36:38.401438 | orchestrator | Friday 29 August 2025 17:30:37 +0000 (0:00:03.355) 0:00:14.192 ********* 2025-08-29 17:36:38.401450 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.401461 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.401552 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.401563 | orchestrator | 2025-08-29 17:36:38.401596 | orchestrator | TASK [loadbalancer : Ensuring proxysql service config subdirectories exist] **** 2025-08-29 17:36:38.401609 | orchestrator | Friday 29 August 2025 17:30:38 +0000 (0:00:01.029) 0:00:15.221 ********* 2025-08-29 17:36:38.401630 | orchestrator | changed: [testbed-node-1] => (item=users) 2025-08-29 17:36:38.401641 | orchestrator | changed: [testbed-node-0] => (item=users) 2025-08-29 17:36:38.401652 | orchestrator | changed: [testbed-node-2] => (item=users) 2025-08-29 17:36:38.401663 | orchestrator | changed: [testbed-node-1] => (item=rules) 2025-08-29 17:36:38.401787 | orchestrator | changed: [testbed-node-0] => (item=rules) 2025-08-29 17:36:38.401800 | orchestrator | changed: [testbed-node-2] => (item=rules) 2025-08-29 17:36:38.401849 | orchestrator | 2025-08-29 17:36:38.401860 | orchestrator | TASK [loadbalancer : Ensuring keepalived checks subdir exists] ***************** 2025-08-29 17:36:38.401871 | orchestrator | Friday 29 August 2025 17:30:40 +0000 (0:00:02.220) 0:00:17.442 ********* 2025-08-29 17:36:38.401882 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.401893 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.401903 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.401914 | orchestrator | 2025-08-29 17:36:38.401925 | orchestrator | TASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] ******************* 2025-08-29 17:36:38.401936 | orchestrator | Friday 29 August 2025 17:30:43 +0000 (0:00:02.226) 0:00:19.668 ********* 2025-08-29 17:36:38.401946 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.401957 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.401968 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.401978 | orchestrator | 2025-08-29 17:36:38.401991 | orchestrator | TASK [loadbalancer : Removing checks for services which are disabled] ********** 2025-08-29 17:36:38.402010 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:01.455) 0:00:21.124 ********* 2025-08-29 17:36:38.402283 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.402328 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.402347 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.402361 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.2.20250711', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}})  2025-08-29 17:36:38.402464 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.402484 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.402496 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.402508 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.402527 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.2.20250711', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}})  2025-08-29 17:36:38.402539 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.402651 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.402663 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.402687 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.402699 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.2.20250711', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}})  2025-08-29 17:36:38.402710 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.402721 | orchestrator | 2025-08-29 17:36:38.402732 | orchestrator | TASK [loadbalancer : Copying checks for services which are enabled] ************ 2025-08-29 17:36:38.402743 | orchestrator | Friday 29 August 2025 17:30:46 +0000 (0:00:01.698) 0:00:22.823 ********* 2025-08-29 17:36:38.402755 | orchestrator | changed: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.402787 | orchestrator | changed: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.402799 | orchestrator | changed: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.402816 | orchestrator | changed: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403050 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.403088 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.2.20250711', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}})  2025-08-29 17:36:38.403101 | orchestrator | changed: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403112 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.403157 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.2.20250711', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}})  2025-08-29 17:36:38.403177 | orchestrator | changed: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403204 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.403216 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/haproxy-ssh:9.2.20250711', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea', '__omit_place_holder__08e5db9e67356fedf8ec8208e62755b5f6c351ea'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}})  2025-08-29 17:36:38.403227 | orchestrator | 2025-08-29 17:36:38.403307 | orchestrator | TASK [loadbalancer : Copying over config.json files for services] ************** 2025-08-29 17:36:38.403675 | orchestrator | Friday 29 August 2025 17:30:49 +0000 (0:00:03.192) 0:00:26.015 ********* 2025-08-29 17:36:38.403703 | orchestrator | changed: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403725 | orchestrator | changed: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403762 | orchestrator | changed: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403794 | orchestrator | changed: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403830 | orchestrator | changed: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403851 | orchestrator | changed: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.403873 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.403893 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.403912 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.403931 | orchestrator | 2025-08-29 17:36:38.403949 | orchestrator | TASK [loadbalancer : Copying over haproxy.cfg] ********************************* 2025-08-29 17:36:38.404258 | orchestrator | Friday 29 August 2025 17:30:52 +0000 (0:00:03.332) 0:00:29.348 ********* 2025-08-29 17:36:38.404273 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2) 2025-08-29 17:36:38.404296 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2) 2025-08-29 17:36:38.404308 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2) 2025-08-29 17:36:38.404338 | orchestrator | 2025-08-29 17:36:38.404356 | orchestrator | TASK [loadbalancer : Copying over proxysql config] ***************************** 2025-08-29 17:36:38.404401 | orchestrator | Friday 29 August 2025 17:30:54 +0000 (0:00:02.031) 0:00:31.380 ********* 2025-08-29 17:36:38.404420 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2) 2025-08-29 17:36:38.404440 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2) 2025-08-29 17:36:38.404458 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2) 2025-08-29 17:36:38.404474 | orchestrator | 2025-08-29 17:36:38.404500 | orchestrator | TASK [loadbalancer : Copying over haproxy single external frontend config] ***** 2025-08-29 17:36:38.404628 | orchestrator | Friday 29 August 2025 17:30:59 +0000 (0:00:04.684) 0:00:36.064 ********* 2025-08-29 17:36:38.404647 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.404993 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.405025 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.405043 | orchestrator | 2025-08-29 17:36:38.405061 | orchestrator | TASK [loadbalancer : Copying over custom haproxy services configuration] ******* 2025-08-29 17:36:38.405078 | orchestrator | Friday 29 August 2025 17:31:00 +0000 (0:00:00.919) 0:00:36.983 ********* 2025-08-29 17:36:38.405097 | orchestrator | changed: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg) 2025-08-29 17:36:38.405118 | orchestrator | changed: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg) 2025-08-29 17:36:38.405136 | orchestrator | changed: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg) 2025-08-29 17:36:38.405324 | orchestrator | 2025-08-29 17:36:38.405346 | orchestrator | TASK [loadbalancer : Copying over keepalived.conf] ***************************** 2025-08-29 17:36:38.405480 | orchestrator | Friday 29 August 2025 17:31:02 +0000 (0:00:02.661) 0:00:39.645 ********* 2025-08-29 17:36:38.405496 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2) 2025-08-29 17:36:38.405507 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2) 2025-08-29 17:36:38.405517 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2) 2025-08-29 17:36:38.405528 | orchestrator | 2025-08-29 17:36:38.405539 | orchestrator | TASK [loadbalancer : Copying over haproxy.pem] ********************************* 2025-08-29 17:36:38.405549 | orchestrator | Friday 29 August 2025 17:31:04 +0000 (0:00:01.883) 0:00:41.528 ********* 2025-08-29 17:36:38.405560 | orchestrator | changed: [testbed-node-0] => (item=haproxy.pem) 2025-08-29 17:36:38.405571 | orchestrator | changed: [testbed-node-1] => (item=haproxy.pem) 2025-08-29 17:36:38.405581 | orchestrator | changed: [testbed-node-2] => (item=haproxy.pem) 2025-08-29 17:36:38.405592 | orchestrator | 2025-08-29 17:36:38.405603 | orchestrator | TASK [loadbalancer : Copying over haproxy-internal.pem] ************************ 2025-08-29 17:36:38.405614 | orchestrator | Friday 29 August 2025 17:31:07 +0000 (0:00:02.388) 0:00:43.917 ********* 2025-08-29 17:36:38.405624 | orchestrator | changed: [testbed-node-1] => (item=haproxy-internal.pem) 2025-08-29 17:36:38.405635 | orchestrator | changed: [testbed-node-2] => (item=haproxy-internal.pem) 2025-08-29 17:36:38.405661 | orchestrator | changed: [testbed-node-0] => (item=haproxy-internal.pem) 2025-08-29 17:36:38.405672 | orchestrator | 2025-08-29 17:36:38.405683 | orchestrator | TASK [loadbalancer : include_tasks] ******************************************** 2025-08-29 17:36:38.405694 | orchestrator | Friday 29 August 2025 17:31:09 +0000 (0:00:02.298) 0:00:46.216 ********* 2025-08-29 17:36:38.405704 | orchestrator | included: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.405715 | orchestrator | 2025-08-29 17:36:38.405725 | orchestrator | TASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] *** 2025-08-29 17:36:38.405858 | orchestrator | Friday 29 August 2025 17:31:10 +0000 (0:00:00.639) 0:00:46.855 ********* 2025-08-29 17:36:38.405875 | orchestrator | changed: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.405916 | orchestrator | changed: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.405946 | orchestrator | changed: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.405959 | orchestrator | changed: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.405971 | orchestrator | changed: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.405982 | orchestrator | changed: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.406009 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.406097 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.406131 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.406152 | orchestrator | 2025-08-29 17:36:38.406171 | orchestrator | TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] *** 2025-08-29 17:36:38.406190 | orchestrator | Friday 29 August 2025 17:31:14 +0000 (0:00:03.884) 0:00:50.740 ********* 2025-08-29 17:36:38.406216 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.406238 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.406291 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.406731 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.406751 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.406776 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.406795 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.406804 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.406818 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.406827 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.406835 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.406843 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.406852 | orchestrator | 2025-08-29 17:36:38.406860 | orchestrator | TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] *** 2025-08-29 17:36:38.406874 | orchestrator | Friday 29 August 2025 17:31:15 +0000 (0:00:01.004) 0:00:51.745 ********* 2025-08-29 17:36:38.406883 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.406891 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.406905 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.406914 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.406926 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.406935 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.406943 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.406951 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.406959 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.406972 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407097 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407108 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.407116 | orchestrator | 2025-08-29 17:36:38.407124 | orchestrator | TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ******** 2025-08-29 17:36:38.407132 | orchestrator | Friday 29 August 2025 17:31:16 +0000 (0:00:01.033) 0:00:52.778 ********* 2025-08-29 17:36:38.407147 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407160 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407168 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407186 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.407195 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407203 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407212 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407220 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.407233 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407246 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407255 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407263 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.407270 | orchestrator | 2025-08-29 17:36:38.407283 | orchestrator | TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] *** 2025-08-29 17:36:38.407291 | orchestrator | Friday 29 August 2025 17:31:16 +0000 (0:00:00.662) 0:00:53.441 ********* 2025-08-29 17:36:38.407299 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407308 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407316 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407324 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.407332 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407346 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407359 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407414 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.407425 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407433 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407441 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407450 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.407458 | orchestrator | 2025-08-29 17:36:38.407466 | orchestrator | TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] ***** 2025-08-29 17:36:38.407474 | orchestrator | Friday 29 August 2025 17:31:17 +0000 (0:00:00.922) 0:00:54.363 ********* 2025-08-29 17:36:38.407482 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407498 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407517 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407537 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.407551 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407566 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407580 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407591 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.407600 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407613 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407622 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407635 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.407643 | orchestrator | 2025-08-29 17:36:38.407655 | orchestrator | TASK [service-cert-copy : proxysql | Copying over extra CA certificates] ******* 2025-08-29 17:36:38.407664 | orchestrator | Friday 29 August 2025 17:31:18 +0000 (0:00:01.064) 0:00:55.427 ********* 2025-08-29 17:36:38.407680 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407701 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407714 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407727 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.407739 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407751 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407772 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407794 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.407813 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407829 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407841 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407853 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.407868 | orchestrator | 2025-08-29 17:36:38.407882 | orchestrator | TASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] *** 2025-08-29 17:36:38.407895 | orchestrator | Friday 29 August 2025 17:31:19 +0000 (0:00:00.699) 0:00:56.126 ********* 2025-08-29 17:36:38.407907 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407920 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.407940 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.407964 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.407983 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.407997 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.408011 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.408026 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.408039 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.408052 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.408061 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.408075 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.408083 | orchestrator | 2025-08-29 17:36:38.408091 | orchestrator | TASK [service-cert-copy : proxysql | Copying over backend internal TLS key] **** 2025-08-29 17:36:38.408105 | orchestrator | Friday 29 August 2025 17:31:20 +0000 (0:00:01.130) 0:00:57.256 ********* 2025-08-29 17:36:38.408113 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.408126 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.408134 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.408143 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.408151 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.408159 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.408167 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.408184 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.408196 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})  2025-08-29 17:36:38.408206 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})  2025-08-29 17:36:38.408214 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})  2025-08-29 17:36:38.408222 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.408230 | orchestrator | 2025-08-29 17:36:38.408238 | orchestrator | TASK [loadbalancer : Copying over haproxy start script] ************************ 2025-08-29 17:36:38.408247 | orchestrator | Friday 29 August 2025 17:31:21 +0000 (0:00:01.274) 0:00:58.531 ********* 2025-08-29 17:36:38.408275 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2) 2025-08-29 17:36:38.408284 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2) 2025-08-29 17:36:38.408292 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2) 2025-08-29 17:36:38.408300 | orchestrator | 2025-08-29 17:36:38.408308 | orchestrator | TASK [loadbalancer : Copying over proxysql start script] *********************** 2025-08-29 17:36:38.408316 | orchestrator | Friday 29 August 2025 17:31:23 +0000 (0:00:02.016) 0:01:00.548 ********* 2025-08-29 17:36:38.408324 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2) 2025-08-29 17:36:38.408332 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2) 2025-08-29 17:36:38.408340 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2) 2025-08-29 17:36:38.408348 | orchestrator | 2025-08-29 17:36:38.408356 | orchestrator | TASK [loadbalancer : Copying files for haproxy-ssh] **************************** 2025-08-29 17:36:38.408364 | orchestrator | Friday 29 August 2025 17:31:25 +0000 (0:00:01.386) 0:01:01.935 ********* 2025-08-29 17:36:38.408397 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'})  2025-08-29 17:36:38.408407 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'})  2025-08-29 17:36:38.408421 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'})  2025-08-29 17:36:38.408429 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'})  2025-08-29 17:36:38.408437 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.408444 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'})  2025-08-29 17:36:38.408452 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.408460 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'})  2025-08-29 17:36:38.408468 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.408476 | orchestrator | 2025-08-29 17:36:38.408483 | orchestrator | TASK [loadbalancer : Check loadbalancer containers] **************************** 2025-08-29 17:36:38.408491 | orchestrator | Friday 29 August 2025 17:31:26 +0000 (0:00:00.816) 0:01:02.751 ********* 2025-08-29 17:36:38.408505 | orchestrator | changed: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.408519 | orchestrator | changed: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.408528 | orchestrator | changed: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/haproxy:2.6.12.20250711', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) 2025-08-29 17:36:38.408539 | orchestrator | changed: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.408548 | orchestrator | changed: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.408562 | orchestrator | changed: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/proxysql:2.7.3.20250711', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) 2025-08-29 17:36:38.408571 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.408586 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.408600 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keepalived:2.2.7.20250711', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) 2025-08-29 17:36:38.408610 | orchestrator | 2025-08-29 17:36:38.408619 | orchestrator | TASK [include_role : aodh] ***************************************************** 2025-08-29 17:36:38.408628 | orchestrator | Friday 29 August 2025 17:31:28 +0000 (0:00:02.486) 0:01:05.238 ********* 2025-08-29 17:36:38.408637 | orchestrator | included: aodh for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.408645 | orchestrator | 2025-08-29 17:36:38.408654 | orchestrator | TASK [haproxy-config : Copying over aodh haproxy config] *********************** 2025-08-29 17:36:38.408662 | orchestrator | Friday 29 August 2025 17:31:29 +0000 (0:00:00.522) 0:01:05.761 ********* 2025-08-29 17:36:38.408678 | orchestrator | changed: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) 2025-08-29 17:36:38.408702 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.408716 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408737 | orchestrator | changed: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) 2025-08-29 17:36:38.408758 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408772 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.408786 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408809 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408825 | orchestrator | changed: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) 2025-08-29 17:36:38.408839 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.408858 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408870 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408879 | orchestrator | 2025-08-29 17:36:38.408887 | orchestrator | TASK [haproxy-config : Add configuration for aodh when using single external frontend] *** 2025-08-29 17:36:38.408895 | orchestrator | Friday 29 August 2025 17:31:32 +0000 (0:00:03.526) 0:01:09.287 ********* 2025-08-29 17:36:38.408903 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})  2025-08-29 17:36:38.408917 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.408925 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408938 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})  2025-08-29 17:36:38.408946 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408959 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.408967 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.408975 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408988 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.408996 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.409004 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-api:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})  2025-08-29 17:36:38.409012 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-evaluator:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.409028 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-listener:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409047 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/aodh-notifier:19.0.0.20250711', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409062 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.409075 | orchestrator | 2025-08-29 17:36:38.409090 | orchestrator | TASK [haproxy-config : Configuring firewall for aodh] ************************** 2025-08-29 17:36:38.409114 | orchestrator | Friday 29 August 2025 17:31:33 +0000 (0:00:00.887) 0:01:10.174 ********* 2025-08-29 17:36:38.409127 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}})  2025-08-29 17:36:38.409143 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}})  2025-08-29 17:36:38.409156 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.409169 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}})  2025-08-29 17:36:38.409181 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}})  2025-08-29 17:36:38.409193 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.409207 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}})  2025-08-29 17:36:38.409219 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}})  2025-08-29 17:36:38.409233 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.409245 | orchestrator | 2025-08-29 17:36:38.409258 | orchestrator | TASK [proxysql-config : Copying over aodh ProxySQL users config] *************** 2025-08-29 17:36:38.409270 | orchestrator | Friday 29 August 2025 17:31:34 +0000 (0:00:00.928) 0:01:11.103 ********* 2025-08-29 17:36:38.409283 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.409295 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.409308 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.409320 | orchestrator | 2025-08-29 17:36:38.409332 | orchestrator | TASK [proxysql-config : Copying over aodh ProxySQL rules config] *************** 2025-08-29 17:36:38.409345 | orchestrator | Friday 29 August 2025 17:31:36 +0000 (0:00:01.918) 0:01:13.022 ********* 2025-08-29 17:36:38.409357 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.409392 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.409408 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.409421 | orchestrator | 2025-08-29 17:36:38.409433 | orchestrator | TASK [include_role : barbican] ************************************************* 2025-08-29 17:36:38.409445 | orchestrator | Friday 29 August 2025 17:31:38 +0000 (0:00:01.858) 0:01:14.880 ********* 2025-08-29 17:36:38.409457 | orchestrator | included: barbican for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.409468 | orchestrator | 2025-08-29 17:36:38.409481 | orchestrator | TASK [haproxy-config : Copying over barbican haproxy config] ******************* 2025-08-29 17:36:38.409493 | orchestrator | Friday 29 August 2025 17:31:38 +0000 (0:00:00.559) 0:01:15.440 ********* 2025-08-29 17:36:38.409530 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.409573 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409588 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409602 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.409616 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409631 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409653 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.409680 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409764 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409780 | orchestrator | 2025-08-29 17:36:38.409794 | orchestrator | TASK [haproxy-config : Add configuration for barbican when using single external frontend] *** 2025-08-29 17:36:38.409808 | orchestrator | Friday 29 August 2025 17:31:42 +0000 (0:00:03.495) 0:01:18.935 ********* 2025-08-29 17:36:38.409822 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.409836 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409856 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409878 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.409898 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.409912 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409926 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.409939 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.409952 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.409972 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.410005 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.410078 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.410096 | orchestrator | 2025-08-29 17:36:38.410110 | orchestrator | TASK [haproxy-config : Configuring firewall for barbican] ********************** 2025-08-29 17:36:38.410123 | orchestrator | Friday 29 August 2025 17:31:43 +0000 (0:00:00.770) 0:01:19.706 ********* 2025-08-29 17:36:38.410137 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}})  2025-08-29 17:36:38.410151 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}})  2025-08-29 17:36:38.410165 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.410179 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}})  2025-08-29 17:36:38.410193 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}})  2025-08-29 17:36:38.410207 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.410221 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}})  2025-08-29 17:36:38.410234 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}})  2025-08-29 17:36:38.410248 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.410261 | orchestrator | 2025-08-29 17:36:38.410275 | orchestrator | TASK [proxysql-config : Copying over barbican ProxySQL users config] *********** 2025-08-29 17:36:38.410288 | orchestrator | Friday 29 August 2025 17:31:43 +0000 (0:00:00.784) 0:01:20.490 ********* 2025-08-29 17:36:38.410302 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.410316 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.410330 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.410343 | orchestrator | 2025-08-29 17:36:38.410356 | orchestrator | TASK [proxysql-config : Copying over barbican ProxySQL rules config] *********** 2025-08-29 17:36:38.410368 | orchestrator | Friday 29 August 2025 17:31:45 +0000 (0:00:01.219) 0:01:21.710 ********* 2025-08-29 17:36:38.410447 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.410502 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.410516 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.410531 | orchestrator | 2025-08-29 17:36:38.410544 | orchestrator | TASK [include_role : blazar] *************************************************** 2025-08-29 17:36:38.410556 | orchestrator | Friday 29 August 2025 17:31:47 +0000 (0:00:02.633) 0:01:24.344 ********* 2025-08-29 17:36:38.410580 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.410594 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.410608 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.410621 | orchestrator | 2025-08-29 17:36:38.410634 | orchestrator | TASK [include_role : ceph-rgw] ************************************************* 2025-08-29 17:36:38.410647 | orchestrator | Friday 29 August 2025 17:31:48 +0000 (0:00:00.389) 0:01:24.734 ********* 2025-08-29 17:36:38.410660 | orchestrator | included: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.410673 | orchestrator | 2025-08-29 17:36:38.410686 | orchestrator | TASK [haproxy-config : Copying over ceph-rgw haproxy config] ******************* 2025-08-29 17:36:38.410699 | orchestrator | Friday 29 August 2025 17:31:48 +0000 (0:00:00.588) 0:01:25.322 ********* 2025-08-29 17:36:38.410780 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) 2025-08-29 17:36:38.410807 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) 2025-08-29 17:36:38.410822 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}}) 2025-08-29 17:36:38.410835 | orchestrator | 2025-08-29 17:36:38.410849 | orchestrator | TASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] *** 2025-08-29 17:36:38.410861 | orchestrator | Friday 29 August 2025 17:31:50 +0000 (0:00:02.222) 0:01:27.545 ********* 2025-08-29 17:36:38.410873 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})  2025-08-29 17:36:38.410897 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.410911 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})  2025-08-29 17:36:38.410924 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.410947 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}}}})  2025-08-29 17:36:38.410962 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.410977 | orchestrator | 2025-08-29 17:36:38.410997 | orchestrator | TASK [haproxy-config : Configuring firewall for ceph-rgw] ********************** 2025-08-29 17:36:38.411011 | orchestrator | Friday 29 August 2025 17:31:52 +0000 (0:00:01.419) 0:01:28.965 ********* 2025-08-29 17:36:38.411025 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}})  2025-08-29 17:36:38.411040 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}})  2025-08-29 17:36:38.411054 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.411066 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}})  2025-08-29 17:36:38.411088 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}})  2025-08-29 17:36:38.411103 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.411117 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}})  2025-08-29 17:36:38.411130 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:8081 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:8081 check inter 2000 rise 2 fall 5']}})  2025-08-29 17:36:38.411139 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.411146 | orchestrator | 2025-08-29 17:36:38.411154 | orchestrator | TASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] *********** 2025-08-29 17:36:38.411162 | orchestrator | Friday 29 August 2025 17:31:53 +0000 (0:00:01.457) 0:01:30.423 ********* 2025-08-29 17:36:38.411170 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.411178 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.411185 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.411193 | orchestrator | 2025-08-29 17:36:38.411201 | orchestrator | TASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] *********** 2025-08-29 17:36:38.411209 | orchestrator | Friday 29 August 2025 17:31:54 +0000 (0:00:00.357) 0:01:30.781 ********* 2025-08-29 17:36:38.411216 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.411224 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.411232 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.411240 | orchestrator | 2025-08-29 17:36:38.411248 | orchestrator | TASK [include_role : cinder] *************************************************** 2025-08-29 17:36:38.411262 | orchestrator | Friday 29 August 2025 17:31:55 +0000 (0:00:01.172) 0:01:31.953 ********* 2025-08-29 17:36:38.411270 | orchestrator | included: cinder for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.411278 | orchestrator | 2025-08-29 17:36:38.411286 | orchestrator | TASK [haproxy-config : Copying over cinder haproxy config] ********************* 2025-08-29 17:36:38.411294 | orchestrator | Friday 29 August 2025 17:31:56 +0000 (0:00:00.753) 0:01:32.707 ********* 2025-08-29 17:36:38.411307 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.411317 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411333 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411343 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411359 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.411395 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411409 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411423 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411431 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.411439 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411452 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411464 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411477 | orchestrator | 2025-08-29 17:36:38.411485 | orchestrator | TASK [haproxy-config : Add configuration for cinder when using single external frontend] *** 2025-08-29 17:36:38.411492 | orchestrator | Friday 29 August 2025 17:31:59 +0000 (0:00:03.226) 0:01:35.934 ********* 2025-08-29 17:36:38.411500 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.411508 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411516 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411528 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411537 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.411549 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.411561 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411569 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411577 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411585 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.411598 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.411610 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411623 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411632 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.411639 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.411647 | orchestrator | 2025-08-29 17:36:38.411654 | orchestrator | TASK [haproxy-config : Configuring firewall for cinder] ************************ 2025-08-29 17:36:38.411662 | orchestrator | Friday 29 August 2025 17:31:59 +0000 (0:00:00.590) 0:01:36.524 ********* 2025-08-29 17:36:38.411670 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}})  2025-08-29 17:36:38.411715 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}})  2025-08-29 17:36:38.411724 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.411730 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}})  2025-08-29 17:36:38.411737 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}})  2025-08-29 17:36:38.411744 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.411756 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}})  2025-08-29 17:36:38.411763 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}})  2025-08-29 17:36:38.411775 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.411782 | orchestrator | 2025-08-29 17:36:38.411788 | orchestrator | TASK [proxysql-config : Copying over cinder ProxySQL users config] ************* 2025-08-29 17:36:38.411795 | orchestrator | Friday 29 August 2025 17:32:01 +0000 (0:00:01.307) 0:01:37.831 ********* 2025-08-29 17:36:38.411802 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.411808 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.411815 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.411821 | orchestrator | 2025-08-29 17:36:38.411828 | orchestrator | TASK [proxysql-config : Copying over cinder ProxySQL rules config] ************* 2025-08-29 17:36:38.411839 | orchestrator | Friday 29 August 2025 17:32:02 +0000 (0:00:01.448) 0:01:39.280 ********* 2025-08-29 17:36:38.411845 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.411852 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.411859 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.411865 | orchestrator | 2025-08-29 17:36:38.411872 | orchestrator | TASK [include_role : cloudkitty] *********************************************** 2025-08-29 17:36:38.411879 | orchestrator | Friday 29 August 2025 17:32:04 +0000 (0:00:02.095) 0:01:41.376 ********* 2025-08-29 17:36:38.411885 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.411892 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.411899 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.411905 | orchestrator | 2025-08-29 17:36:38.411912 | orchestrator | TASK [include_role : cyborg] *************************************************** 2025-08-29 17:36:38.411918 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.310) 0:01:41.686 ********* 2025-08-29 17:36:38.411925 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.411932 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.411938 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.411945 | orchestrator | 2025-08-29 17:36:38.411952 | orchestrator | TASK [include_role : designate] ************************************************ 2025-08-29 17:36:38.411958 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.310) 0:01:41.997 ********* 2025-08-29 17:36:38.411965 | orchestrator | included: designate for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.411971 | orchestrator | 2025-08-29 17:36:38.411978 | orchestrator | TASK [haproxy-config : Copying over designate haproxy config] ****************** 2025-08-29 17:36:38.411984 | orchestrator | Friday 29 August 2025 17:32:06 +0000 (0:00:01.084) 0:01:43.081 ********* 2025-08-29 17:36:38.411991 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:36:38.411999 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:36:38.412011 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412024 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412035 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412043 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412050 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412057 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:36:38.412068 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:36:38.412080 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412088 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412095 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412102 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412110 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412134 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:36:38.412150 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:36:38.412163 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412170 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412177 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412184 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412191 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412203 | orchestrator | 2025-08-29 17:36:38.412210 | orchestrator | TASK [haproxy-config : Add configuration for designate when using single external frontend] *** 2025-08-29 17:36:38.412217 | orchestrator | Friday 29 August 2025 17:32:11 +0000 (0:00:05.307) 0:01:48.389 ********* 2025-08-29 17:36:38.412228 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:36:38.412238 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:36:38.412246 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412253 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412260 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412331 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412345 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412356 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.412399 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:36:38.412414 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:36:38.412425 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:36:38.412438 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:36:38.412454 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412466 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412477 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412485 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412492 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412499 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412510 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412517 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412529 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412536 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.412547 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/designate-sink:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.412554 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.412560 | orchestrator | 2025-08-29 17:36:38.412567 | orchestrator | TASK [haproxy-config : Configuring firewall for designate] ********************* 2025-08-29 17:36:38.412574 | orchestrator | Friday 29 August 2025 17:32:13 +0000 (0:00:01.302) 0:01:49.691 ********* 2025-08-29 17:36:38.412581 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}})  2025-08-29 17:36:38.412588 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}})  2025-08-29 17:36:38.412595 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.412602 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}})  2025-08-29 17:36:38.412608 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}})  2025-08-29 17:36:38.412620 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.412627 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}})  2025-08-29 17:36:38.412634 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}})  2025-08-29 17:36:38.412641 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.412647 | orchestrator | 2025-08-29 17:36:38.412654 | orchestrator | TASK [proxysql-config : Copying over designate ProxySQL users config] ********** 2025-08-29 17:36:38.412660 | orchestrator | Friday 29 August 2025 17:32:14 +0000 (0:00:01.273) 0:01:50.965 ********* 2025-08-29 17:36:38.412667 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.412674 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.412680 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.412687 | orchestrator | 2025-08-29 17:36:38.412694 | orchestrator | TASK [proxysql-config : Copying over designate ProxySQL rules config] ********** 2025-08-29 17:36:38.412700 | orchestrator | Friday 29 August 2025 17:32:15 +0000 (0:00:01.396) 0:01:52.362 ********* 2025-08-29 17:36:38.412707 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.412713 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.412720 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.412727 | orchestrator | 2025-08-29 17:36:38.412733 | orchestrator | TASK [include_role : etcd] ***************************************************** 2025-08-29 17:36:38.412740 | orchestrator | Friday 29 August 2025 17:32:17 +0000 (0:00:02.005) 0:01:54.367 ********* 2025-08-29 17:36:38.412746 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.412753 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.412760 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.412766 | orchestrator | 2025-08-29 17:36:38.412773 | orchestrator | TASK [include_role : glance] *************************************************** 2025-08-29 17:36:38.412779 | orchestrator | Friday 29 August 2025 17:32:18 +0000 (0:00:00.484) 0:01:54.852 ********* 2025-08-29 17:36:38.412786 | orchestrator | included: glance for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.412792 | orchestrator | 2025-08-29 17:36:38.412799 | orchestrator | TASK [haproxy-config : Copying over glance haproxy config] ********************* 2025-08-29 17:36:38.412805 | orchestrator | Friday 29 August 2025 17:32:18 +0000 (0:00:00.787) 0:01:55.640 ********* 2025-08-29 17:36:38.412824 | orchestrator | changed: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:36:38.412838 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20250711', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.413001 | orchestrator | changed: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:36:38.413021 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20250711', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.413034 | orchestrator | changed: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:36:38.413046 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20250711', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.413062 | orchestrator | 2025-08-29 17:36:38.413070 | orchestrator | TASK [haproxy-config : Add configuration for glance when using single external frontend] *** 2025-08-29 17:36:38.413076 | orchestrator | Friday 29 August 2025 17:32:22 +0000 (0:00:03.929) 0:01:59.569 ********* 2025-08-29 17:36:38.413087 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:36:38.413099 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20250711', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.413111 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.413119 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:36:38.413135 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20250711', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.413147 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.413155 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:36:38.413171 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/glance-tls-proxy:29.0.1.20250711', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.413184 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.413191 | orchestrator | 2025-08-29 17:36:38.413198 | orchestrator | TASK [haproxy-config : Configuring firewall for glance] ************************ 2025-08-29 17:36:38.413205 | orchestrator | Friday 29 August 2025 17:32:26 +0000 (0:00:03.364) 0:02:02.934 ********* 2025-08-29 17:36:38.413211 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}})  2025-08-29 17:36:38.413219 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}})  2025-08-29 17:36:38.413226 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.413233 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}})  2025-08-29 17:36:38.413241 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}})  2025-08-29 17:36:38.413248 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}})  2025-08-29 17:36:38.413255 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.413266 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}})  2025-08-29 17:36:38.413277 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.413284 | orchestrator | 2025-08-29 17:36:38.413291 | orchestrator | TASK [proxysql-config : Copying over glance ProxySQL users config] ************* 2025-08-29 17:36:38.413298 | orchestrator | Friday 29 August 2025 17:32:29 +0000 (0:00:02.930) 0:02:05.864 ********* 2025-08-29 17:36:38.413305 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.413315 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.413321 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.413328 | orchestrator | 2025-08-29 17:36:38.413335 | orchestrator | TASK [proxysql-config : Copying over glance ProxySQL rules config] ************* 2025-08-29 17:36:38.413341 | orchestrator | Friday 29 August 2025 17:32:30 +0000 (0:00:01.281) 0:02:07.145 ********* 2025-08-29 17:36:38.413348 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.413355 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.413361 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.413368 | orchestrator | 2025-08-29 17:36:38.413421 | orchestrator | TASK [include_role : gnocchi] ************************************************** 2025-08-29 17:36:38.413428 | orchestrator | Friday 29 August 2025 17:32:32 +0000 (0:00:02.081) 0:02:09.226 ********* 2025-08-29 17:36:38.413435 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.413442 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.413448 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.413455 | orchestrator | 2025-08-29 17:36:38.413462 | orchestrator | TASK [include_role : grafana] ************************************************** 2025-08-29 17:36:38.413469 | orchestrator | Friday 29 August 2025 17:32:33 +0000 (0:00:00.570) 0:02:09.797 ********* 2025-08-29 17:36:38.413475 | orchestrator | included: grafana for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.413482 | orchestrator | 2025-08-29 17:36:38.413489 | orchestrator | TASK [haproxy-config : Copying over grafana haproxy config] ******************** 2025-08-29 17:36:38.413495 | orchestrator | Friday 29 August 2025 17:32:34 +0000 (0:00:00.859) 0:02:10.657 ********* 2025-08-29 17:36:38.413503 | orchestrator | changed: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:36:38.413510 | orchestrator | changed: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:36:38.413518 | orchestrator | changed: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:36:38.413530 | orchestrator | 2025-08-29 17:36:38.413537 | orchestrator | TASK [haproxy-config : Add configuration for grafana when using single external frontend] *** 2025-08-29 17:36:38.413544 | orchestrator | Friday 29 August 2025 17:32:37 +0000 (0:00:03.054) 0:02:13.711 ********* 2025-08-29 17:36:38.413555 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:36:38.413562 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.413573 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:36:38.413581 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.413588 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:36:38.413595 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.413602 | orchestrator | 2025-08-29 17:36:38.413608 | orchestrator | TASK [haproxy-config : Configuring firewall for grafana] *********************** 2025-08-29 17:36:38.413615 | orchestrator | Friday 29 August 2025 17:32:37 +0000 (0:00:00.901) 0:02:14.613 ********* 2025-08-29 17:36:38.413622 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}})  2025-08-29 17:36:38.413629 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}})  2025-08-29 17:36:38.413636 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.413643 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}})  2025-08-29 17:36:38.413650 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}})  2025-08-29 17:36:38.413661 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.413668 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}})  2025-08-29 17:36:38.413675 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}})  2025-08-29 17:36:38.413682 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.413689 | orchestrator | 2025-08-29 17:36:38.413696 | orchestrator | TASK [proxysql-config : Copying over grafana ProxySQL users config] ************ 2025-08-29 17:36:38.413702 | orchestrator | Friday 29 August 2025 17:32:38 +0000 (0:00:00.626) 0:02:15.240 ********* 2025-08-29 17:36:38.413709 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.413716 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.413723 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.413729 | orchestrator | 2025-08-29 17:36:38.413736 | orchestrator | TASK [proxysql-config : Copying over grafana ProxySQL rules config] ************ 2025-08-29 17:36:38.413743 | orchestrator | Friday 29 August 2025 17:32:39 +0000 (0:00:01.347) 0:02:16.587 ********* 2025-08-29 17:36:38.413750 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.413757 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.413764 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.413770 | orchestrator | 2025-08-29 17:36:38.413781 | orchestrator | TASK [include_role : heat] ***************************************************** 2025-08-29 17:36:38.413788 | orchestrator | Friday 29 August 2025 17:32:41 +0000 (0:00:02.028) 0:02:18.616 ********* 2025-08-29 17:36:38.413795 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.413801 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.413808 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.413814 | orchestrator | 2025-08-29 17:36:38.413821 | orchestrator | TASK [include_role : horizon] ************************************************** 2025-08-29 17:36:38.413828 | orchestrator | Friday 29 August 2025 17:32:42 +0000 (0:00:00.480) 0:02:19.096 ********* 2025-08-29 17:36:38.413835 | orchestrator | included: horizon for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.413841 | orchestrator | 2025-08-29 17:36:38.413848 | orchestrator | TASK [haproxy-config : Copying over horizon haproxy config] ******************** 2025-08-29 17:36:38.413855 | orchestrator | Friday 29 August 2025 17:32:43 +0000 (0:00:00.861) 0:02:19.958 ********* 2025-08-29 17:36:38.413866 | orchestrator | changed: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:36:38.413887 | orchestrator | changed: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:36:38.413897 | orchestrator | changed: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:36:38.413911 | orchestrator | 2025-08-29 17:36:38.413918 | orchestrator | TASK [haproxy-config : Add configuration for horizon when using single external frontend] *** 2025-08-29 17:36:38.413925 | orchestrator | Friday 29 August 2025 17:32:46 +0000 (0:00:03.329) 0:02:23.288 ********* 2025-08-29 17:36:38.413940 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:36:38.413949 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.413956 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:36:38.413968 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.413983 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:36:38.413991 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.414002 | orchestrator | 2025-08-29 17:36:38.414009 | orchestrator | TASK [haproxy-config : Configuring firewall for horizon] *********************** 2025-08-29 17:36:38.414050 | orchestrator | Friday 29 August 2025 17:32:47 +0000 (0:00:00.941) 0:02:24.229 ********* 2025-08-29 17:36:38.414058 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}})  2025-08-29 17:36:38.414067 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}})  2025-08-29 17:36:38.414076 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}})  2025-08-29 17:36:38.414084 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}})  2025-08-29 17:36:38.414092 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}})  2025-08-29 17:36:38.414098 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}})  2025-08-29 17:36:38.414106 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}})  2025-08-29 17:36:38.414113 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.414124 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}})  2025-08-29 17:36:38.414131 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}})  2025-08-29 17:36:38.414141 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}})  2025-08-29 17:36:38.414149 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.414155 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}})  2025-08-29 17:36:38.414162 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}})  2025-08-29 17:36:38.414174 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}})  2025-08-29 17:36:38.414181 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}})  2025-08-29 17:36:38.414188 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}})  2025-08-29 17:36:38.414195 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.414201 | orchestrator | 2025-08-29 17:36:38.414208 | orchestrator | TASK [proxysql-config : Copying over horizon ProxySQL users config] ************ 2025-08-29 17:36:38.414215 | orchestrator | Friday 29 August 2025 17:32:48 +0000 (0:00:00.835) 0:02:25.065 ********* 2025-08-29 17:36:38.414222 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.414228 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.414235 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.414241 | orchestrator | 2025-08-29 17:36:38.414248 | orchestrator | TASK [proxysql-config : Copying over horizon ProxySQL rules config] ************ 2025-08-29 17:36:38.414255 | orchestrator | Friday 29 August 2025 17:32:49 +0000 (0:00:01.230) 0:02:26.295 ********* 2025-08-29 17:36:38.414261 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.414268 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.414274 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.414281 | orchestrator | 2025-08-29 17:36:38.414288 | orchestrator | TASK [include_role : influxdb] ************************************************* 2025-08-29 17:36:38.414294 | orchestrator | Friday 29 August 2025 17:32:51 +0000 (0:00:01.969) 0:02:28.265 ********* 2025-08-29 17:36:38.414301 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.414308 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.414315 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.414321 | orchestrator | 2025-08-29 17:36:38.414328 | orchestrator | TASK [include_role : ironic] *************************************************** 2025-08-29 17:36:38.414335 | orchestrator | Friday 29 August 2025 17:32:52 +0000 (0:00:00.555) 0:02:28.820 ********* 2025-08-29 17:36:38.414341 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.414348 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.414355 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.414361 | orchestrator | 2025-08-29 17:36:38.414368 | orchestrator | TASK [include_role : keystone] ************************************************* 2025-08-29 17:36:38.414390 | orchestrator | Friday 29 August 2025 17:32:52 +0000 (0:00:00.329) 0:02:29.150 ********* 2025-08-29 17:36:38.414397 | orchestrator | included: keystone for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.414403 | orchestrator | 2025-08-29 17:36:38.414410 | orchestrator | TASK [haproxy-config : Copying over keystone haproxy config] ******************* 2025-08-29 17:36:38.414416 | orchestrator | Friday 29 August 2025 17:32:53 +0000 (0:00:00.944) 0:02:30.094 ********* 2025-08-29 17:36:38.414439 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:36:38.414453 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:36:38.414461 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:36:38.414468 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:36:38.414476 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:36:38.414488 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:36:38.414514 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:36:38.414522 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:36:38.414529 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:36:38.414536 | orchestrator | 2025-08-29 17:36:38.414543 | orchestrator | TASK [haproxy-config : Add configuration for keystone when using single external frontend] *** 2025-08-29 17:36:38.414550 | orchestrator | Friday 29 August 2025 17:32:57 +0000 (0:00:03.637) 0:02:33.732 ********* 2025-08-29 17:36:38.414557 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:36:38.414568 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:36:38.414584 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:36:38.414591 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.414599 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:36:38.414606 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:36:38.414613 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:36:38.414620 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.414632 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:36:38.414650 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:36:38.414657 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:36:38.414664 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.414671 | orchestrator | 2025-08-29 17:36:38.414678 | orchestrator | TASK [haproxy-config : Configuring firewall for keystone] ********************** 2025-08-29 17:36:38.414685 | orchestrator | Friday 29 August 2025 17:32:57 +0000 (0:00:00.724) 0:02:34.456 ********* 2025-08-29 17:36:38.414692 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}})  2025-08-29 17:36:38.414699 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}})  2025-08-29 17:36:38.414706 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.414736 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}})  2025-08-29 17:36:38.414744 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}})  2025-08-29 17:36:38.414751 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.414758 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}})  2025-08-29 17:36:38.414766 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}})  2025-08-29 17:36:38.414780 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.414787 | orchestrator | 2025-08-29 17:36:38.414794 | orchestrator | TASK [proxysql-config : Copying over keystone ProxySQL users config] *********** 2025-08-29 17:36:38.414800 | orchestrator | Friday 29 August 2025 17:32:58 +0000 (0:00:00.972) 0:02:35.429 ********* 2025-08-29 17:36:38.414807 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.414814 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.414820 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.414827 | orchestrator | 2025-08-29 17:36:38.414834 | orchestrator | TASK [proxysql-config : Copying over keystone ProxySQL rules config] *********** 2025-08-29 17:36:38.414840 | orchestrator | Friday 29 August 2025 17:33:00 +0000 (0:00:01.551) 0:02:36.980 ********* 2025-08-29 17:36:38.414847 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.414854 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.414860 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.414867 | orchestrator | 2025-08-29 17:36:38.414874 | orchestrator | TASK [include_role : letsencrypt] ********************************************** 2025-08-29 17:36:38.414885 | orchestrator | Friday 29 August 2025 17:33:02 +0000 (0:00:02.145) 0:02:39.125 ********* 2025-08-29 17:36:38.414892 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.414898 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.414905 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.414911 | orchestrator | 2025-08-29 17:36:38.414918 | orchestrator | TASK [include_role : magnum] *************************************************** 2025-08-29 17:36:38.414925 | orchestrator | Friday 29 August 2025 17:33:02 +0000 (0:00:00.321) 0:02:39.447 ********* 2025-08-29 17:36:38.414932 | orchestrator | included: magnum for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.414938 | orchestrator | 2025-08-29 17:36:38.414945 | orchestrator | TASK [haproxy-config : Copying over magnum haproxy config] ********************* 2025-08-29 17:36:38.414951 | orchestrator | Friday 29 August 2025 17:33:03 +0000 (0:00:01.032) 0:02:40.479 ********* 2025-08-29 17:36:38.414959 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:36:38.414966 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.414989 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:36:38.415002 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415020 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:36:38.415037 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415044 | orchestrator | 2025-08-29 17:36:38.415051 | orchestrator | TASK [haproxy-config : Add configuration for magnum when using single external frontend] *** 2025-08-29 17:36:38.415058 | orchestrator | Friday 29 August 2025 17:33:08 +0000 (0:00:04.393) 0:02:44.872 ********* 2025-08-29 17:36:38.415065 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:36:38.415077 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415084 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.415095 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:36:38.415106 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415113 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.415120 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:36:38.415127 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415138 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.415145 | orchestrator | 2025-08-29 17:36:38.415151 | orchestrator | TASK [haproxy-config : Configuring firewall for magnum] ************************ 2025-08-29 17:36:38.415158 | orchestrator | Friday 29 August 2025 17:33:08 +0000 (0:00:00.586) 0:02:45.459 ********* 2025-08-29 17:36:38.415166 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}})  2025-08-29 17:36:38.415173 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}})  2025-08-29 17:36:38.415180 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.415186 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}})  2025-08-29 17:36:38.415193 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}})  2025-08-29 17:36:38.415200 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.415207 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}})  2025-08-29 17:36:38.415214 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}})  2025-08-29 17:36:38.415224 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.415231 | orchestrator | 2025-08-29 17:36:38.415238 | orchestrator | TASK [proxysql-config : Copying over magnum ProxySQL users config] ************* 2025-08-29 17:36:38.415244 | orchestrator | Friday 29 August 2025 17:33:09 +0000 (0:00:00.814) 0:02:46.273 ********* 2025-08-29 17:36:38.415251 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.415258 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.415264 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.415271 | orchestrator | 2025-08-29 17:36:38.415277 | orchestrator | TASK [proxysql-config : Copying over magnum ProxySQL rules config] ************* 2025-08-29 17:36:38.415284 | orchestrator | Friday 29 August 2025 17:33:11 +0000 (0:00:01.433) 0:02:47.706 ********* 2025-08-29 17:36:38.415291 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.415297 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.415304 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.415310 | orchestrator | 2025-08-29 17:36:38.415317 | orchestrator | TASK [include_role : manila] *************************************************** 2025-08-29 17:36:38.415328 | orchestrator | Friday 29 August 2025 17:33:12 +0000 (0:00:01.924) 0:02:49.631 ********* 2025-08-29 17:36:38.415335 | orchestrator | included: manila for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.415341 | orchestrator | 2025-08-29 17:36:38.415348 | orchestrator | TASK [haproxy-config : Copying over manila haproxy config] ********************* 2025-08-29 17:36:38.415355 | orchestrator | Friday 29 August 2025 17:33:14 +0000 (0:00:01.044) 0:02:50.675 ********* 2025-08-29 17:36:38.415362 | orchestrator | changed: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) 2025-08-29 17:36:38.415391 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415399 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415406 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415418 | orchestrator | changed: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) 2025-08-29 17:36:38.415428 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415441 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415448 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415455 | orchestrator | changed: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) 2025-08-29 17:36:38.415462 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415474 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415484 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415495 | orchestrator | 2025-08-29 17:36:38.415502 | orchestrator | TASK [haproxy-config : Add configuration for manila when using single external frontend] *** 2025-08-29 17:36:38.415509 | orchestrator | Friday 29 August 2025 17:33:17 +0000 (0:00:03.941) 0:02:54.616 ********* 2025-08-29 17:36:38.415516 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})  2025-08-29 17:36:38.415523 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415530 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415537 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415543 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.415558 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})  2025-08-29 17:36:38.415570 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415577 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415584 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415591 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.415598 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/release/manila-api:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})  2025-08-29 17:36:38.415608 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/release/manila-scheduler:19.0.2.20250711', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415621 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/release/manila-share:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415632 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/release/manila-data:19.0.2.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.415639 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.415646 | orchestrator | 2025-08-29 17:36:38.415653 | orchestrator | TASK [haproxy-config : Configuring firewall for manila] ************************ 2025-08-29 17:36:38.415660 | orchestrator | Friday 29 August 2025 17:33:19 +0000 (0:00:01.075) 0:02:55.692 ********* 2025-08-29 17:36:38.415666 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}})  2025-08-29 17:36:38.415673 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}})  2025-08-29 17:36:38.415680 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.415686 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}})  2025-08-29 17:36:38.415693 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}})  2025-08-29 17:36:38.415700 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.415707 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}})  2025-08-29 17:36:38.415713 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}})  2025-08-29 17:36:38.415720 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.415727 | orchestrator | 2025-08-29 17:36:38.415733 | orchestrator | TASK [proxysql-config : Copying over manila ProxySQL users config] ************* 2025-08-29 17:36:38.415740 | orchestrator | Friday 29 August 2025 17:33:19 +0000 (0:00:00.864) 0:02:56.557 ********* 2025-08-29 17:36:38.415747 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.415753 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.415760 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.415767 | orchestrator | 2025-08-29 17:36:38.415773 | orchestrator | TASK [proxysql-config : Copying over manila ProxySQL rules config] ************* 2025-08-29 17:36:38.415780 | orchestrator | Friday 29 August 2025 17:33:21 +0000 (0:00:01.239) 0:02:57.797 ********* 2025-08-29 17:36:38.415787 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.415793 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.415800 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.415806 | orchestrator | 2025-08-29 17:36:38.415813 | orchestrator | TASK [include_role : mariadb] ************************************************** 2025-08-29 17:36:38.415824 | orchestrator | Friday 29 August 2025 17:33:23 +0000 (0:00:01.964) 0:02:59.761 ********* 2025-08-29 17:36:38.415831 | orchestrator | included: mariadb for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.415837 | orchestrator | 2025-08-29 17:36:38.415844 | orchestrator | TASK [mariadb : Ensure mysql monitor user exist] ******************************* 2025-08-29 17:36:38.415851 | orchestrator | Friday 29 August 2025 17:33:24 +0000 (0:00:01.274) 0:03:01.035 ********* 2025-08-29 17:36:38.415858 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:36:38.415864 | orchestrator | 2025-08-29 17:36:38.415871 | orchestrator | TASK [haproxy-config : Copying over mariadb haproxy config] ******************** 2025-08-29 17:36:38.415878 | orchestrator | Friday 29 August 2025 17:33:27 +0000 (0:00:02.725) 0:03:03.761 ********* 2025-08-29 17:36:38.415895 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:36:38.415904 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}})  2025-08-29 17:36:38.415911 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.415922 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:36:38.415937 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}})  2025-08-29 17:36:38.415944 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.415951 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:36:38.415959 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}})  2025-08-29 17:36:38.415970 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.415977 | orchestrator | 2025-08-29 17:36:38.415984 | orchestrator | TASK [haproxy-config : Add configuration for mariadb when using single external frontend] *** 2025-08-29 17:36:38.415990 | orchestrator | Friday 29 August 2025 17:33:29 +0000 (0:00:02.561) 0:03:06.323 ********* 2025-08-29 17:36:38.416006 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:36:38.416014 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}})  2025-08-29 17:36:38.416022 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416029 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:36:38.416043 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}})  2025-08-29 17:36:38.416051 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416062 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:36:38.416069 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/release/mariadb-clustercheck:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}})  2025-08-29 17:36:38.416080 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416087 | orchestrator | 2025-08-29 17:36:38.416094 | orchestrator | TASK [haproxy-config : Configuring firewall for mariadb] *********************** 2025-08-29 17:36:38.416101 | orchestrator | Friday 29 August 2025 17:33:32 +0000 (0:00:03.212) 0:03:09.535 ********* 2025-08-29 17:36:38.416108 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}})  2025-08-29 17:36:38.416257 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}})  2025-08-29 17:36:38.416268 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416280 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}})  2025-08-29 17:36:38.416287 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}})  2025-08-29 17:36:38.416294 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416301 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}})  2025-08-29 17:36:38.416308 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}})  2025-08-29 17:36:38.416361 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416369 | orchestrator | 2025-08-29 17:36:38.416391 | orchestrator | TASK [proxysql-config : Copying over mariadb ProxySQL users config] ************ 2025-08-29 17:36:38.416398 | orchestrator | Friday 29 August 2025 17:33:36 +0000 (0:00:03.149) 0:03:12.685 ********* 2025-08-29 17:36:38.416404 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.416411 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.416418 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.416424 | orchestrator | 2025-08-29 17:36:38.416431 | orchestrator | TASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************ 2025-08-29 17:36:38.416438 | orchestrator | Friday 29 August 2025 17:33:38 +0000 (0:00:02.684) 0:03:15.370 ********* 2025-08-29 17:36:38.416444 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416451 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416457 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416464 | orchestrator | 2025-08-29 17:36:38.416471 | orchestrator | TASK [include_role : masakari] ************************************************* 2025-08-29 17:36:38.416477 | orchestrator | Friday 29 August 2025 17:33:40 +0000 (0:00:01.791) 0:03:17.161 ********* 2025-08-29 17:36:38.416484 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416491 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416497 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416504 | orchestrator | 2025-08-29 17:36:38.416511 | orchestrator | TASK [include_role : memcached] ************************************************ 2025-08-29 17:36:38.416517 | orchestrator | Friday 29 August 2025 17:33:41 +0000 (0:00:00.563) 0:03:17.724 ********* 2025-08-29 17:36:38.416524 | orchestrator | included: memcached for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.416531 | orchestrator | 2025-08-29 17:36:38.416538 | orchestrator | TASK [haproxy-config : Copying over memcached haproxy config] ****************** 2025-08-29 17:36:38.416544 | orchestrator | Friday 29 August 2025 17:33:42 +0000 (0:00:01.090) 0:03:18.815 ********* 2025-08-29 17:36:38.416606 | orchestrator | changed: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.18.20250711', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) 2025-08-29 17:36:38.416624 | orchestrator | changed: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.18.20250711', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) 2025-08-29 17:36:38.416632 | orchestrator | changed: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.18.20250711', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) 2025-08-29 17:36:38.416644 | orchestrator | 2025-08-29 17:36:38.416651 | orchestrator | TASK [haproxy-config : Add configuration for memcached when using single external frontend] *** 2025-08-29 17:36:38.416658 | orchestrator | Friday 29 August 2025 17:33:43 +0000 (0:00:01.783) 0:03:20.599 ********* 2025-08-29 17:36:38.416665 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.18.20250711', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})  2025-08-29 17:36:38.416672 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.18.20250711', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})  2025-08-29 17:36:38.416679 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416686 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416740 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/release/memcached:1.6.18.20250711', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})  2025-08-29 17:36:38.416750 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416756 | orchestrator | 2025-08-29 17:36:38.416763 | orchestrator | TASK [haproxy-config : Configuring firewall for memcached] ********************* 2025-08-29 17:36:38.416770 | orchestrator | Friday 29 August 2025 17:33:44 +0000 (0:00:00.656) 0:03:21.255 ********* 2025-08-29 17:36:38.416785 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}})  2025-08-29 17:36:38.416794 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}})  2025-08-29 17:36:38.416805 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416812 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416819 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}})  2025-08-29 17:36:38.416826 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416832 | orchestrator | 2025-08-29 17:36:38.416839 | orchestrator | TASK [proxysql-config : Copying over memcached ProxySQL users config] ********** 2025-08-29 17:36:38.416846 | orchestrator | Friday 29 August 2025 17:33:45 +0000 (0:00:00.591) 0:03:21.847 ********* 2025-08-29 17:36:38.416853 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416860 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416867 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416893 | orchestrator | 2025-08-29 17:36:38.416900 | orchestrator | TASK [proxysql-config : Copying over memcached ProxySQL rules config] ********** 2025-08-29 17:36:38.416907 | orchestrator | Friday 29 August 2025 17:33:45 +0000 (0:00:00.464) 0:03:22.311 ********* 2025-08-29 17:36:38.416914 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416921 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416927 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416934 | orchestrator | 2025-08-29 17:36:38.416941 | orchestrator | TASK [include_role : mistral] ************************************************** 2025-08-29 17:36:38.416947 | orchestrator | Friday 29 August 2025 17:33:46 +0000 (0:00:01.292) 0:03:23.604 ********* 2025-08-29 17:36:38.416954 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.416961 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.416967 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.416974 | orchestrator | 2025-08-29 17:36:38.416981 | orchestrator | TASK [include_role : neutron] ************************************************** 2025-08-29 17:36:38.416987 | orchestrator | Friday 29 August 2025 17:33:47 +0000 (0:00:00.539) 0:03:24.143 ********* 2025-08-29 17:36:38.416994 | orchestrator | included: neutron for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.417001 | orchestrator | 2025-08-29 17:36:38.417008 | orchestrator | TASK [haproxy-config : Copying over neutron haproxy config] ******************** 2025-08-29 17:36:38.417014 | orchestrator | Friday 29 August 2025 17:33:48 +0000 (0:00:01.170) 0:03:25.314 ********* 2025-08-29 17:36:38.417021 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:36:38.417081 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.1.20250711', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417101 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417109 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417128 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', "healthcheck_port 'neutron-l3-agent ' 5672"], 'timeout': '30'}}})  2025-08-29 17:36:38.417136 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:36:38.417191 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417211 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.1.20250711', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417219 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.1.20250711', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417226 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417233 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417240 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417292 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417310 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', "healthcheck_port 'neutron-l3-agent ' 5672"], 'timeout': '30'}}})  2025-08-29 17:36:38.417318 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417325 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.417332 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.1.20250711', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417339 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417346 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417420 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417435 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417442 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.417450 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417457 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417464 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417512 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.1.20250711', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417531 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417538 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.1.20250711', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417546 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.417555 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.417562 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.417617 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.417631 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-ovn-vpn-agent', 'value': {'container_name': 'neutron_ovn_vpn_agent', 'image': 'registry.osism.tech/kolla/release/neutron-ovn-vpn-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-ovn-vpn-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-vpn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port python 6642', '&&', 'healthcheck_port neutron-ovn-vpn-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417638 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-ovn-vpn-agent', 'value': {'container_name': 'neutron_ovn_vpn_agent', 'image': 'registry.osism.tech/kolla/release/neutron-ovn-vpn-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-ovn-vpn-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-vpn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port python 6642', '&&', 'healthcheck_port neutron-ovn-vpn-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417645 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:36:38.417653 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.1.20250711', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417709 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417725 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417733 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', "healthcheck_port 'neutron-l3-agent ' 5672"], 'timeout': '30'}}})  2025-08-29 17:36:38.417740 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417747 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.1.20250711', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417755 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417825 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417840 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.417858 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417865 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417873 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.417880 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.1.20250711', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417936 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.417951 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.417959 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-ovn-vpn-agent', 'value': {'container_name': 'neutron_ovn_vpn_agent', 'image': 'registry.osism.tech/kolla/release/neutron-ovn-vpn-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-ovn-vpn-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-vpn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port python 6642', '&&', 'healthcheck_port neutron-ovn-vpn-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.417966 | orchestrator | 2025-08-29 17:36:38.417973 | orchestrator | TASK [haproxy-config : Add configuration for neutron when using single external frontend] *** 2025-08-29 17:36:38.417980 | orchestrator | Friday 29 August 2025 17:33:53 +0000 (0:00:04.542) 0:03:29.857 ********* 2025-08-29 17:36:38.417996 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:36:38.418003 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.1.20250711', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418126 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418143 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418151 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', "healthcheck_port 'neutron-l3-agent ' 5672"], 'timeout': '30'}}})  2025-08-29 17:36:38.418158 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418165 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.1.20250711', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418177 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:36:38.418230 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418244 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.1.20250711', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418252 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:36:38.418259 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418271 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418279 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/release/neutron-openvswitch-agent:25.2.1.20250711', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418340 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418352 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.418424 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/release/neutron-linuxbridge-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418434 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418448 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', "healthcheck_port 'neutron-l3-agent ' 5672"], 'timeout': '30'}}})  2025-08-29 17:36:38.418505 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/release/neutron-dhcp-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418529 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418537 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418544 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418551 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.1.20250711', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418564 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/release/neutron-l3-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', "healthcheck_port 'neutron-l3-agent ' 5672"], 'timeout': '30'}}})  2025-08-29 17:36:38.418571 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418624 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/release/neutron-sriov-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418638 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.1.20250711', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418645 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418653 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.418669 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/release/neutron-mlnx-agent:25.2.1.20250711', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418677 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.418715 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418727 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/release/neutron-eswitchd:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418735 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418759 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418771 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.418779 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418831 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.418845 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-ovn-vpn-agent', 'value': {'container_name': 'neutron_ovn_vpn_agent', 'image': 'registry.osism.tech/kolla/release/neutron-ovn-vpn-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-ovn-vpn-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-vpn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port python 6642', '&&', 'healthcheck_port neutron-ovn-vpn-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418852 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.1.20250711', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418860 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/release/neutron-bgp-dragent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418872 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.418889 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.418897 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/release/neutron-infoblox-ipam-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418925 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.418938 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metering-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}})  2025-08-29 17:36:38.418945 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-ovn-vpn-agent', 'value': {'container_name': 'neutron_ovn_vpn_agent', 'image': 'registry.osism.tech/kolla/release/neutron-ovn-vpn-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-ovn-vpn-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-vpn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port python 6642', '&&', 'healthcheck_port neutron-ovn-vpn-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418957 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/release/ironic-neutron-agent:25.2.1.20250711', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.418964 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.418971 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/neutron-tls-proxy:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}})  2025-08-29 17:36:38.418978 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/release/neutron-ovn-agent:25.2.1.20250711', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:36:38.419005 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-ovn-vpn-agent', 'value': {'container_name': 'neutron_ovn_vpn_agent', 'image': 'registry.osism.tech/kolla/release/neutron-ovn-vpn-agent:25.2.1.20250711', 'privileged': True, 'enabled': False, 'group': 'neutron-ovn-vpn-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-vpn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port python 6642', '&&', 'healthcheck_port neutron-ovn-vpn-agent 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.419017 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.419027 | orchestrator | 2025-08-29 17:36:38.419043 | orchestrator | TASK [haproxy-config : Configuring firewall for neutron] *********************** 2025-08-29 17:36:38.419054 | orchestrator | Friday 29 August 2025 17:33:55 +0000 (0:00:01.906) 0:03:31.764 ********* 2025-08-29 17:36:38.419064 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}})  2025-08-29 17:36:38.419076 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}})  2025-08-29 17:36:38.419093 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.419182 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}})  2025-08-29 17:36:38.419197 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}})  2025-08-29 17:36:38.419207 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.419217 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}})  2025-08-29 17:36:38.419227 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}})  2025-08-29 17:36:38.419237 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.419248 | orchestrator | 2025-08-29 17:36:38.419259 | orchestrator | TASK [proxysql-config : Copying over neutron ProxySQL users config] ************ 2025-08-29 17:36:38.419271 | orchestrator | Friday 29 August 2025 17:33:56 +0000 (0:00:01.664) 0:03:33.428 ********* 2025-08-29 17:36:38.419281 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.419292 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.419303 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.419313 | orchestrator | 2025-08-29 17:36:38.419324 | orchestrator | TASK [proxysql-config : Copying over neutron ProxySQL rules config] ************ 2025-08-29 17:36:38.419335 | orchestrator | Friday 29 August 2025 17:33:57 +0000 (0:00:01.218) 0:03:34.647 ********* 2025-08-29 17:36:38.419346 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.419356 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.419368 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.419396 | orchestrator | 2025-08-29 17:36:38.419403 | orchestrator | TASK [include_role : placement] ************************************************ 2025-08-29 17:36:38.419410 | orchestrator | Friday 29 August 2025 17:34:00 +0000 (0:00:02.060) 0:03:36.708 ********* 2025-08-29 17:36:38.419417 | orchestrator | included: placement for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.419423 | orchestrator | 2025-08-29 17:36:38.419430 | orchestrator | TASK [haproxy-config : Copying over placement haproxy config] ****************** 2025-08-29 17:36:38.419436 | orchestrator | Friday 29 August 2025 17:34:01 +0000 (0:00:01.501) 0:03:38.209 ********* 2025-08-29 17:36:38.419444 | orchestrator | changed: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.419504 | orchestrator | changed: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.419523 | orchestrator | changed: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.419530 | orchestrator | 2025-08-29 17:36:38.419537 | orchestrator | TASK [haproxy-config : Add configuration for placement when using single external frontend] *** 2025-08-29 17:36:38.419544 | orchestrator | Friday 29 August 2025 17:34:04 +0000 (0:00:03.426) 0:03:41.636 ********* 2025-08-29 17:36:38.419551 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.419558 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.419565 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.419572 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.419599 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.419616 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.419623 | orchestrator | 2025-08-29 17:36:38.419629 | orchestrator | TASK [haproxy-config : Configuring firewall for placement] ********************* 2025-08-29 17:36:38.419636 | orchestrator | Friday 29 August 2025 17:34:05 +0000 (0:00:00.994) 0:03:42.630 ********* 2025-08-29 17:36:38.419643 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}})  2025-08-29 17:36:38.419650 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}})  2025-08-29 17:36:38.419657 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.419663 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}})  2025-08-29 17:36:38.419670 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}})  2025-08-29 17:36:38.419677 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.419684 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}})  2025-08-29 17:36:38.419690 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}})  2025-08-29 17:36:38.419697 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.419704 | orchestrator | 2025-08-29 17:36:38.419710 | orchestrator | TASK [proxysql-config : Copying over placement ProxySQL users config] ********** 2025-08-29 17:36:38.419717 | orchestrator | Friday 29 August 2025 17:34:06 +0000 (0:00:00.781) 0:03:43.411 ********* 2025-08-29 17:36:38.419723 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.419730 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.419736 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.419743 | orchestrator | 2025-08-29 17:36:38.419749 | orchestrator | TASK [proxysql-config : Copying over placement ProxySQL rules config] ********** 2025-08-29 17:36:38.419756 | orchestrator | Friday 29 August 2025 17:34:08 +0000 (0:00:01.262) 0:03:44.674 ********* 2025-08-29 17:36:38.419763 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.419769 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.419776 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.419783 | orchestrator | 2025-08-29 17:36:38.419789 | orchestrator | TASK [include_role : nova] ***************************************************** 2025-08-29 17:36:38.419796 | orchestrator | Friday 29 August 2025 17:34:10 +0000 (0:00:02.080) 0:03:46.755 ********* 2025-08-29 17:36:38.419803 | orchestrator | included: nova for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.419809 | orchestrator | 2025-08-29 17:36:38.419816 | orchestrator | TASK [haproxy-config : Copying over nova haproxy config] *********************** 2025-08-29 17:36:38.419827 | orchestrator | Friday 29 August 2025 17:34:11 +0000 (0:00:01.442) 0:03:48.197 ********* 2025-08-29 17:36:38.419857 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.419871 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.419880 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.419889 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.419898 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.419915 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.419946 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.419956 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.419964 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.419972 | orchestrator | 2025-08-29 17:36:38.419980 | orchestrator | TASK [haproxy-config : Add configuration for nova when using single external frontend] *** 2025-08-29 17:36:38.419987 | orchestrator | Friday 29 August 2025 17:34:15 +0000 (0:00:04.244) 0:03:52.442 ********* 2025-08-29 17:36:38.419995 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.420027 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.420039 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.420047 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.420055 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.420064 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.420077 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.420084 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.420115 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.420124 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.420132 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/release/nova-super-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.420139 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.420147 | orchestrator | 2025-08-29 17:36:38.420154 | orchestrator | TASK [haproxy-config : Configuring firewall for nova] ************************** 2025-08-29 17:36:38.420161 | orchestrator | Friday 29 August 2025 17:34:16 +0000 (0:00:00.663) 0:03:53.106 ********* 2025-08-29 17:36:38.420169 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420182 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420191 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420198 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420206 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.420212 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420219 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420226 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420233 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420240 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.420265 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420273 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420280 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420290 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}})  2025-08-29 17:36:38.420297 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.420304 | orchestrator | 2025-08-29 17:36:38.420311 | orchestrator | TASK [proxysql-config : Copying over nova ProxySQL users config] *************** 2025-08-29 17:36:38.420318 | orchestrator | Friday 29 August 2025 17:34:17 +0000 (0:00:01.259) 0:03:54.366 ********* 2025-08-29 17:36:38.420324 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.420331 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.420338 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.420359 | orchestrator | 2025-08-29 17:36:38.420366 | orchestrator | TASK [proxysql-config : Copying over nova ProxySQL rules config] *************** 2025-08-29 17:36:38.420391 | orchestrator | Friday 29 August 2025 17:34:19 +0000 (0:00:01.599) 0:03:55.965 ********* 2025-08-29 17:36:38.420399 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.420406 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.420412 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.420419 | orchestrator | 2025-08-29 17:36:38.420425 | orchestrator | TASK [include_role : nova-cell] ************************************************ 2025-08-29 17:36:38.420432 | orchestrator | Friday 29 August 2025 17:34:21 +0000 (0:00:02.086) 0:03:58.052 ********* 2025-08-29 17:36:38.420443 | orchestrator | included: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.420450 | orchestrator | 2025-08-29 17:36:38.420457 | orchestrator | TASK [nova-cell : Configure loadbalancer for nova-novncproxy] ****************** 2025-08-29 17:36:38.420463 | orchestrator | Friday 29 August 2025 17:34:22 +0000 (0:00:01.579) 0:03:59.631 ********* 2025-08-29 17:36:38.420470 | orchestrator | included: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy) 2025-08-29 17:36:38.420477 | orchestrator | 2025-08-29 17:36:38.420484 | orchestrator | TASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] *** 2025-08-29 17:36:38.420491 | orchestrator | Friday 29 August 2025 17:34:23 +0000 (0:00:00.830) 0:04:00.461 ********* 2025-08-29 17:36:38.420498 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) 2025-08-29 17:36:38.420506 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) 2025-08-29 17:36:38.420513 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) 2025-08-29 17:36:38.420520 | orchestrator | 2025-08-29 17:36:38.420527 | orchestrator | TASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] *** 2025-08-29 17:36:38.420534 | orchestrator | Friday 29 August 2025 17:34:27 +0000 (0:00:04.076) 0:04:04.538 ********* 2025-08-29 17:36:38.420564 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420573 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.420583 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420591 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420605 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.420612 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.420618 | orchestrator | 2025-08-29 17:36:38.420625 | orchestrator | TASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] ***** 2025-08-29 17:36:38.420632 | orchestrator | Friday 29 August 2025 17:34:29 +0000 (0:00:01.433) 0:04:05.971 ********* 2025-08-29 17:36:38.420639 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}})  2025-08-29 17:36:38.420646 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}})  2025-08-29 17:36:38.420653 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.420660 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}})  2025-08-29 17:36:38.420667 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}})  2025-08-29 17:36:38.420674 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.420680 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}})  2025-08-29 17:36:38.420687 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}})  2025-08-29 17:36:38.420694 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.420701 | orchestrator | 2025-08-29 17:36:38.420707 | orchestrator | TASK [proxysql-config : Copying over nova-cell ProxySQL users config] ********** 2025-08-29 17:36:38.420714 | orchestrator | Friday 29 August 2025 17:34:30 +0000 (0:00:01.579) 0:04:07.551 ********* 2025-08-29 17:36:38.420721 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.420727 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.420734 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.420740 | orchestrator | 2025-08-29 17:36:38.420747 | orchestrator | TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] ********** 2025-08-29 17:36:38.420753 | orchestrator | Friday 29 August 2025 17:34:33 +0000 (0:00:02.719) 0:04:10.271 ********* 2025-08-29 17:36:38.420760 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.420767 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.420773 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.420780 | orchestrator | 2025-08-29 17:36:38.420786 | orchestrator | TASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] ************* 2025-08-29 17:36:38.420793 | orchestrator | Friday 29 August 2025 17:34:36 +0000 (0:00:03.073) 0:04:13.344 ********* 2025-08-29 17:36:38.420800 | orchestrator | included: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy) 2025-08-29 17:36:38.420807 | orchestrator | 2025-08-29 17:36:38.420814 | orchestrator | TASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] *** 2025-08-29 17:36:38.420843 | orchestrator | Friday 29 August 2025 17:34:38 +0000 (0:00:01.550) 0:04:14.895 ********* 2025-08-29 17:36:38.420857 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420864 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.420874 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420882 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.420888 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420895 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.420902 | orchestrator | 2025-08-29 17:36:38.420909 | orchestrator | TASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] *** 2025-08-29 17:36:38.420915 | orchestrator | Friday 29 August 2025 17:34:39 +0000 (0:00:01.364) 0:04:16.260 ********* 2025-08-29 17:36:38.420922 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420929 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.420936 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420943 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.420950 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}})  2025-08-29 17:36:38.420956 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.420967 | orchestrator | 2025-08-29 17:36:38.420974 | orchestrator | TASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] *** 2025-08-29 17:36:38.420981 | orchestrator | Friday 29 August 2025 17:34:41 +0000 (0:00:01.422) 0:04:17.682 ********* 2025-08-29 17:36:38.420988 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.420994 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.421001 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.421007 | orchestrator | 2025-08-29 17:36:38.421014 | orchestrator | TASK [proxysql-config : Copying over nova-cell ProxySQL users config] ********** 2025-08-29 17:36:38.421040 | orchestrator | Friday 29 August 2025 17:34:42 +0000 (0:00:01.906) 0:04:19.588 ********* 2025-08-29 17:36:38.421048 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.421055 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.421062 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.421069 | orchestrator | 2025-08-29 17:36:38.421075 | orchestrator | TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] ********** 2025-08-29 17:36:38.421082 | orchestrator | Friday 29 August 2025 17:34:45 +0000 (0:00:02.342) 0:04:21.931 ********* 2025-08-29 17:36:38.421089 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.421095 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.421102 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.421109 | orchestrator | 2025-08-29 17:36:38.421116 | orchestrator | TASK [nova-cell : Configure loadbalancer for nova-serialproxy] ***************** 2025-08-29 17:36:38.421122 | orchestrator | Friday 29 August 2025 17:34:48 +0000 (0:00:03.008) 0:04:24.940 ********* 2025-08-29 17:36:38.421132 | orchestrator | included: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy) 2025-08-29 17:36:38.421139 | orchestrator | 2025-08-29 17:36:38.421146 | orchestrator | TASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] *** 2025-08-29 17:36:38.421153 | orchestrator | Friday 29 August 2025 17:34:49 +0000 (0:00:00.832) 0:04:25.773 ********* 2025-08-29 17:36:38.421160 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}})  2025-08-29 17:36:38.421166 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.421173 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}})  2025-08-29 17:36:38.421180 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.421187 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}})  2025-08-29 17:36:38.421194 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.421200 | orchestrator | 2025-08-29 17:36:38.421207 | orchestrator | TASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] *** 2025-08-29 17:36:38.421214 | orchestrator | Friday 29 August 2025 17:34:50 +0000 (0:00:01.411) 0:04:27.184 ********* 2025-08-29 17:36:38.421227 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}})  2025-08-29 17:36:38.421234 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.421241 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}})  2025-08-29 17:36:38.421248 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.421276 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}})  2025-08-29 17:36:38.421284 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.421291 | orchestrator | 2025-08-29 17:36:38.421298 | orchestrator | TASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] **** 2025-08-29 17:36:38.421304 | orchestrator | Friday 29 August 2025 17:34:51 +0000 (0:00:01.399) 0:04:28.583 ********* 2025-08-29 17:36:38.421311 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.421318 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.421324 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.421331 | orchestrator | 2025-08-29 17:36:38.421341 | orchestrator | TASK [proxysql-config : Copying over nova-cell ProxySQL users config] ********** 2025-08-29 17:36:38.421348 | orchestrator | Friday 29 August 2025 17:34:53 +0000 (0:00:01.442) 0:04:30.026 ********* 2025-08-29 17:36:38.421355 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.421361 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.421368 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.421387 | orchestrator | 2025-08-29 17:36:38.421394 | orchestrator | TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] ********** 2025-08-29 17:36:38.421401 | orchestrator | Friday 29 August 2025 17:34:55 +0000 (0:00:02.495) 0:04:32.522 ********* 2025-08-29 17:36:38.421408 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.421414 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.421421 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.421428 | orchestrator | 2025-08-29 17:36:38.421434 | orchestrator | TASK [include_role : octavia] ************************************************** 2025-08-29 17:36:38.421441 | orchestrator | Friday 29 August 2025 17:34:58 +0000 (0:00:03.058) 0:04:35.581 ********* 2025-08-29 17:36:38.421448 | orchestrator | included: octavia for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.421454 | orchestrator | 2025-08-29 17:36:38.421461 | orchestrator | TASK [haproxy-config : Copying over octavia haproxy config] ******************** 2025-08-29 17:36:38.421468 | orchestrator | Friday 29 August 2025 17:35:00 +0000 (0:00:01.596) 0:04:37.178 ********* 2025-08-29 17:36:38.421475 | orchestrator | changed: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.421487 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})  2025-08-29 17:36:38.421494 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421557 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421571 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.421579 | orchestrator | changed: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.421591 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})  2025-08-29 17:36:38.421598 | orchestrator | changed: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.421605 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421631 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})  2025-08-29 17:36:38.421644 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421651 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421663 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421670 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.421677 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.421684 | orchestrator | 2025-08-29 17:36:38.421691 | orchestrator | TASK [haproxy-config : Add configuration for octavia when using single external frontend] *** 2025-08-29 17:36:38.421697 | orchestrator | Friday 29 August 2025 17:35:04 +0000 (0:00:03.697) 0:04:40.875 ********* 2025-08-29 17:36:38.421724 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.421736 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})  2025-08-29 17:36:38.421743 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421755 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421762 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.421769 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.421776 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.421802 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})  2025-08-29 17:36:38.421817 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421828 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421848 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.421859 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.421870 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-api:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.421881 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-driver-agent:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})  2025-08-29 17:36:38.421922 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-health-manager:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421940 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-housekeeping:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})  2025-08-29 17:36:38.421960 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/octavia-worker:15.0.1.20250711', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:36:38.421971 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.421981 | orchestrator | 2025-08-29 17:36:38.421992 | orchestrator | TASK [haproxy-config : Configuring firewall for octavia] *********************** 2025-08-29 17:36:38.422003 | orchestrator | Friday 29 August 2025 17:35:05 +0000 (0:00:01.112) 0:04:41.987 ********* 2025-08-29 17:36:38.422047 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}})  2025-08-29 17:36:38.422062 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}})  2025-08-29 17:36:38.422074 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.422086 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}})  2025-08-29 17:36:38.422098 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}})  2025-08-29 17:36:38.422110 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.422121 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}})  2025-08-29 17:36:38.422133 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}})  2025-08-29 17:36:38.422141 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.422147 | orchestrator | 2025-08-29 17:36:38.422154 | orchestrator | TASK [proxysql-config : Copying over octavia ProxySQL users config] ************ 2025-08-29 17:36:38.422161 | orchestrator | Friday 29 August 2025 17:35:06 +0000 (0:00:01.283) 0:04:43.271 ********* 2025-08-29 17:36:38.422167 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.422174 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.422180 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.422186 | orchestrator | 2025-08-29 17:36:38.422193 | orchestrator | TASK [proxysql-config : Copying over octavia ProxySQL rules config] ************ 2025-08-29 17:36:38.422200 | orchestrator | Friday 29 August 2025 17:35:08 +0000 (0:00:01.626) 0:04:44.897 ********* 2025-08-29 17:36:38.422206 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.422213 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.422219 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.422226 | orchestrator | 2025-08-29 17:36:38.422232 | orchestrator | TASK [include_role : opensearch] *********************************************** 2025-08-29 17:36:38.422239 | orchestrator | Friday 29 August 2025 17:35:10 +0000 (0:00:02.237) 0:04:47.135 ********* 2025-08-29 17:36:38.422246 | orchestrator | included: opensearch for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.422252 | orchestrator | 2025-08-29 17:36:38.422259 | orchestrator | TASK [haproxy-config : Copying over opensearch haproxy config] ***************** 2025-08-29 17:36:38.422265 | orchestrator | Friday 29 August 2025 17:35:12 +0000 (0:00:01.659) 0:04:48.794 ********* 2025-08-29 17:36:38.422313 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:36:38.422323 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:36:38.422330 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:36:38.422338 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:36:38.422364 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:36:38.422418 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:36:38.422433 | orchestrator | 2025-08-29 17:36:38.422444 | orchestrator | TASK [haproxy-config : Add configuration for opensearch when using single external frontend] *** 2025-08-29 17:36:38.422454 | orchestrator | Friday 29 August 2025 17:35:17 +0000 (0:00:05.027) 0:04:53.822 ********* 2025-08-29 17:36:38.422466 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:36:38.422479 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:36:38.422499 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.422546 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:36:38.422560 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:36:38.422568 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.422575 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:36:38.422582 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:36:38.422594 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.422601 | orchestrator | 2025-08-29 17:36:38.422607 | orchestrator | TASK [haproxy-config : Configuring firewall for opensearch] ******************** 2025-08-29 17:36:38.422614 | orchestrator | Friday 29 August 2025 17:35:17 +0000 (0:00:00.632) 0:04:54.454 ********* 2025-08-29 17:36:38.422621 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}})  2025-08-29 17:36:38.422688 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}})  2025-08-29 17:36:38.422698 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}})  2025-08-29 17:36:38.422705 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.422712 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}})  2025-08-29 17:36:38.422723 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}})  2025-08-29 17:36:38.422730 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}})  2025-08-29 17:36:38.422737 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.422744 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}})  2025-08-29 17:36:38.422751 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}})  2025-08-29 17:36:38.422758 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}})  2025-08-29 17:36:38.422765 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.422771 | orchestrator | 2025-08-29 17:36:38.422778 | orchestrator | TASK [proxysql-config : Copying over opensearch ProxySQL users config] ********* 2025-08-29 17:36:38.422785 | orchestrator | Friday 29 August 2025 17:35:19 +0000 (0:00:01.599) 0:04:56.054 ********* 2025-08-29 17:36:38.422791 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.422798 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.422804 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.422811 | orchestrator | 2025-08-29 17:36:38.422818 | orchestrator | TASK [proxysql-config : Copying over opensearch ProxySQL rules config] ********* 2025-08-29 17:36:38.422824 | orchestrator | Friday 29 August 2025 17:35:19 +0000 (0:00:00.410) 0:04:56.464 ********* 2025-08-29 17:36:38.422831 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.422838 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.422845 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.422851 | orchestrator | 2025-08-29 17:36:38.422858 | orchestrator | TASK [include_role : prometheus] *********************************************** 2025-08-29 17:36:38.422864 | orchestrator | Friday 29 August 2025 17:35:21 +0000 (0:00:01.300) 0:04:57.765 ********* 2025-08-29 17:36:38.422871 | orchestrator | included: prometheus for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.422883 | orchestrator | 2025-08-29 17:36:38.422890 | orchestrator | TASK [haproxy-config : Copying over prometheus haproxy config] ***************** 2025-08-29 17:36:38.422897 | orchestrator | Friday 29 August 2025 17:35:22 +0000 (0:00:01.651) 0:04:59.417 ********* 2025-08-29 17:36:38.422904 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) 2025-08-29 17:36:38.422912 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:36:38.422941 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.422953 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.422960 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.422968 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) 2025-08-29 17:36:38.422980 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) 2025-08-29 17:36:38.422987 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:36:38.423014 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423022 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:36:38.423034 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423041 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423048 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423059 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423066 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423077 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) 2025-08-29 17:36:38.423085 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}})  2025-08-29 17:36:38.423093 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423173 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423202 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423210 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) 2025-08-29 17:36:38.423227 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}})  2025-08-29 17:36:38.423238 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) 2025-08-29 17:36:38.423246 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423257 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}})  2025-08-29 17:36:38.423264 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423271 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423282 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423303 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423310 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423317 | orchestrator | 2025-08-29 17:36:38.423324 | orchestrator | TASK [haproxy-config : Add configuration for prometheus when using single external frontend] *** 2025-08-29 17:36:38.423331 | orchestrator | Friday 29 August 2025 17:35:26 +0000 (0:00:04.194) 0:05:03.611 ********* 2025-08-29 17:36:38.423345 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})  2025-08-29 17:36:38.423353 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:36:38.423360 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423366 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423431 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423445 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})  2025-08-29 17:36:38.423452 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})  2025-08-29 17:36:38.423465 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:36:38.423472 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}})  2025-08-29 17:36:38.423479 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423490 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423500 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423507 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423519 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423526 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423533 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.423540 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})  2025-08-29 17:36:38.423551 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}})  2025-08-29 17:36:38.423566 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423578 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423585 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423592 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.423599 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})  2025-08-29 17:36:38.423607 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:36:38.423614 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423625 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423635 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423647 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})  2025-08-29 17:36:38.423655 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/release/prometheus-openstack-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}})  2025-08-29 17:36:38.423661 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423668 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:36:38.423679 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:36:38.423686 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.423693 | orchestrator | 2025-08-29 17:36:38.423699 | orchestrator | TASK [haproxy-config : Configuring firewall for prometheus] ******************** 2025-08-29 17:36:38.423706 | orchestrator | Friday 29 August 2025 17:35:27 +0000 (0:00:00.897) 0:05:04.509 ********* 2025-08-29 17:36:38.423713 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}})  2025-08-29 17:36:38.423730 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}})  2025-08-29 17:36:38.423738 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}})  2025-08-29 17:36:38.423745 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}})  2025-08-29 17:36:38.423753 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.423759 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}})  2025-08-29 17:36:38.423766 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}})  2025-08-29 17:36:38.423774 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}})  2025-08-29 17:36:38.423781 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}})  2025-08-29 17:36:38.423787 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.423794 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}})  2025-08-29 17:36:38.423801 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}})  2025-08-29 17:36:38.423808 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}})  2025-08-29 17:36:38.423814 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}})  2025-08-29 17:36:38.423821 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.423828 | orchestrator | 2025-08-29 17:36:38.423834 | orchestrator | TASK [proxysql-config : Copying over prometheus ProxySQL users config] ********* 2025-08-29 17:36:38.423841 | orchestrator | Friday 29 August 2025 17:35:29 +0000 (0:00:01.281) 0:05:05.790 ********* 2025-08-29 17:36:38.423848 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.423854 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.423861 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.423867 | orchestrator | 2025-08-29 17:36:38.423878 | orchestrator | TASK [proxysql-config : Copying over prometheus ProxySQL rules config] ********* 2025-08-29 17:36:38.423885 | orchestrator | Friday 29 August 2025 17:35:29 +0000 (0:00:00.459) 0:05:06.250 ********* 2025-08-29 17:36:38.423895 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.423902 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.423908 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.423915 | orchestrator | 2025-08-29 17:36:38.423921 | orchestrator | TASK [include_role : rabbitmq] ************************************************* 2025-08-29 17:36:38.423928 | orchestrator | Friday 29 August 2025 17:35:30 +0000 (0:00:01.372) 0:05:07.622 ********* 2025-08-29 17:36:38.423934 | orchestrator | included: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.423941 | orchestrator | 2025-08-29 17:36:38.423948 | orchestrator | TASK [haproxy-config : Copying over rabbitmq haproxy config] ******************* 2025-08-29 17:36:38.423954 | orchestrator | Friday 29 August 2025 17:35:32 +0000 (0:00:01.419) 0:05:09.041 ********* 2025-08-29 17:36:38.423963 | orchestrator | changed: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:36:38.423971 | orchestrator | changed: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:36:38.423978 | orchestrator | changed: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) 2025-08-29 17:36:38.423990 | orchestrator | 2025-08-29 17:36:38.423996 | orchestrator | TASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] *** 2025-08-29 17:36:38.424002 | orchestrator | Friday 29 August 2025 17:35:34 +0000 (0:00:02.597) 0:05:11.639 ********* 2025-08-29 17:36:38.424013 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})  2025-08-29 17:36:38.424020 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424029 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})  2025-08-29 17:36:38.424036 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424042 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})  2025-08-29 17:36:38.424049 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424055 | orchestrator | 2025-08-29 17:36:38.424061 | orchestrator | TASK [haproxy-config : Configuring firewall for rabbitmq] ********************** 2025-08-29 17:36:38.424067 | orchestrator | Friday 29 August 2025 17:35:35 +0000 (0:00:00.463) 0:05:12.102 ********* 2025-08-29 17:36:38.424074 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}})  2025-08-29 17:36:38.424080 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424086 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}})  2025-08-29 17:36:38.424096 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424102 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}})  2025-08-29 17:36:38.424109 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424115 | orchestrator | 2025-08-29 17:36:38.424121 | orchestrator | TASK [proxysql-config : Copying over rabbitmq ProxySQL users config] *********** 2025-08-29 17:36:38.424127 | orchestrator | Friday 29 August 2025 17:35:36 +0000 (0:00:00.633) 0:05:12.736 ********* 2025-08-29 17:36:38.424133 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424139 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424145 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424151 | orchestrator | 2025-08-29 17:36:38.424158 | orchestrator | TASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] *********** 2025-08-29 17:36:38.424164 | orchestrator | Friday 29 August 2025 17:35:36 +0000 (0:00:00.818) 0:05:13.554 ********* 2025-08-29 17:36:38.424170 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424176 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424182 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424188 | orchestrator | 2025-08-29 17:36:38.424194 | orchestrator | TASK [include_role : skyline] ************************************************** 2025-08-29 17:36:38.424203 | orchestrator | Friday 29 August 2025 17:35:38 +0000 (0:00:01.477) 0:05:15.031 ********* 2025-08-29 17:36:38.424210 | orchestrator | included: skyline for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:36:38.424216 | orchestrator | 2025-08-29 17:36:38.424222 | orchestrator | TASK [haproxy-config : Copying over skyline haproxy config] ******************** 2025-08-29 17:36:38.424228 | orchestrator | Friday 29 August 2025 17:35:39 +0000 (0:00:01.590) 0:05:16.621 ********* 2025-08-29 17:36:38.424239 | orchestrator | changed: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.424250 | orchestrator | changed: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.424262 | orchestrator | changed: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.424278 | orchestrator | changed: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.424298 | orchestrator | changed: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.424310 | orchestrator | changed: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) 2025-08-29 17:36:38.424319 | orchestrator | 2025-08-29 17:36:38.424329 | orchestrator | TASK [haproxy-config : Add configuration for skyline when using single external frontend] *** 2025-08-29 17:36:38.424339 | orchestrator | Friday 29 August 2025 17:35:46 +0000 (0:00:06.779) 0:05:23.400 ********* 2025-08-29 17:36:38.424349 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.424367 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.424395 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424417 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.424430 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.424441 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424451 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-apiserver:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.424469 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/skyline-console:5.0.1.20250711', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})  2025-08-29 17:36:38.424476 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424483 | orchestrator | 2025-08-29 17:36:38.424489 | orchestrator | TASK [haproxy-config : Configuring firewall for skyline] *********************** 2025-08-29 17:36:38.424495 | orchestrator | Friday 29 August 2025 17:35:47 +0000 (0:00:00.710) 0:05:24.111 ********* 2025-08-29 17:36:38.424502 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424511 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424518 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424525 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424535 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424541 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424548 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424554 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424560 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424571 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424578 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424584 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424590 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424596 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}})  2025-08-29 17:36:38.424603 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424609 | orchestrator | 2025-08-29 17:36:38.424615 | orchestrator | TASK [proxysql-config : Copying over skyline ProxySQL users config] ************ 2025-08-29 17:36:38.424621 | orchestrator | Friday 29 August 2025 17:35:48 +0000 (0:00:01.009) 0:05:25.121 ********* 2025-08-29 17:36:38.424627 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.424633 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.424640 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.424646 | orchestrator | 2025-08-29 17:36:38.424652 | orchestrator | TASK [proxysql-config : Copying over skyline ProxySQL rules config] ************ 2025-08-29 17:36:38.424658 | orchestrator | Friday 29 August 2025 17:35:50 +0000 (0:00:01.983) 0:05:27.105 ********* 2025-08-29 17:36:38.424664 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.424670 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.424676 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.424682 | orchestrator | 2025-08-29 17:36:38.424689 | orchestrator | TASK [include_role : swift] **************************************************** 2025-08-29 17:36:38.424695 | orchestrator | Friday 29 August 2025 17:35:52 +0000 (0:00:02.207) 0:05:29.312 ********* 2025-08-29 17:36:38.424701 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424707 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424713 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424719 | orchestrator | 2025-08-29 17:36:38.424725 | orchestrator | TASK [include_role : tacker] *************************************************** 2025-08-29 17:36:38.424731 | orchestrator | Friday 29 August 2025 17:35:53 +0000 (0:00:00.367) 0:05:29.679 ********* 2025-08-29 17:36:38.424737 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424743 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424749 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424755 | orchestrator | 2025-08-29 17:36:38.424762 | orchestrator | TASK [include_role : trove] **************************************************** 2025-08-29 17:36:38.424768 | orchestrator | Friday 29 August 2025 17:35:53 +0000 (0:00:00.369) 0:05:30.049 ********* 2025-08-29 17:36:38.424774 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424780 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424786 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424792 | orchestrator | 2025-08-29 17:36:38.424798 | orchestrator | TASK [include_role : venus] **************************************************** 2025-08-29 17:36:38.424804 | orchestrator | Friday 29 August 2025 17:35:53 +0000 (0:00:00.327) 0:05:30.377 ********* 2025-08-29 17:36:38.424814 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424820 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424826 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424832 | orchestrator | 2025-08-29 17:36:38.424838 | orchestrator | TASK [include_role : watcher] ************************************************** 2025-08-29 17:36:38.424844 | orchestrator | Friday 29 August 2025 17:35:54 +0000 (0:00:00.670) 0:05:31.047 ********* 2025-08-29 17:36:38.424857 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424863 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424870 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424876 | orchestrator | 2025-08-29 17:36:38.424882 | orchestrator | TASK [include_role : zun] ****************************************************** 2025-08-29 17:36:38.424888 | orchestrator | Friday 29 August 2025 17:35:54 +0000 (0:00:00.323) 0:05:31.371 ********* 2025-08-29 17:36:38.424894 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.424900 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.424906 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.424912 | orchestrator | 2025-08-29 17:36:38.424922 | orchestrator | RUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] ******* 2025-08-29 17:36:38.424928 | orchestrator | Friday 29 August 2025 17:35:55 +0000 (0:00:00.533) 0:05:31.905 ********* 2025-08-29 17:36:38.424934 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.424941 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.424947 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.424953 | orchestrator | 2025-08-29 17:36:38.424959 | orchestrator | RUNNING HANDLER [loadbalancer : Group HA nodes by status] ********************** 2025-08-29 17:36:38.424965 | orchestrator | Friday 29 August 2025 17:35:56 +0000 (0:00:00.991) 0:05:32.897 ********* 2025-08-29 17:36:38.424971 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.424977 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.424984 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.424990 | orchestrator | 2025-08-29 17:36:38.424996 | orchestrator | RUNNING HANDLER [loadbalancer : Stop backup keepalived container] ************** 2025-08-29 17:36:38.425002 | orchestrator | Friday 29 August 2025 17:35:56 +0000 (0:00:00.340) 0:05:33.238 ********* 2025-08-29 17:36:38.425008 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.425014 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.425020 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.425026 | orchestrator | 2025-08-29 17:36:38.425033 | orchestrator | RUNNING HANDLER [loadbalancer : Stop backup haproxy container] ***************** 2025-08-29 17:36:38.425039 | orchestrator | Friday 29 August 2025 17:35:57 +0000 (0:00:00.932) 0:05:34.170 ********* 2025-08-29 17:36:38.425045 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.425051 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.425057 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.425063 | orchestrator | 2025-08-29 17:36:38.425069 | orchestrator | RUNNING HANDLER [loadbalancer : Stop backup proxysql container] **************** 2025-08-29 17:36:38.425075 | orchestrator | Friday 29 August 2025 17:35:58 +0000 (0:00:00.906) 0:05:35.077 ********* 2025-08-29 17:36:38.425081 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.425087 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.425093 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.425099 | orchestrator | 2025-08-29 17:36:38.425106 | orchestrator | RUNNING HANDLER [loadbalancer : Start backup haproxy container] **************** 2025-08-29 17:36:38.425112 | orchestrator | Friday 29 August 2025 17:35:59 +0000 (0:00:01.226) 0:05:36.303 ********* 2025-08-29 17:36:38.425118 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.425124 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.425130 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.425136 | orchestrator | 2025-08-29 17:36:38.425142 | orchestrator | RUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] ************** 2025-08-29 17:36:38.425148 | orchestrator | Friday 29 August 2025 17:36:08 +0000 (0:00:08.385) 0:05:44.689 ********* 2025-08-29 17:36:38.425155 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.425161 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.425167 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.425173 | orchestrator | 2025-08-29 17:36:38.425179 | orchestrator | RUNNING HANDLER [loadbalancer : Start backup proxysql container] *************** 2025-08-29 17:36:38.425185 | orchestrator | Friday 29 August 2025 17:36:08 +0000 (0:00:00.754) 0:05:45.443 ********* 2025-08-29 17:36:38.425191 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.425198 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.425209 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.425215 | orchestrator | 2025-08-29 17:36:38.425221 | orchestrator | RUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] ************* 2025-08-29 17:36:38.425227 | orchestrator | Friday 29 August 2025 17:36:21 +0000 (0:00:13.209) 0:05:58.653 ********* 2025-08-29 17:36:38.425233 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.425239 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.425245 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.425251 | orchestrator | 2025-08-29 17:36:38.425258 | orchestrator | RUNNING HANDLER [loadbalancer : Start backup keepalived container] ************* 2025-08-29 17:36:38.425264 | orchestrator | Friday 29 August 2025 17:36:22 +0000 (0:00:00.749) 0:05:59.402 ********* 2025-08-29 17:36:38.425270 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:36:38.425276 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:36:38.425282 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:36:38.425288 | orchestrator | 2025-08-29 17:36:38.425294 | orchestrator | RUNNING HANDLER [loadbalancer : Stop master haproxy container] ***************** 2025-08-29 17:36:38.425300 | orchestrator | Friday 29 August 2025 17:36:30 +0000 (0:00:08.211) 0:06:07.613 ********* 2025-08-29 17:36:38.425307 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.425313 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.425319 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.425325 | orchestrator | 2025-08-29 17:36:38.425331 | orchestrator | RUNNING HANDLER [loadbalancer : Stop master proxysql container] **************** 2025-08-29 17:36:38.425337 | orchestrator | Friday 29 August 2025 17:36:31 +0000 (0:00:00.331) 0:06:07.944 ********* 2025-08-29 17:36:38.425343 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.425349 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.425355 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.425361 | orchestrator | 2025-08-29 17:36:38.425367 | orchestrator | RUNNING HANDLER [loadbalancer : Stop master keepalived container] ************** 2025-08-29 17:36:38.425387 | orchestrator | Friday 29 August 2025 17:36:31 +0000 (0:00:00.347) 0:06:08.291 ********* 2025-08-29 17:36:38.425393 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.425400 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.425409 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.425416 | orchestrator | 2025-08-29 17:36:38.425422 | orchestrator | RUNNING HANDLER [loadbalancer : Start master haproxy container] **************** 2025-08-29 17:36:38.425428 | orchestrator | Friday 29 August 2025 17:36:31 +0000 (0:00:00.358) 0:06:08.650 ********* 2025-08-29 17:36:38.425434 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.425440 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.425446 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.425453 | orchestrator | 2025-08-29 17:36:38.425459 | orchestrator | RUNNING HANDLER [loadbalancer : Start master proxysql container] *************** 2025-08-29 17:36:38.425465 | orchestrator | Friday 29 August 2025 17:36:32 +0000 (0:00:00.736) 0:06:09.387 ********* 2025-08-29 17:36:38.425471 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.425477 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.425483 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.425489 | orchestrator | 2025-08-29 17:36:38.425496 | orchestrator | RUNNING HANDLER [loadbalancer : Start master keepalived container] ************* 2025-08-29 17:36:38.425505 | orchestrator | Friday 29 August 2025 17:36:33 +0000 (0:00:00.358) 0:06:09.745 ********* 2025-08-29 17:36:38.425511 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:36:38.425517 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:36:38.425524 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:36:38.425530 | orchestrator | 2025-08-29 17:36:38.425536 | orchestrator | RUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] ************* 2025-08-29 17:36:38.425542 | orchestrator | Friday 29 August 2025 17:36:33 +0000 (0:00:00.337) 0:06:10.082 ********* 2025-08-29 17:36:38.425548 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.425555 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.425561 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.425572 | orchestrator | 2025-08-29 17:36:38.425578 | orchestrator | RUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************ 2025-08-29 17:36:38.425584 | orchestrator | Friday 29 August 2025 17:36:34 +0000 (0:00:01.261) 0:06:11.344 ********* 2025-08-29 17:36:38.425590 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:36:38.425596 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:36:38.425603 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:36:38.425609 | orchestrator | 2025-08-29 17:36:38.425615 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:36:38.425621 | orchestrator | testbed-node-0 : ok=123  changed=76  unreachable=0 failed=0 skipped=97  rescued=0 ignored=0 2025-08-29 17:36:38.425628 | orchestrator | testbed-node-1 : ok=122  changed=76  unreachable=0 failed=0 skipped=97  rescued=0 ignored=0 2025-08-29 17:36:38.425634 | orchestrator | testbed-node-2 : ok=122  changed=76  unreachable=0 failed=0 skipped=97  rescued=0 ignored=0 2025-08-29 17:36:38.425640 | orchestrator | 2025-08-29 17:36:38.425647 | orchestrator | 2025-08-29 17:36:38.425653 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:36:38.425659 | orchestrator | Friday 29 August 2025 17:36:35 +0000 (0:00:01.286) 0:06:12.631 ********* 2025-08-29 17:36:38.425665 | orchestrator | =============================================================================== 2025-08-29 17:36:38.425671 | orchestrator | loadbalancer : Start backup proxysql container ------------------------- 13.21s 2025-08-29 17:36:38.425678 | orchestrator | loadbalancer : Start backup haproxy container --------------------------- 8.39s 2025-08-29 17:36:38.425684 | orchestrator | loadbalancer : Start backup keepalived container ------------------------ 8.21s 2025-08-29 17:36:38.425690 | orchestrator | haproxy-config : Copying over skyline haproxy config -------------------- 6.78s 2025-08-29 17:36:38.425696 | orchestrator | haproxy-config : Copying over designate haproxy config ------------------ 5.31s 2025-08-29 17:36:38.425702 | orchestrator | haproxy-config : Copying over opensearch haproxy config ----------------- 5.03s 2025-08-29 17:36:38.425708 | orchestrator | loadbalancer : Copying over proxysql config ----------------------------- 4.68s 2025-08-29 17:36:38.425715 | orchestrator | haproxy-config : Copying over neutron haproxy config -------------------- 4.54s 2025-08-29 17:36:38.425721 | orchestrator | haproxy-config : Copying over magnum haproxy config --------------------- 4.39s 2025-08-29 17:36:38.425727 | orchestrator | haproxy-config : Copying over nova haproxy config ----------------------- 4.24s 2025-08-29 17:36:38.425733 | orchestrator | haproxy-config : Copying over prometheus haproxy config ----------------- 4.19s 2025-08-29 17:36:38.425739 | orchestrator | haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.08s 2025-08-29 17:36:38.425745 | orchestrator | haproxy-config : Copying over manila haproxy config --------------------- 3.94s 2025-08-29 17:36:38.425752 | orchestrator | haproxy-config : Copying over glance haproxy config --------------------- 3.93s 2025-08-29 17:36:38.425758 | orchestrator | service-cert-copy : loadbalancer | Copying over extra CA certificates --- 3.88s 2025-08-29 17:36:38.425764 | orchestrator | haproxy-config : Copying over octavia haproxy config -------------------- 3.70s 2025-08-29 17:36:38.425771 | orchestrator | haproxy-config : Copying over keystone haproxy config ------------------- 3.64s 2025-08-29 17:36:38.425777 | orchestrator | haproxy-config : Copying over aodh haproxy config ----------------------- 3.53s 2025-08-29 17:36:38.425783 | orchestrator | haproxy-config : Copying over barbican haproxy config ------------------- 3.50s 2025-08-29 17:36:38.425789 | orchestrator | haproxy-config : Copying over placement haproxy config ------------------ 3.43s 2025-08-29 17:36:38.425795 | orchestrator | 2025-08-29 17:36:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:41.451853 | orchestrator | 2025-08-29 17:36:41 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:41.451938 | orchestrator | 2025-08-29 17:36:41 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:41.452715 | orchestrator | 2025-08-29 17:36:41 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:41.452909 | orchestrator | 2025-08-29 17:36:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:44.490516 | orchestrator | 2025-08-29 17:36:44 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:44.490619 | orchestrator | 2025-08-29 17:36:44 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:44.491493 | orchestrator | 2025-08-29 17:36:44 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:44.491540 | orchestrator | 2025-08-29 17:36:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:47.604508 | orchestrator | 2025-08-29 17:36:47 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:47.606511 | orchestrator | 2025-08-29 17:36:47 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:47.610123 | orchestrator | 2025-08-29 17:36:47 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:47.610614 | orchestrator | 2025-08-29 17:36:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:50.651519 | orchestrator | 2025-08-29 17:36:50 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:50.651720 | orchestrator | 2025-08-29 17:36:50 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:50.652247 | orchestrator | 2025-08-29 17:36:50 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:50.652345 | orchestrator | 2025-08-29 17:36:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:53.690425 | orchestrator | 2025-08-29 17:36:53 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:53.691216 | orchestrator | 2025-08-29 17:36:53 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:53.692042 | orchestrator | 2025-08-29 17:36:53 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:53.692252 | orchestrator | 2025-08-29 17:36:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:56.733698 | orchestrator | 2025-08-29 17:36:56 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:56.734559 | orchestrator | 2025-08-29 17:36:56 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:56.735912 | orchestrator | 2025-08-29 17:36:56 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:56.736013 | orchestrator | 2025-08-29 17:36:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:36:59.773743 | orchestrator | 2025-08-29 17:36:59 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:36:59.776462 | orchestrator | 2025-08-29 17:36:59 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:36:59.778881 | orchestrator | 2025-08-29 17:36:59 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:36:59.779356 | orchestrator | 2025-08-29 17:36:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:02.812920 | orchestrator | 2025-08-29 17:37:02 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:02.813026 | orchestrator | 2025-08-29 17:37:02 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:02.813575 | orchestrator | 2025-08-29 17:37:02 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:02.813644 | orchestrator | 2025-08-29 17:37:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:05.850936 | orchestrator | 2025-08-29 17:37:05 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:05.852537 | orchestrator | 2025-08-29 17:37:05 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:05.853792 | orchestrator | 2025-08-29 17:37:05 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:05.853818 | orchestrator | 2025-08-29 17:37:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:08.900082 | orchestrator | 2025-08-29 17:37:08 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:08.900185 | orchestrator | 2025-08-29 17:37:08 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:08.900200 | orchestrator | 2025-08-29 17:37:08 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:08.900212 | orchestrator | 2025-08-29 17:37:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:11.928004 | orchestrator | 2025-08-29 17:37:11 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:11.928060 | orchestrator | 2025-08-29 17:37:11 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:11.928069 | orchestrator | 2025-08-29 17:37:11 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:11.928088 | orchestrator | 2025-08-29 17:37:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:14.959291 | orchestrator | 2025-08-29 17:37:14 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:14.960401 | orchestrator | 2025-08-29 17:37:14 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:14.961121 | orchestrator | 2025-08-29 17:37:14 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:14.961249 | orchestrator | 2025-08-29 17:37:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:18.021882 | orchestrator | 2025-08-29 17:37:18 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:18.023878 | orchestrator | 2025-08-29 17:37:18 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:18.025509 | orchestrator | 2025-08-29 17:37:18 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:18.025595 | orchestrator | 2025-08-29 17:37:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:21.076168 | orchestrator | 2025-08-29 17:37:21 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:21.076920 | orchestrator | 2025-08-29 17:37:21 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:21.078115 | orchestrator | 2025-08-29 17:37:21 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:21.078146 | orchestrator | 2025-08-29 17:37:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:24.124099 | orchestrator | 2025-08-29 17:37:24 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:24.125688 | orchestrator | 2025-08-29 17:37:24 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:24.127053 | orchestrator | 2025-08-29 17:37:24 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:24.127325 | orchestrator | 2025-08-29 17:37:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:27.175870 | orchestrator | 2025-08-29 17:37:27 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:27.177058 | orchestrator | 2025-08-29 17:37:27 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:27.179100 | orchestrator | 2025-08-29 17:37:27 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:27.179329 | orchestrator | 2025-08-29 17:37:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:30.221370 | orchestrator | 2025-08-29 17:37:30 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:30.222781 | orchestrator | 2025-08-29 17:37:30 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:30.222891 | orchestrator | 2025-08-29 17:37:30 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:30.222915 | orchestrator | 2025-08-29 17:37:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:33.266958 | orchestrator | 2025-08-29 17:37:33 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:33.272137 | orchestrator | 2025-08-29 17:37:33 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:33.273157 | orchestrator | 2025-08-29 17:37:33 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:33.273246 | orchestrator | 2025-08-29 17:37:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:36.325490 | orchestrator | 2025-08-29 17:37:36 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:36.326633 | orchestrator | 2025-08-29 17:37:36 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:36.328135 | orchestrator | 2025-08-29 17:37:36 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:36.328171 | orchestrator | 2025-08-29 17:37:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:39.366719 | orchestrator | 2025-08-29 17:37:39 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:39.368745 | orchestrator | 2025-08-29 17:37:39 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:39.370944 | orchestrator | 2025-08-29 17:37:39 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:39.371012 | orchestrator | 2025-08-29 17:37:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:42.407955 | orchestrator | 2025-08-29 17:37:42 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:42.409723 | orchestrator | 2025-08-29 17:37:42 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:42.412109 | orchestrator | 2025-08-29 17:37:42 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:42.412135 | orchestrator | 2025-08-29 17:37:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:45.464859 | orchestrator | 2025-08-29 17:37:45 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:45.468051 | orchestrator | 2025-08-29 17:37:45 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:45.469224 | orchestrator | 2025-08-29 17:37:45 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:45.469260 | orchestrator | 2025-08-29 17:37:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:48.512789 | orchestrator | 2025-08-29 17:37:48 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:48.514987 | orchestrator | 2025-08-29 17:37:48 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:48.515028 | orchestrator | 2025-08-29 17:37:48 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:48.515040 | orchestrator | 2025-08-29 17:37:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:51.560234 | orchestrator | 2025-08-29 17:37:51 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:51.561805 | orchestrator | 2025-08-29 17:37:51 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:51.564651 | orchestrator | 2025-08-29 17:37:51 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:51.564759 | orchestrator | 2025-08-29 17:37:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:54.609921 | orchestrator | 2025-08-29 17:37:54 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:54.611700 | orchestrator | 2025-08-29 17:37:54 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:54.615810 | orchestrator | 2025-08-29 17:37:54 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:54.615907 | orchestrator | 2025-08-29 17:37:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:37:57.657623 | orchestrator | 2025-08-29 17:37:57 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:37:57.659185 | orchestrator | 2025-08-29 17:37:57 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:37:57.660149 | orchestrator | 2025-08-29 17:37:57 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:37:57.660197 | orchestrator | 2025-08-29 17:37:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:00.719276 | orchestrator | 2025-08-29 17:38:00 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:00.721461 | orchestrator | 2025-08-29 17:38:00 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:00.724912 | orchestrator | 2025-08-29 17:38:00 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:00.725442 | orchestrator | 2025-08-29 17:38:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:03.778695 | orchestrator | 2025-08-29 17:38:03 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:03.780561 | orchestrator | 2025-08-29 17:38:03 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:03.782298 | orchestrator | 2025-08-29 17:38:03 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:03.782341 | orchestrator | 2025-08-29 17:38:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:06.838342 | orchestrator | 2025-08-29 17:38:06 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:06.843827 | orchestrator | 2025-08-29 17:38:06 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:06.847242 | orchestrator | 2025-08-29 17:38:06 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:06.847571 | orchestrator | 2025-08-29 17:38:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:09.890215 | orchestrator | 2025-08-29 17:38:09 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:09.890340 | orchestrator | 2025-08-29 17:38:09 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:09.890364 | orchestrator | 2025-08-29 17:38:09 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:09.890374 | orchestrator | 2025-08-29 17:38:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:12.939188 | orchestrator | 2025-08-29 17:38:12 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:12.942642 | orchestrator | 2025-08-29 17:38:12 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:12.945731 | orchestrator | 2025-08-29 17:38:12 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:12.946659 | orchestrator | 2025-08-29 17:38:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:15.999899 | orchestrator | 2025-08-29 17:38:15 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:16.002616 | orchestrator | 2025-08-29 17:38:16 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:16.004579 | orchestrator | 2025-08-29 17:38:16 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:16.004640 | orchestrator | 2025-08-29 17:38:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:19.055081 | orchestrator | 2025-08-29 17:38:19 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:19.055683 | orchestrator | 2025-08-29 17:38:19 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:19.056943 | orchestrator | 2025-08-29 17:38:19 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:19.056976 | orchestrator | 2025-08-29 17:38:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:22.105542 | orchestrator | 2025-08-29 17:38:22 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:22.107643 | orchestrator | 2025-08-29 17:38:22 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:22.110238 | orchestrator | 2025-08-29 17:38:22 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:22.110278 | orchestrator | 2025-08-29 17:38:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:25.158337 | orchestrator | 2025-08-29 17:38:25 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:25.159078 | orchestrator | 2025-08-29 17:38:25 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:25.160980 | orchestrator | 2025-08-29 17:38:25 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:25.161025 | orchestrator | 2025-08-29 17:38:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:28.210296 | orchestrator | 2025-08-29 17:38:28 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:28.212482 | orchestrator | 2025-08-29 17:38:28 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:28.215078 | orchestrator | 2025-08-29 17:38:28 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:28.215295 | orchestrator | 2025-08-29 17:38:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:31.265661 | orchestrator | 2025-08-29 17:38:31 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:31.266943 | orchestrator | 2025-08-29 17:38:31 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:31.268689 | orchestrator | 2025-08-29 17:38:31 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:31.268724 | orchestrator | 2025-08-29 17:38:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:34.315820 | orchestrator | 2025-08-29 17:38:34 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:34.317913 | orchestrator | 2025-08-29 17:38:34 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:34.319622 | orchestrator | 2025-08-29 17:38:34 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:34.319820 | orchestrator | 2025-08-29 17:38:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:37.366898 | orchestrator | 2025-08-29 17:38:37 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:37.370264 | orchestrator | 2025-08-29 17:38:37 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:37.373059 | orchestrator | 2025-08-29 17:38:37 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:37.373370 | orchestrator | 2025-08-29 17:38:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:40.414988 | orchestrator | 2025-08-29 17:38:40 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:40.416557 | orchestrator | 2025-08-29 17:38:40 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:40.418280 | orchestrator | 2025-08-29 17:38:40 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:40.418572 | orchestrator | 2025-08-29 17:38:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:43.464680 | orchestrator | 2025-08-29 17:38:43 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:43.466195 | orchestrator | 2025-08-29 17:38:43 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:43.468296 | orchestrator | 2025-08-29 17:38:43 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:43.468328 | orchestrator | 2025-08-29 17:38:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:46.525747 | orchestrator | 2025-08-29 17:38:46 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:46.526230 | orchestrator | 2025-08-29 17:38:46 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:46.528030 | orchestrator | 2025-08-29 17:38:46 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:46.528284 | orchestrator | 2025-08-29 17:38:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:49.582713 | orchestrator | 2025-08-29 17:38:49 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:49.583601 | orchestrator | 2025-08-29 17:38:49 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:49.584364 | orchestrator | 2025-08-29 17:38:49 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:49.584526 | orchestrator | 2025-08-29 17:38:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:52.629014 | orchestrator | 2025-08-29 17:38:52 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:52.633594 | orchestrator | 2025-08-29 17:38:52 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:52.635997 | orchestrator | 2025-08-29 17:38:52 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:52.636125 | orchestrator | 2025-08-29 17:38:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:55.690145 | orchestrator | 2025-08-29 17:38:55 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:55.692321 | orchestrator | 2025-08-29 17:38:55 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:55.694217 | orchestrator | 2025-08-29 17:38:55 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:55.694295 | orchestrator | 2025-08-29 17:38:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:38:58.742879 | orchestrator | 2025-08-29 17:38:58 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:38:58.743098 | orchestrator | 2025-08-29 17:38:58 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:38:58.744512 | orchestrator | 2025-08-29 17:38:58 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state STARTED 2025-08-29 17:38:58.744551 | orchestrator | 2025-08-29 17:38:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:01.796147 | orchestrator | 2025-08-29 17:39:01 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:01.798193 | orchestrator | 2025-08-29 17:39:01 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:01.801537 | orchestrator | 2025-08-29 17:39:01 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:01.808705 | orchestrator | 2025-08-29 17:39:01 | INFO  | Task a36066c6-d231-4483-a7a5-38ea5e3561ae is in state SUCCESS 2025-08-29 17:39:01.810741 | orchestrator | 2025-08-29 17:39:01.810778 | orchestrator | 2025-08-29 17:39:01.810791 | orchestrator | PLAY [Prepare deployment of Ceph services] ************************************* 2025-08-29 17:39:01.810803 | orchestrator | 2025-08-29 17:39:01.810814 | orchestrator | TASK [ceph-facts : Include facts.yml] ****************************************** 2025-08-29 17:39:01.810826 | orchestrator | Friday 29 August 2025 17:27:41 +0000 (0:00:00.678) 0:00:00.678 ********* 2025-08-29 17:39:01.810838 | orchestrator | included: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.810851 | orchestrator | 2025-08-29 17:39:01.810862 | orchestrator | TASK [ceph-facts : Check if it is atomic host] ********************************* 2025-08-29 17:39:01.810872 | orchestrator | Friday 29 August 2025 17:27:42 +0000 (0:00:00.998) 0:00:01.676 ********* 2025-08-29 17:39:01.810883 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.810895 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.810906 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.810970 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.810984 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.811037 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.811086 | orchestrator | 2025-08-29 17:39:01.811097 | orchestrator | TASK [ceph-facts : Set_fact is_atomic] ***************************************** 2025-08-29 17:39:01.811108 | orchestrator | Friday 29 August 2025 17:27:44 +0000 (0:00:01.651) 0:00:03.327 ********* 2025-08-29 17:39:01.811119 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.811130 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.811141 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.811152 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.811219 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.811231 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.811242 | orchestrator | 2025-08-29 17:39:01.811253 | orchestrator | TASK [ceph-facts : Check if podman binary is present] ************************** 2025-08-29 17:39:01.811290 | orchestrator | Friday 29 August 2025 17:27:44 +0000 (0:00:00.666) 0:00:03.994 ********* 2025-08-29 17:39:01.811419 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.811435 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.811447 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.811459 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.811471 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.811484 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.811497 | orchestrator | 2025-08-29 17:39:01.811535 | orchestrator | TASK [ceph-facts : Set_fact container_binary] ********************************** 2025-08-29 17:39:01.811549 | orchestrator | Friday 29 August 2025 17:27:45 +0000 (0:00:00.927) 0:00:04.922 ********* 2025-08-29 17:39:01.811562 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.811575 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.811588 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.811625 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.811639 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.811652 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.811664 | orchestrator | 2025-08-29 17:39:01.811676 | orchestrator | TASK [ceph-facts : Set_fact ceph_cmd] ****************************************** 2025-08-29 17:39:01.811753 | orchestrator | Friday 29 August 2025 17:27:46 +0000 (0:00:00.735) 0:00:05.657 ********* 2025-08-29 17:39:01.811764 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.811775 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.811786 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.811797 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.811808 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.811819 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.811831 | orchestrator | 2025-08-29 17:39:01.811842 | orchestrator | TASK [ceph-facts : Set_fact discovered_interpreter_python] ********************* 2025-08-29 17:39:01.811928 | orchestrator | Friday 29 August 2025 17:27:47 +0000 (0:00:00.606) 0:00:06.264 ********* 2025-08-29 17:39:01.811942 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.811954 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.811966 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.811979 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.811991 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.812003 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.812015 | orchestrator | 2025-08-29 17:39:01.812028 | orchestrator | TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] *** 2025-08-29 17:39:01.812040 | orchestrator | Friday 29 August 2025 17:27:48 +0000 (0:00:00.880) 0:00:07.145 ********* 2025-08-29 17:39:01.812053 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.812066 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.812078 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.812090 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.812102 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.812114 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.812126 | orchestrator | 2025-08-29 17:39:01.812139 | orchestrator | TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ****************** 2025-08-29 17:39:01.812151 | orchestrator | Friday 29 August 2025 17:27:49 +0000 (0:00:01.571) 0:00:08.716 ********* 2025-08-29 17:39:01.812163 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.812175 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.812187 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.812199 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.812211 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.812224 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.812236 | orchestrator | 2025-08-29 17:39:01.812248 | orchestrator | TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************ 2025-08-29 17:39:01.812260 | orchestrator | Friday 29 August 2025 17:27:50 +0000 (0:00:01.081) 0:00:09.797 ********* 2025-08-29 17:39:01.812273 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:39:01.812285 | orchestrator | ok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:39:01.812351 | orchestrator | ok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:39:01.812437 | orchestrator | 2025-08-29 17:39:01.812450 | orchestrator | TASK [ceph-facts : Set_fact container_exec_cmd] ******************************** 2025-08-29 17:39:01.812527 | orchestrator | Friday 29 August 2025 17:27:51 +0000 (0:00:00.616) 0:00:10.414 ********* 2025-08-29 17:39:01.812539 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.812551 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.812561 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.812572 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.812583 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.812604 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.812615 | orchestrator | 2025-08-29 17:39:01.812669 | orchestrator | TASK [ceph-facts : Find a running mon container] ******************************* 2025-08-29 17:39:01.812683 | orchestrator | Friday 29 August 2025 17:27:52 +0000 (0:00:01.394) 0:00:11.808 ********* 2025-08-29 17:39:01.812694 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:39:01.812705 | orchestrator | ok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:39:01.812716 | orchestrator | ok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:39:01.812727 | orchestrator | 2025-08-29 17:39:01.812775 | orchestrator | TASK [ceph-facts : Check for a ceph mon socket] ******************************** 2025-08-29 17:39:01.812786 | orchestrator | Friday 29 August 2025 17:27:55 +0000 (0:00:03.158) 0:00:14.966 ********* 2025-08-29 17:39:01.812798 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-0)  2025-08-29 17:39:01.812808 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-1)  2025-08-29 17:39:01.812899 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-2)  2025-08-29 17:39:01.812911 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.812922 | orchestrator | 2025-08-29 17:39:01.812933 | orchestrator | TASK [ceph-facts : Check if the ceph mon socket is in-use] ********************* 2025-08-29 17:39:01.812944 | orchestrator | Friday 29 August 2025 17:27:56 +0000 (0:00:00.745) 0:00:15.712 ********* 2025-08-29 17:39:01.812958 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.812972 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.812983 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.812995 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.813006 | orchestrator | 2025-08-29 17:39:01.813017 | orchestrator | TASK [ceph-facts : Set_fact running_mon - non_container] *********************** 2025-08-29 17:39:01.813028 | orchestrator | Friday 29 August 2025 17:27:57 +0000 (0:00:00.872) 0:00:16.584 ********* 2025-08-29 17:39:01.813042 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.813056 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.813076 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.813088 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.813099 | orchestrator | 2025-08-29 17:39:01.813110 | orchestrator | TASK [ceph-facts : Set_fact running_mon - container] *************************** 2025-08-29 17:39:01.813179 | orchestrator | Friday 29 August 2025 17:27:57 +0000 (0:00:00.500) 0:00:17.085 ********* 2025-08-29 17:39:01.813200 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2025-08-29 17:27:53.382728', 'end': '2025-08-29 17:27:53.694282', 'delta': '0:00:00.311554', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.813224 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2025-08-29 17:27:54.388864', 'end': '2025-08-29 17:27:54.696812', 'delta': '0:00:00.307948', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.813237 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2025-08-29 17:27:55.158969', 'end': '2025-08-29 17:27:55.443432', 'delta': '0:00:00.284463', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.813249 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.813259 | orchestrator | 2025-08-29 17:39:01.813270 | orchestrator | TASK [ceph-facts : Set_fact _container_exec_cmd] ******************************* 2025-08-29 17:39:01.813281 | orchestrator | Friday 29 August 2025 17:27:58 +0000 (0:00:00.325) 0:00:17.410 ********* 2025-08-29 17:39:01.813292 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.813303 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.813314 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.813324 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.813335 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.813346 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.813357 | orchestrator | 2025-08-29 17:39:01.813368 | orchestrator | TASK [ceph-facts : Get current fsid if cluster is already running] ************* 2025-08-29 17:39:01.813379 | orchestrator | Friday 29 August 2025 17:28:01 +0000 (0:00:02.742) 0:00:20.152 ********* 2025-08-29 17:39:01.813448 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.813470 | orchestrator | 2025-08-29 17:39:01.813559 | orchestrator | TASK [ceph-facts : Set_fact current_fsid rc 1] ********************************* 2025-08-29 17:39:01.813572 | orchestrator | Friday 29 August 2025 17:28:01 +0000 (0:00:00.774) 0:00:20.927 ********* 2025-08-29 17:39:01.813583 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.813594 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.813605 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.813616 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.813626 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.813637 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.813648 | orchestrator | 2025-08-29 17:39:01.813659 | orchestrator | TASK [ceph-facts : Get current fsid] ******************************************* 2025-08-29 17:39:01.813670 | orchestrator | Friday 29 August 2025 17:28:02 +0000 (0:00:01.182) 0:00:22.110 ********* 2025-08-29 17:39:01.813766 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.813777 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.813787 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.813798 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.813809 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.813820 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.813831 | orchestrator | 2025-08-29 17:39:01.813842 | orchestrator | TASK [ceph-facts : Set_fact fsid] ********************************************** 2025-08-29 17:39:01.813852 | orchestrator | Friday 29 August 2025 17:28:04 +0000 (0:00:01.460) 0:00:23.570 ********* 2025-08-29 17:39:01.813863 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.813874 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.813885 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.813896 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.813907 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.813944 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.813956 | orchestrator | 2025-08-29 17:39:01.813968 | orchestrator | TASK [ceph-facts : Set_fact fsid from current_fsid] **************************** 2025-08-29 17:39:01.814005 | orchestrator | Friday 29 August 2025 17:28:05 +0000 (0:00:00.578) 0:00:24.148 ********* 2025-08-29 17:39:01.814061 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.814105 | orchestrator | 2025-08-29 17:39:01.814118 | orchestrator | TASK [ceph-facts : Generate cluster fsid] ************************************** 2025-08-29 17:39:01.814129 | orchestrator | Friday 29 August 2025 17:28:05 +0000 (0:00:00.103) 0:00:24.251 ********* 2025-08-29 17:39:01.814140 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.814151 | orchestrator | 2025-08-29 17:39:01.814162 | orchestrator | TASK [ceph-facts : Set_fact fsid] ********************************************** 2025-08-29 17:39:01.814173 | orchestrator | Friday 29 August 2025 17:28:05 +0000 (0:00:00.182) 0:00:24.434 ********* 2025-08-29 17:39:01.814218 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.814229 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.814257 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.814268 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.814279 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.814290 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.814301 | orchestrator | 2025-08-29 17:39:01.814331 | orchestrator | TASK [ceph-facts : Resolve device link(s)] ************************************* 2025-08-29 17:39:01.814351 | orchestrator | Friday 29 August 2025 17:28:06 +0000 (0:00:01.013) 0:00:25.447 ********* 2025-08-29 17:39:01.814363 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.814374 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.814401 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.814413 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.814424 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.814437 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.814455 | orchestrator | 2025-08-29 17:39:01.814472 | orchestrator | TASK [ceph-facts : Set_fact build devices from resolved symlinks] ************** 2025-08-29 17:39:01.814489 | orchestrator | Friday 29 August 2025 17:28:07 +0000 (0:00:01.091) 0:00:26.539 ********* 2025-08-29 17:39:01.814530 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.814550 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.814567 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.814586 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.814602 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.814620 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.814637 | orchestrator | 2025-08-29 17:39:01.814655 | orchestrator | TASK [ceph-facts : Resolve dedicated_device link(s)] *************************** 2025-08-29 17:39:01.814674 | orchestrator | Friday 29 August 2025 17:28:08 +0000 (0:00:00.840) 0:00:27.380 ********* 2025-08-29 17:39:01.814691 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.814709 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.814725 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.814741 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.814757 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.814774 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.814791 | orchestrator | 2025-08-29 17:39:01.814809 | orchestrator | TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] **** 2025-08-29 17:39:01.814828 | orchestrator | Friday 29 August 2025 17:28:09 +0000 (0:00:00.929) 0:00:28.310 ********* 2025-08-29 17:39:01.814846 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.814862 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.814880 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.814897 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.814914 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.814930 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.814947 | orchestrator | 2025-08-29 17:39:01.814966 | orchestrator | TASK [ceph-facts : Resolve bluestore_wal_device link(s)] *********************** 2025-08-29 17:39:01.814986 | orchestrator | Friday 29 August 2025 17:28:09 +0000 (0:00:00.545) 0:00:28.855 ********* 2025-08-29 17:39:01.815006 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.815026 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.815044 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.815062 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.815080 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.815099 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.815118 | orchestrator | 2025-08-29 17:39:01.815138 | orchestrator | TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] *** 2025-08-29 17:39:01.815156 | orchestrator | Friday 29 August 2025 17:28:10 +0000 (0:00:00.868) 0:00:29.723 ********* 2025-08-29 17:39:01.815176 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.815194 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.815213 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.815231 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.815248 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.815259 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.815273 | orchestrator | 2025-08-29 17:39:01.815292 | orchestrator | TASK [ceph-facts : Collect existed devices] ************************************ 2025-08-29 17:39:01.815310 | orchestrator | Friday 29 August 2025 17:28:11 +0000 (0:00:01.120) 0:00:30.844 ********* 2025-08-29 17:39:01.815329 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815360 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815477 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815517 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815542 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815554 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815566 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815577 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815593 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part1', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part14', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part15', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part16', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.815635 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.815648 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.815660 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815671 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815682 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815693 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815704 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815715 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815734 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815745 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815784 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part1', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part14', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part15', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part16', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.815799 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-20-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.815811 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815829 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815841 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815852 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815875 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815885 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815895 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815905 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815916 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part1', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part14', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part15', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part16', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.815942 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.815953 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.815963 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b', 'dm-uuid-LVM-8SW2ubuTvGOnhFOTau03JFgatBW9cwlrT4nSVIJUb9pyMNhhp0p0QmNBRj3JLwK3'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815975 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a', 'dm-uuid-LVM-5q8dGM2db5NrMc0chc8fHWOTnADjoH0OBy9rcTAnAINXxTj11VL1o8QghnCwxTrh'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815986 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.815996 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.816006 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816024 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816034 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816044 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816065 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2', 'dm-uuid-LVM-GUlkxvsBUEoBNxRHTurm2I6vdb71iOQT09A53JwNvatREyMn391iO8eZzncdLhOV'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816076 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816087 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2', 'dm-uuid-LVM-gVhXWy9gnL13qoE5gEc3ufITkM48FBGMjmEmplIXepwpRgq3AOhOj1huYwO1lOSE'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816097 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816106 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816123 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816133 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816143 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816153 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816167 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816184 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816195 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816213 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816223 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816239 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Ln6jHt-qsqU-X1le-h5ao-5dTL-pKYQ-cOGmh1', 'scsi-0QEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322', 'scsi-SQEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816250 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816267 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uPkwe4-XhPc-lzhV-DN2T-ebZn-SA0c-JkEUrj', 'scsi-0QEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f', 'scsi-SQEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816318 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-0A5zqG-61nB-VreW-lmbz-tCSc-hwpf-PGsoeL', 'scsi-0QEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3', 'scsi-SQEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816341 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-IdAuJW-KdwX-YTF8-XPAR-1cwa-KggL-hFVogq', 'scsi-0QEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6', 'scsi-SQEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816352 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5', 'scsi-SQEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816363 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474', 'scsi-SQEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816379 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816418 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-17-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816435 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.816451 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.816468 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe', 'dm-uuid-LVM-tOwLLnFJLaczcEpQ6NBpurAfznGF1LI9EZRJAMwx4l3TllcGTbf8jWoatcKCkzDd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816492 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288', 'dm-uuid-LVM-y0l4STA9mSjwec3ylN7nLUCBF2I0AWgcWG3yGA1QpITOXDaU9fQLHWqzTqgQaUq4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816519 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816539 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816550 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816569 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816579 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816589 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816599 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816609 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:39:01.816633 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816652 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-CYMrCT-OsQ3-cdVr-EcBN-TKOH-xE0q-pkdaCm', 'scsi-0QEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6', 'scsi-SQEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816662 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XNIeJ-OvKH-SWIj-T9QN-B9GX-Jw9F-bWcj1x', 'scsi-0QEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6', 'scsi-SQEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816673 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290', 'scsi-SQEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816687 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-18-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:39:01.816703 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.816714 | orchestrator | 2025-08-29 17:39:01.816724 | orchestrator | TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] *** 2025-08-29 17:39:01.816734 | orchestrator | Friday 29 August 2025 17:28:13 +0000 (0:00:01.884) 0:00:32.728 ********* 2025-08-29 17:39:01.816745 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816762 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816772 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816782 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816792 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816802 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816824 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816835 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816852 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part1', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part14', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part15', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part16', 'scsi-SQEMU_QEMU_HARDDISK_ddf3c727-22d6-4195-a504-158fe82cbe5d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816872 | orchestrator | skipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-13-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816884 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816899 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816909 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816920 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816930 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816940 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816960 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816977 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816988 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part1', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part14', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part15', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part16', 'scsi-SQEMU_QEMU_HARDDISK_c024de2c-1d0f-4cd2-a007-d9b7894b983c-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.816999 | orchestrator | skipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-20-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817009 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.817033 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817049 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.817060 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817070 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817080 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817090 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817100 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817119 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817135 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817146 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b', 'dm-uuid-LVM-8SW2ubuTvGOnhFOTau03JFgatBW9cwlrT4nSVIJUb9pyMNhhp0p0QmNBRj3JLwK3'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817157 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part1', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part14', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part15', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part16', 'scsi-SQEMU_QEMU_HARDDISK_681b03f9-5a09-4bb1-a1f3-7c10614aa46d-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817178 | orchestrator | skipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-15-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817195 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a', 'dm-uuid-LVM-5q8dGM2db5NrMc0chc8fHWOTnADjoH0OBy9rcTAnAINXxTj11VL1o8QghnCwxTrh'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817206 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817216 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817227 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817237 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.817247 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817512 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817539 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817551 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817569 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817605 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817647 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Ln6jHt-qsqU-X1le-h5ao-5dTL-pKYQ-cOGmh1', 'scsi-0QEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322', 'scsi-SQEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817669 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uPkwe4-XhPc-lzhV-DN2T-ebZn-SA0c-JkEUrj', 'scsi-0QEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f', 'scsi-SQEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817688 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5', 'scsi-SQEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817704 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2', 'dm-uuid-LVM-GUlkxvsBUEoBNxRHTurm2I6vdb71iOQT09A53JwNvatREyMn391iO8eZzncdLhOV'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817740 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-17-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817752 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2', 'dm-uuid-LVM-gVhXWy9gnL13qoE5gEc3ufITkM48FBGMjmEmplIXepwpRgq3AOhOj1huYwO1lOSE'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817762 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817773 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.817783 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817793 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817804 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817829 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817840 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817850 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817860 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817881 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817905 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-0A5zqG-61nB-VreW-lmbz-tCSc-hwpf-PGsoeL', 'scsi-0QEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3', 'scsi-SQEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817915 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-IdAuJW-KdwX-YTF8-XPAR-1cwa-KggL-hFVogq', 'scsi-0QEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6', 'scsi-SQEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817926 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474', 'scsi-SQEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817942 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe', 'dm-uuid-LVM-tOwLLnFJLaczcEpQ6NBpurAfznGF1LI9EZRJAMwx4l3TllcGTbf8jWoatcKCkzDd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817962 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817973 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.817983 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288', 'dm-uuid-LVM-y0l4STA9mSjwec3ylN7nLUCBF2I0AWgcWG3yGA1QpITOXDaU9fQLHWqzTqgQaUq4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.817994 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818004 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818014 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818137 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818163 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818175 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818187 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818199 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818243 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818262 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-CYMrCT-OsQ3-cdVr-EcBN-TKOH-xE0q-pkdaCm', 'scsi-0QEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6', 'scsi-SQEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818274 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XNIeJ-OvKH-SWIj-T9QN-B9GX-Jw9F-bWcj1x', 'scsi-0QEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6', 'scsi-SQEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818286 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290', 'scsi-SQEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818304 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-18-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:39:01.818315 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.818326 | orchestrator | 2025-08-29 17:39:01.818338 | orchestrator | TASK [ceph-facts : Check if the ceph conf exists] ****************************** 2025-08-29 17:39:01.818349 | orchestrator | Friday 29 August 2025 17:28:14 +0000 (0:00:01.182) 0:00:33.910 ********* 2025-08-29 17:39:01.818364 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.818376 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.818413 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.818440 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.818450 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.818460 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.818469 | orchestrator | 2025-08-29 17:39:01.818484 | orchestrator | TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] *************** 2025-08-29 17:39:01.818501 | orchestrator | Friday 29 August 2025 17:28:16 +0000 (0:00:01.399) 0:00:35.310 ********* 2025-08-29 17:39:01.818516 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.818532 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.818553 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.818574 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.818590 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.818606 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.818622 | orchestrator | 2025-08-29 17:39:01.818638 | orchestrator | TASK [ceph-facts : Read osd pool default crush rule] *************************** 2025-08-29 17:39:01.818654 | orchestrator | Friday 29 August 2025 17:28:16 +0000 (0:00:00.512) 0:00:35.823 ********* 2025-08-29 17:39:01.818670 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.818686 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.818703 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.818729 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.818746 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.818762 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.818778 | orchestrator | 2025-08-29 17:39:01.818794 | orchestrator | TASK [ceph-facts : Set osd_pool_default_crush_rule fact] *********************** 2025-08-29 17:39:01.818806 | orchestrator | Friday 29 August 2025 17:28:17 +0000 (0:00:00.908) 0:00:36.732 ********* 2025-08-29 17:39:01.818815 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.818825 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.818834 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.818843 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.818853 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.818862 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.818872 | orchestrator | 2025-08-29 17:39:01.818881 | orchestrator | TASK [ceph-facts : Read osd pool default crush rule] *************************** 2025-08-29 17:39:01.818891 | orchestrator | Friday 29 August 2025 17:28:18 +0000 (0:00:00.430) 0:00:37.162 ********* 2025-08-29 17:39:01.818910 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.818920 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.818929 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.818938 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.818948 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.818957 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.818967 | orchestrator | 2025-08-29 17:39:01.818976 | orchestrator | TASK [ceph-facts : Set osd_pool_default_crush_rule fact] *********************** 2025-08-29 17:39:01.818986 | orchestrator | Friday 29 August 2025 17:28:18 +0000 (0:00:00.821) 0:00:37.984 ********* 2025-08-29 17:39:01.818995 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.819005 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.819014 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.819023 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.819033 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.819042 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.819052 | orchestrator | 2025-08-29 17:39:01.819061 | orchestrator | TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] ************************* 2025-08-29 17:39:01.819071 | orchestrator | Friday 29 August 2025 17:28:19 +0000 (0:00:01.134) 0:00:39.119 ********* 2025-08-29 17:39:01.819080 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:39:01.819090 | orchestrator | ok: [testbed-node-2] => (item=testbed-node-0) 2025-08-29 17:39:01.819099 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-0) 2025-08-29 17:39:01.819109 | orchestrator | ok: [testbed-node-1] => (item=testbed-node-0) 2025-08-29 17:39:01.819118 | orchestrator | ok: [testbed-node-4] => (item=testbed-node-0) 2025-08-29 17:39:01.819127 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-1) 2025-08-29 17:39:01.819137 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-2) 2025-08-29 17:39:01.819146 | orchestrator | ok: [testbed-node-5] => (item=testbed-node-0) 2025-08-29 17:39:01.819155 | orchestrator | ok: [testbed-node-5] => (item=testbed-node-1) 2025-08-29 17:39:01.819165 | orchestrator | ok: [testbed-node-5] => (item=testbed-node-2) 2025-08-29 17:39:01.819174 | orchestrator | ok: [testbed-node-4] => (item=testbed-node-1) 2025-08-29 17:39:01.819184 | orchestrator | ok: [testbed-node-2] => (item=testbed-node-1) 2025-08-29 17:39:01.819193 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-1) 2025-08-29 17:39:01.819203 | orchestrator | ok: [testbed-node-4] => (item=testbed-node-2) 2025-08-29 17:39:01.819212 | orchestrator | ok: [testbed-node-1] => (item=testbed-node-1) 2025-08-29 17:39:01.819221 | orchestrator | ok: [testbed-node-2] => (item=testbed-node-2) 2025-08-29 17:39:01.819231 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-2) 2025-08-29 17:39:01.819240 | orchestrator | ok: [testbed-node-1] => (item=testbed-node-2) 2025-08-29 17:39:01.819249 | orchestrator | 2025-08-29 17:39:01.819259 | orchestrator | TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] ************************* 2025-08-29 17:39:01.819268 | orchestrator | Friday 29 August 2025 17:28:24 +0000 (0:00:04.805) 0:00:43.924 ********* 2025-08-29 17:39:01.819278 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-0)  2025-08-29 17:39:01.819287 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-1)  2025-08-29 17:39:01.819297 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-2)  2025-08-29 17:39:01.819306 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.819316 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-0)  2025-08-29 17:39:01.819325 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-1)  2025-08-29 17:39:01.819334 | orchestrator | skipping: [testbed-node-1] => (item=testbed-node-2)  2025-08-29 17:39:01.819344 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.819353 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-0)  2025-08-29 17:39:01.819363 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-1)  2025-08-29 17:39:01.819372 | orchestrator | skipping: [testbed-node-2] => (item=testbed-node-2)  2025-08-29 17:39:01.819457 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-0)  2025-08-29 17:39:01.819480 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-1)  2025-08-29 17:39:01.819490 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.819500 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-2)  2025-08-29 17:39:01.819510 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-0)  2025-08-29 17:39:01.819518 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-1)  2025-08-29 17:39:01.819526 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-2)  2025-08-29 17:39:01.819534 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.819542 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.819550 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-0)  2025-08-29 17:39:01.819558 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-1)  2025-08-29 17:39:01.819566 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-2)  2025-08-29 17:39:01.819574 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.819581 | orchestrator | 2025-08-29 17:39:01.819590 | orchestrator | TASK [ceph-facts : Import_tasks set_radosgw_address.yml] *********************** 2025-08-29 17:39:01.819598 | orchestrator | Friday 29 August 2025 17:28:25 +0000 (0:00:00.708) 0:00:44.632 ********* 2025-08-29 17:39:01.819605 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.819613 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.819621 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.819630 | orchestrator | included: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.819638 | orchestrator | 2025-08-29 17:39:01.819646 | orchestrator | TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface from node "{{ ceph_dashboard_call_item }}"] *** 2025-08-29 17:39:01.819655 | orchestrator | Friday 29 August 2025 17:28:27 +0000 (0:00:01.782) 0:00:46.414 ********* 2025-08-29 17:39:01.819663 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.819671 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.819679 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.819687 | orchestrator | 2025-08-29 17:39:01.819695 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] **** 2025-08-29 17:39:01.819703 | orchestrator | Friday 29 August 2025 17:28:27 +0000 (0:00:00.380) 0:00:46.796 ********* 2025-08-29 17:39:01.819710 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.819718 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.819726 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.819734 | orchestrator | 2025-08-29 17:39:01.819742 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] **** 2025-08-29 17:39:01.819750 | orchestrator | Friday 29 August 2025 17:28:28 +0000 (0:00:00.603) 0:00:47.399 ********* 2025-08-29 17:39:01.819758 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.819766 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.819774 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.819782 | orchestrator | 2025-08-29 17:39:01.819790 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] *************** 2025-08-29 17:39:01.819798 | orchestrator | Friday 29 August 2025 17:28:28 +0000 (0:00:00.395) 0:00:47.795 ********* 2025-08-29 17:39:01.819805 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.819813 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.819821 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.819829 | orchestrator | 2025-08-29 17:39:01.819837 | orchestrator | TASK [ceph-facts : Set_fact _interface] **************************************** 2025-08-29 17:39:01.819845 | orchestrator | Friday 29 August 2025 17:28:29 +0000 (0:00:00.585) 0:00:48.380 ********* 2025-08-29 17:39:01.819853 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.819861 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.819869 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.819882 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.819890 | orchestrator | 2025-08-29 17:39:01.819898 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ****** 2025-08-29 17:39:01.819906 | orchestrator | Friday 29 August 2025 17:28:29 +0000 (0:00:00.335) 0:00:48.716 ********* 2025-08-29 17:39:01.819914 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.819922 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.819930 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.819937 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.819945 | orchestrator | 2025-08-29 17:39:01.819953 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ****** 2025-08-29 17:39:01.819961 | orchestrator | Friday 29 August 2025 17:28:29 +0000 (0:00:00.303) 0:00:49.020 ********* 2025-08-29 17:39:01.819969 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.819977 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.819985 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.819992 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.820000 | orchestrator | 2025-08-29 17:39:01.820008 | orchestrator | TASK [ceph-facts : Reset rgw_instances (workaround)] *************************** 2025-08-29 17:39:01.820016 | orchestrator | Friday 29 August 2025 17:28:30 +0000 (0:00:00.477) 0:00:49.497 ********* 2025-08-29 17:39:01.820024 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.820032 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.820040 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.820048 | orchestrator | 2025-08-29 17:39:01.820056 | orchestrator | TASK [ceph-facts : Set_fact rgw_instances] ************************************* 2025-08-29 17:39:01.820064 | orchestrator | Friday 29 August 2025 17:28:30 +0000 (0:00:00.618) 0:00:50.116 ********* 2025-08-29 17:39:01.820072 | orchestrator | ok: [testbed-node-3] => (item=0) 2025-08-29 17:39:01.820080 | orchestrator | ok: [testbed-node-4] => (item=0) 2025-08-29 17:39:01.820088 | orchestrator | ok: [testbed-node-5] => (item=0) 2025-08-29 17:39:01.820096 | orchestrator | 2025-08-29 17:39:01.820104 | orchestrator | TASK [ceph-facts : Set_fact ceph_run_cmd] ************************************** 2025-08-29 17:39:01.820116 | orchestrator | Friday 29 August 2025 17:28:31 +0000 (0:00:00.628) 0:00:50.745 ********* 2025-08-29 17:39:01.820128 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:39:01.820136 | orchestrator | ok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:39:01.820145 | orchestrator | ok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:39:01.820152 | orchestrator | ok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3) 2025-08-29 17:39:01.820160 | orchestrator | ok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4) 2025-08-29 17:39:01.820168 | orchestrator | ok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5) 2025-08-29 17:39:01.820176 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager) 2025-08-29 17:39:01.820184 | orchestrator | 2025-08-29 17:39:01.820192 | orchestrator | TASK [ceph-facts : Set_fact ceph_admin_command] ******************************** 2025-08-29 17:39:01.820200 | orchestrator | Friday 29 August 2025 17:28:32 +0000 (0:00:01.081) 0:00:51.826 ********* 2025-08-29 17:39:01.820208 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:39:01.820216 | orchestrator | ok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:39:01.820224 | orchestrator | ok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:39:01.820232 | orchestrator | ok: [testbed-node-0 -> testbed-node-3(192.168.16.13)] => (item=testbed-node-3) 2025-08-29 17:39:01.820240 | orchestrator | ok: [testbed-node-0 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4) 2025-08-29 17:39:01.820248 | orchestrator | ok: [testbed-node-0 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5) 2025-08-29 17:39:01.820261 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => (item=testbed-manager) 2025-08-29 17:39:01.820269 | orchestrator | 2025-08-29 17:39:01.820277 | orchestrator | TASK [ceph-handler : Include check_running_cluster.yml] ************************ 2025-08-29 17:39:01.820285 | orchestrator | Friday 29 August 2025 17:28:34 +0000 (0:00:01.963) 0:00:53.790 ********* 2025-08-29 17:39:01.820294 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.820303 | orchestrator | 2025-08-29 17:39:01.820311 | orchestrator | TASK [ceph-handler : Include check_running_containers.yml] ********************* 2025-08-29 17:39:01.820319 | orchestrator | Friday 29 August 2025 17:28:35 +0000 (0:00:01.125) 0:00:54.916 ********* 2025-08-29 17:39:01.820327 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.820335 | orchestrator | 2025-08-29 17:39:01.820343 | orchestrator | TASK [ceph-handler : Check for a mon container] ******************************** 2025-08-29 17:39:01.820351 | orchestrator | Friday 29 August 2025 17:28:36 +0000 (0:00:01.206) 0:00:56.122 ********* 2025-08-29 17:39:01.820359 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.820367 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.820375 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.820383 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.820409 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.820418 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.820426 | orchestrator | 2025-08-29 17:39:01.820434 | orchestrator | TASK [ceph-handler : Check for an osd container] ******************************* 2025-08-29 17:39:01.820442 | orchestrator | Friday 29 August 2025 17:28:38 +0000 (0:00:01.134) 0:00:57.256 ********* 2025-08-29 17:39:01.820450 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.820457 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.820465 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.820473 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.820481 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.820489 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.820497 | orchestrator | 2025-08-29 17:39:01.820505 | orchestrator | TASK [ceph-handler : Check for a mds container] ******************************** 2025-08-29 17:39:01.820513 | orchestrator | Friday 29 August 2025 17:28:39 +0000 (0:00:01.427) 0:00:58.684 ********* 2025-08-29 17:39:01.820521 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.820529 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.820537 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.820545 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.820552 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.820560 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.820568 | orchestrator | 2025-08-29 17:39:01.820576 | orchestrator | TASK [ceph-handler : Check for a rgw container] ******************************** 2025-08-29 17:39:01.820584 | orchestrator | Friday 29 August 2025 17:28:40 +0000 (0:00:01.167) 0:00:59.851 ********* 2025-08-29 17:39:01.820592 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.820600 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.820608 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.820616 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.820624 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.820631 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.820639 | orchestrator | 2025-08-29 17:39:01.820647 | orchestrator | TASK [ceph-handler : Check for a mgr container] ******************************** 2025-08-29 17:39:01.820655 | orchestrator | Friday 29 August 2025 17:28:41 +0000 (0:00:01.084) 0:01:00.936 ********* 2025-08-29 17:39:01.820663 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.820671 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.820679 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.820707 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.820715 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.820723 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.820731 | orchestrator | 2025-08-29 17:39:01.820739 | orchestrator | TASK [ceph-handler : Check for a rbd mirror container] ************************* 2025-08-29 17:39:01.820751 | orchestrator | Friday 29 August 2025 17:28:42 +0000 (0:00:00.864) 0:01:01.800 ********* 2025-08-29 17:39:01.820764 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.820773 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.820781 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.820788 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.820796 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.820804 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.820812 | orchestrator | 2025-08-29 17:39:01.820820 | orchestrator | TASK [ceph-handler : Check for a nfs container] ******************************** 2025-08-29 17:39:01.820828 | orchestrator | Friday 29 August 2025 17:28:43 +0000 (0:00:00.573) 0:01:02.374 ********* 2025-08-29 17:39:01.820836 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.820844 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.820851 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.820859 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.820867 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.820875 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.820882 | orchestrator | 2025-08-29 17:39:01.820890 | orchestrator | TASK [ceph-handler : Check for a ceph-crash container] ************************* 2025-08-29 17:39:01.820898 | orchestrator | Friday 29 August 2025 17:28:44 +0000 (0:00:00.926) 0:01:03.300 ********* 2025-08-29 17:39:01.820906 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.820914 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.820922 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.820930 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.820937 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.820945 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.820953 | orchestrator | 2025-08-29 17:39:01.820961 | orchestrator | TASK [ceph-handler : Check for a ceph-exporter container] ********************** 2025-08-29 17:39:01.820969 | orchestrator | Friday 29 August 2025 17:28:45 +0000 (0:00:00.913) 0:01:04.214 ********* 2025-08-29 17:39:01.820977 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.820985 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.820992 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.821000 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.821008 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.821016 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.821023 | orchestrator | 2025-08-29 17:39:01.821031 | orchestrator | TASK [ceph-handler : Include check_socket_non_container.yml] ******************* 2025-08-29 17:39:01.821039 | orchestrator | Friday 29 August 2025 17:28:46 +0000 (0:00:01.456) 0:01:05.672 ********* 2025-08-29 17:39:01.821047 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.821055 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.821063 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.821071 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.821079 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.821086 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.821094 | orchestrator | 2025-08-29 17:39:01.821102 | orchestrator | TASK [ceph-handler : Set_fact handler_mon_status] ****************************** 2025-08-29 17:39:01.821110 | orchestrator | Friday 29 August 2025 17:28:47 +0000 (0:00:00.657) 0:01:06.329 ********* 2025-08-29 17:39:01.821118 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.821126 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.821134 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.821141 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.821149 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.821157 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.821165 | orchestrator | 2025-08-29 17:39:01.821173 | orchestrator | TASK [ceph-handler : Set_fact handler_osd_status] ****************************** 2025-08-29 17:39:01.821186 | orchestrator | Friday 29 August 2025 17:28:48 +0000 (0:00:01.224) 0:01:07.553 ********* 2025-08-29 17:39:01.821194 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.821201 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.821209 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.821217 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.821225 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.821233 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.821241 | orchestrator | 2025-08-29 17:39:01.821249 | orchestrator | TASK [ceph-handler : Set_fact handler_mds_status] ****************************** 2025-08-29 17:39:01.821257 | orchestrator | Friday 29 August 2025 17:28:49 +0000 (0:00:00.930) 0:01:08.483 ********* 2025-08-29 17:39:01.821265 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.821272 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.821280 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.821288 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.821296 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.821303 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.821311 | orchestrator | 2025-08-29 17:39:01.821319 | orchestrator | TASK [ceph-handler : Set_fact handler_rgw_status] ****************************** 2025-08-29 17:39:01.821327 | orchestrator | Friday 29 August 2025 17:28:50 +0000 (0:00:01.230) 0:01:09.714 ********* 2025-08-29 17:39:01.821335 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.821343 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.821351 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.821359 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.821367 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.821374 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.821382 | orchestrator | 2025-08-29 17:39:01.821408 | orchestrator | TASK [ceph-handler : Set_fact handler_nfs_status] ****************************** 2025-08-29 17:39:01.821416 | orchestrator | Friday 29 August 2025 17:28:51 +0000 (0:00:00.711) 0:01:10.425 ********* 2025-08-29 17:39:01.821424 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.821432 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.821440 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.821447 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.821455 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.821463 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.821471 | orchestrator | 2025-08-29 17:39:01.821479 | orchestrator | TASK [ceph-handler : Set_fact handler_rbd_status] ****************************** 2025-08-29 17:39:01.821486 | orchestrator | Friday 29 August 2025 17:28:52 +0000 (0:00:00.778) 0:01:11.204 ********* 2025-08-29 17:39:01.821494 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.821502 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.821510 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.821517 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.821525 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.821533 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.821541 | orchestrator | 2025-08-29 17:39:01.821553 | orchestrator | TASK [ceph-handler : Set_fact handler_mgr_status] ****************************** 2025-08-29 17:39:01.821565 | orchestrator | Friday 29 August 2025 17:28:52 +0000 (0:00:00.540) 0:01:11.744 ********* 2025-08-29 17:39:01.821574 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.821582 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.821589 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.821597 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.821605 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.821612 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.821620 | orchestrator | 2025-08-29 17:39:01.821628 | orchestrator | TASK [ceph-handler : Set_fact handler_crash_status] **************************** 2025-08-29 17:39:01.821636 | orchestrator | Friday 29 August 2025 17:28:53 +0000 (0:00:00.739) 0:01:12.484 ********* 2025-08-29 17:39:01.821644 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.821652 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.821665 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.821673 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.821680 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.821688 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.821696 | orchestrator | 2025-08-29 17:39:01.821704 | orchestrator | TASK [ceph-handler : Set_fact handler_exporter_status] ************************* 2025-08-29 17:39:01.821712 | orchestrator | Friday 29 August 2025 17:28:53 +0000 (0:00:00.634) 0:01:13.119 ********* 2025-08-29 17:39:01.821720 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.821728 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.821735 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.821743 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.821751 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.821758 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.821766 | orchestrator | 2025-08-29 17:39:01.821774 | orchestrator | TASK [ceph-container-common : Generate systemd ceph target file] *************** 2025-08-29 17:39:01.821787 | orchestrator | Friday 29 August 2025 17:28:55 +0000 (0:00:01.184) 0:01:14.303 ********* 2025-08-29 17:39:01.821801 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.821814 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.821826 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.821839 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.821851 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.821865 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.821878 | orchestrator | 2025-08-29 17:39:01.821892 | orchestrator | TASK [ceph-container-common : Enable ceph.target] ****************************** 2025-08-29 17:39:01.821906 | orchestrator | Friday 29 August 2025 17:28:57 +0000 (0:00:01.850) 0:01:16.154 ********* 2025-08-29 17:39:01.821919 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.821933 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.821941 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.821949 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.821956 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.821964 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.821972 | orchestrator | 2025-08-29 17:39:01.821980 | orchestrator | TASK [ceph-container-common : Include prerequisites.yml] *********************** 2025-08-29 17:39:01.821988 | orchestrator | Friday 29 August 2025 17:28:59 +0000 (0:00:02.072) 0:01:18.226 ********* 2025-08-29 17:39:01.821996 | orchestrator | included: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.822004 | orchestrator | 2025-08-29 17:39:01.822012 | orchestrator | TASK [ceph-container-common : Stop lvmetad] ************************************ 2025-08-29 17:39:01.822057 | orchestrator | Friday 29 August 2025 17:29:00 +0000 (0:00:01.337) 0:01:19.563 ********* 2025-08-29 17:39:01.822065 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.822074 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.822082 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.822089 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.822097 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.822105 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.822113 | orchestrator | 2025-08-29 17:39:01.822121 | orchestrator | TASK [ceph-container-common : Disable and mask lvmetad service] **************** 2025-08-29 17:39:01.822129 | orchestrator | Friday 29 August 2025 17:29:01 +0000 (0:00:00.788) 0:01:20.352 ********* 2025-08-29 17:39:01.822137 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.822144 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.822152 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.822160 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.822168 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.822176 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.822183 | orchestrator | 2025-08-29 17:39:01.822191 | orchestrator | TASK [ceph-container-common : Remove ceph udev rules] ************************** 2025-08-29 17:39:01.822199 | orchestrator | Friday 29 August 2025 17:29:01 +0000 (0:00:00.551) 0:01:20.903 ********* 2025-08-29 17:39:01.822214 | orchestrator | ok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules) 2025-08-29 17:39:01.822222 | orchestrator | ok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules) 2025-08-29 17:39:01.822230 | orchestrator | ok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules) 2025-08-29 17:39:01.822238 | orchestrator | ok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules) 2025-08-29 17:39:01.822246 | orchestrator | ok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules) 2025-08-29 17:39:01.822253 | orchestrator | ok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules) 2025-08-29 17:39:01.822261 | orchestrator | ok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules) 2025-08-29 17:39:01.822269 | orchestrator | ok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules) 2025-08-29 17:39:01.822277 | orchestrator | ok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules) 2025-08-29 17:39:01.822285 | orchestrator | ok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules) 2025-08-29 17:39:01.822292 | orchestrator | ok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules) 2025-08-29 17:39:01.822305 | orchestrator | ok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules) 2025-08-29 17:39:01.822313 | orchestrator | 2025-08-29 17:39:01.822334 | orchestrator | TASK [ceph-container-common : Ensure tmpfiles.d is present] ******************** 2025-08-29 17:39:01.822342 | orchestrator | Friday 29 August 2025 17:29:03 +0000 (0:00:01.515) 0:01:22.419 ********* 2025-08-29 17:39:01.822350 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.822358 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.822366 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.822373 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.822381 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.822436 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.822445 | orchestrator | 2025-08-29 17:39:01.822453 | orchestrator | TASK [ceph-container-common : Restore certificates selinux context] ************ 2025-08-29 17:39:01.822461 | orchestrator | Friday 29 August 2025 17:29:04 +0000 (0:00:01.150) 0:01:23.570 ********* 2025-08-29 17:39:01.822469 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.822476 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.822484 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.822492 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.822500 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.822508 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.822516 | orchestrator | 2025-08-29 17:39:01.822524 | orchestrator | TASK [ceph-container-common : Install python3 on osd nodes] ******************** 2025-08-29 17:39:01.822532 | orchestrator | Friday 29 August 2025 17:29:05 +0000 (0:00:00.768) 0:01:24.339 ********* 2025-08-29 17:39:01.822540 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.822547 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.822555 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.822563 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.822571 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.822579 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.822586 | orchestrator | 2025-08-29 17:39:01.822594 | orchestrator | TASK [ceph-container-common : Include registry.yml] **************************** 2025-08-29 17:39:01.822602 | orchestrator | Friday 29 August 2025 17:29:05 +0000 (0:00:00.574) 0:01:24.914 ********* 2025-08-29 17:39:01.822610 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.822618 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.822626 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.822634 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.822641 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.822649 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.822668 | orchestrator | 2025-08-29 17:39:01.822676 | orchestrator | TASK [ceph-container-common : Include fetch_image.yml] ************************* 2025-08-29 17:39:01.822684 | orchestrator | Friday 29 August 2025 17:29:06 +0000 (0:00:00.758) 0:01:25.672 ********* 2025-08-29 17:39:01.822693 | orchestrator | included: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.822701 | orchestrator | 2025-08-29 17:39:01.822709 | orchestrator | TASK [ceph-container-common : Pulling Ceph container image] ******************** 2025-08-29 17:39:01.822717 | orchestrator | Friday 29 August 2025 17:29:07 +0000 (0:00:01.109) 0:01:26.782 ********* 2025-08-29 17:39:01.822725 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.822732 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.822738 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.822745 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.822751 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.822758 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.822765 | orchestrator | 2025-08-29 17:39:01.822771 | orchestrator | TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] *** 2025-08-29 17:39:01.822778 | orchestrator | Friday 29 August 2025 17:30:26 +0000 (0:01:19.345) 0:02:46.128 ********* 2025-08-29 17:39:01.822785 | orchestrator | skipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2)  2025-08-29 17:39:01.822791 | orchestrator | skipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2)  2025-08-29 17:39:01.822798 | orchestrator | skipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4)  2025-08-29 17:39:01.822804 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.822811 | orchestrator | skipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2)  2025-08-29 17:39:01.822818 | orchestrator | skipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2)  2025-08-29 17:39:01.822824 | orchestrator | skipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4)  2025-08-29 17:39:01.822831 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.822838 | orchestrator | skipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2)  2025-08-29 17:39:01.822844 | orchestrator | skipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2)  2025-08-29 17:39:01.822871 | orchestrator | skipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4)  2025-08-29 17:39:01.822878 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.822885 | orchestrator | skipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2)  2025-08-29 17:39:01.822891 | orchestrator | skipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2)  2025-08-29 17:39:01.822898 | orchestrator | skipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4)  2025-08-29 17:39:01.822905 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.822917 | orchestrator | skipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2)  2025-08-29 17:39:01.822928 | orchestrator | skipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2)  2025-08-29 17:39:01.822939 | orchestrator | skipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4)  2025-08-29 17:39:01.822950 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.822961 | orchestrator | skipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2)  2025-08-29 17:39:01.822971 | orchestrator | skipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2)  2025-08-29 17:39:01.822988 | orchestrator | skipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4)  2025-08-29 17:39:01.823006 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823018 | orchestrator | 2025-08-29 17:39:01.823027 | orchestrator | TASK [ceph-container-common : Pulling node-exporter container image] *********** 2025-08-29 17:39:01.823034 | orchestrator | Friday 29 August 2025 17:30:27 +0000 (0:00:00.678) 0:02:46.806 ********* 2025-08-29 17:39:01.823041 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823047 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823060 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823067 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823073 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823080 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823086 | orchestrator | 2025-08-29 17:39:01.823093 | orchestrator | TASK [ceph-container-common : Export local ceph dev image] ********************* 2025-08-29 17:39:01.823099 | orchestrator | Friday 29 August 2025 17:30:28 +0000 (0:00:00.463) 0:02:47.269 ********* 2025-08-29 17:39:01.823106 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823112 | orchestrator | 2025-08-29 17:39:01.823119 | orchestrator | TASK [ceph-container-common : Copy ceph dev image file] ************************ 2025-08-29 17:39:01.823126 | orchestrator | Friday 29 August 2025 17:30:28 +0000 (0:00:00.200) 0:02:47.469 ********* 2025-08-29 17:39:01.823132 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823139 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823145 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823152 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823158 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823165 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823171 | orchestrator | 2025-08-29 17:39:01.823178 | orchestrator | TASK [ceph-container-common : Load ceph dev image] ***************************** 2025-08-29 17:39:01.823185 | orchestrator | Friday 29 August 2025 17:30:29 +0000 (0:00:00.702) 0:02:48.172 ********* 2025-08-29 17:39:01.823191 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823198 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823204 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823210 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823217 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823223 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823230 | orchestrator | 2025-08-29 17:39:01.823237 | orchestrator | TASK [ceph-container-common : Remove tmp ceph dev image file] ****************** 2025-08-29 17:39:01.823243 | orchestrator | Friday 29 August 2025 17:30:29 +0000 (0:00:00.605) 0:02:48.777 ********* 2025-08-29 17:39:01.823250 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823256 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823263 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823269 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823276 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823282 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823289 | orchestrator | 2025-08-29 17:39:01.823296 | orchestrator | TASK [ceph-container-common : Get ceph version] ******************************** 2025-08-29 17:39:01.823302 | orchestrator | Friday 29 August 2025 17:30:30 +0000 (0:00:00.707) 0:02:49.485 ********* 2025-08-29 17:39:01.823309 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.823316 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.823322 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.823329 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.823335 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.823342 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.823348 | orchestrator | 2025-08-29 17:39:01.823355 | orchestrator | TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] *** 2025-08-29 17:39:01.823362 | orchestrator | Friday 29 August 2025 17:30:33 +0000 (0:00:03.255) 0:02:52.741 ********* 2025-08-29 17:39:01.823368 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.823375 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.823381 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.823407 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.823414 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.823420 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.823427 | orchestrator | 2025-08-29 17:39:01.823434 | orchestrator | TASK [ceph-container-common : Include release.yml] ***************************** 2025-08-29 17:39:01.823440 | orchestrator | Friday 29 August 2025 17:30:34 +0000 (0:00:00.891) 0:02:53.633 ********* 2025-08-29 17:39:01.823447 | orchestrator | included: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.823460 | orchestrator | 2025-08-29 17:39:01.823467 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release jewel] ********************* 2025-08-29 17:39:01.823473 | orchestrator | Friday 29 August 2025 17:30:35 +0000 (0:00:01.289) 0:02:54.923 ********* 2025-08-29 17:39:01.823480 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823487 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823493 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823500 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823506 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823513 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823520 | orchestrator | 2025-08-29 17:39:01.823526 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release kraken] ******************** 2025-08-29 17:39:01.823533 | orchestrator | Friday 29 August 2025 17:30:36 +0000 (0:00:00.734) 0:02:55.657 ********* 2025-08-29 17:39:01.823539 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823546 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823553 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823559 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823566 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823573 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823579 | orchestrator | 2025-08-29 17:39:01.823586 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release luminous] ****************** 2025-08-29 17:39:01.823593 | orchestrator | Friday 29 August 2025 17:30:37 +0000 (0:00:01.001) 0:02:56.658 ********* 2025-08-29 17:39:01.823599 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823606 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823612 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823619 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823626 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823632 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823639 | orchestrator | 2025-08-29 17:39:01.823649 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release mimic] ********************* 2025-08-29 17:39:01.823660 | orchestrator | Friday 29 August 2025 17:30:38 +0000 (0:00:00.546) 0:02:57.204 ********* 2025-08-29 17:39:01.823667 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823673 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823680 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823687 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823693 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823700 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823707 | orchestrator | 2025-08-29 17:39:01.823714 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release nautilus] ****************** 2025-08-29 17:39:01.823720 | orchestrator | Friday 29 August 2025 17:30:38 +0000 (0:00:00.653) 0:02:57.858 ********* 2025-08-29 17:39:01.823727 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823734 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823740 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823747 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823753 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823760 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823767 | orchestrator | 2025-08-29 17:39:01.823774 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release octopus] ******************* 2025-08-29 17:39:01.823780 | orchestrator | Friday 29 August 2025 17:30:39 +0000 (0:00:00.767) 0:02:58.625 ********* 2025-08-29 17:39:01.823787 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823794 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823800 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823807 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823813 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823820 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823827 | orchestrator | 2025-08-29 17:39:01.823838 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release pacific] ******************* 2025-08-29 17:39:01.823845 | orchestrator | Friday 29 August 2025 17:30:40 +0000 (0:00:00.986) 0:02:59.612 ********* 2025-08-29 17:39:01.823851 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823858 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823865 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823871 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823878 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823884 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823891 | orchestrator | 2025-08-29 17:39:01.823898 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release quincy] ******************** 2025-08-29 17:39:01.823907 | orchestrator | Friday 29 August 2025 17:30:41 +0000 (0:00:00.742) 0:03:00.354 ********* 2025-08-29 17:39:01.823918 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.823929 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.823946 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.823959 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.823969 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.823979 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.823990 | orchestrator | 2025-08-29 17:39:01.824001 | orchestrator | TASK [ceph-container-common : Set_fact ceph_release reef] ********************** 2025-08-29 17:39:01.824010 | orchestrator | Friday 29 August 2025 17:30:42 +0000 (0:00:01.011) 0:03:01.366 ********* 2025-08-29 17:39:01.824021 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.824030 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.824039 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.824048 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.824058 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.824069 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.824078 | orchestrator | 2025-08-29 17:39:01.824088 | orchestrator | TASK [ceph-config : Include create_ceph_initial_dirs.yml] ********************** 2025-08-29 17:39:01.824098 | orchestrator | Friday 29 August 2025 17:30:43 +0000 (0:00:01.033) 0:03:02.399 ********* 2025-08-29 17:39:01.824108 | orchestrator | included: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.824119 | orchestrator | 2025-08-29 17:39:01.824129 | orchestrator | TASK [ceph-config : Create ceph initial directories] *************************** 2025-08-29 17:39:01.824139 | orchestrator | Friday 29 August 2025 17:30:44 +0000 (0:00:01.075) 0:03:03.475 ********* 2025-08-29 17:39:01.824148 | orchestrator | changed: [testbed-node-0] => (item=/etc/ceph) 2025-08-29 17:39:01.824159 | orchestrator | changed: [testbed-node-2] => (item=/etc/ceph) 2025-08-29 17:39:01.824169 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/) 2025-08-29 17:39:01.824179 | orchestrator | changed: [testbed-node-3] => (item=/etc/ceph) 2025-08-29 17:39:01.824190 | orchestrator | changed: [testbed-node-1] => (item=/etc/ceph) 2025-08-29 17:39:01.824201 | orchestrator | changed: [testbed-node-5] => (item=/etc/ceph) 2025-08-29 17:39:01.824211 | orchestrator | changed: [testbed-node-4] => (item=/etc/ceph) 2025-08-29 17:39:01.824221 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/) 2025-08-29 17:39:01.824232 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/mon) 2025-08-29 17:39:01.824243 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/) 2025-08-29 17:39:01.824254 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/) 2025-08-29 17:39:01.824266 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/) 2025-08-29 17:39:01.824278 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/mon) 2025-08-29 17:39:01.824288 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/) 2025-08-29 17:39:01.824298 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/mon) 2025-08-29 17:39:01.824309 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/mon) 2025-08-29 17:39:01.824321 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/osd) 2025-08-29 17:39:01.824342 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/osd) 2025-08-29 17:39:01.824353 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/mon) 2025-08-29 17:39:01.824364 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/mon) 2025-08-29 17:39:01.824380 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/osd) 2025-08-29 17:39:01.824420 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/osd) 2025-08-29 17:39:01.824431 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/mds) 2025-08-29 17:39:01.824442 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/mds) 2025-08-29 17:39:01.824452 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/osd) 2025-08-29 17:39:01.824462 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/osd) 2025-08-29 17:39:01.824472 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/mds) 2025-08-29 17:39:01.824483 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/mds) 2025-08-29 17:39:01.824493 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/tmp) 2025-08-29 17:39:01.824505 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/tmp) 2025-08-29 17:39:01.824514 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/mds) 2025-08-29 17:39:01.824524 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/mds) 2025-08-29 17:39:01.824534 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/tmp) 2025-08-29 17:39:01.824544 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/tmp) 2025-08-29 17:39:01.824554 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/crash) 2025-08-29 17:39:01.824565 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/tmp) 2025-08-29 17:39:01.824576 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/tmp) 2025-08-29 17:39:01.824585 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/crash) 2025-08-29 17:39:01.824596 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/crash) 2025-08-29 17:39:01.824607 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/crash) 2025-08-29 17:39:01.824617 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/radosgw) 2025-08-29 17:39:01.824627 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/crash) 2025-08-29 17:39:01.824639 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/crash) 2025-08-29 17:39:01.824649 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/radosgw) 2025-08-29 17:39:01.824660 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/radosgw) 2025-08-29 17:39:01.824672 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/radosgw) 2025-08-29 17:39:01.824684 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw) 2025-08-29 17:39:01.824695 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/radosgw) 2025-08-29 17:39:01.824706 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/radosgw) 2025-08-29 17:39:01.824716 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw) 2025-08-29 17:39:01.824728 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw) 2025-08-29 17:39:01.824739 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw) 2025-08-29 17:39:01.824749 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr) 2025-08-29 17:39:01.824760 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw) 2025-08-29 17:39:01.824787 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw) 2025-08-29 17:39:01.824798 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr) 2025-08-29 17:39:01.824819 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr) 2025-08-29 17:39:01.824829 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr) 2025-08-29 17:39:01.824839 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds) 2025-08-29 17:39:01.824864 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr) 2025-08-29 17:39:01.824874 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr) 2025-08-29 17:39:01.824885 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds) 2025-08-29 17:39:01.824895 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds) 2025-08-29 17:39:01.824905 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds) 2025-08-29 17:39:01.824916 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd) 2025-08-29 17:39:01.824927 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds) 2025-08-29 17:39:01.824939 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds) 2025-08-29 17:39:01.824949 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd) 2025-08-29 17:39:01.824959 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd) 2025-08-29 17:39:01.824971 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd) 2025-08-29 17:39:01.824982 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd) 2025-08-29 17:39:01.824994 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd) 2025-08-29 17:39:01.825004 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd) 2025-08-29 17:39:01.825015 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd) 2025-08-29 17:39:01.825027 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd) 2025-08-29 17:39:01.825038 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd) 2025-08-29 17:39:01.825049 | orchestrator | changed: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror) 2025-08-29 17:39:01.825060 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd) 2025-08-29 17:39:01.825077 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror) 2025-08-29 17:39:01.825098 | orchestrator | changed: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror) 2025-08-29 17:39:01.825109 | orchestrator | changed: [testbed-node-2] => (item=/var/run/ceph) 2025-08-29 17:39:01.825120 | orchestrator | changed: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror) 2025-08-29 17:39:01.825130 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd) 2025-08-29 17:39:01.825139 | orchestrator | changed: [testbed-node-3] => (item=/var/run/ceph) 2025-08-29 17:39:01.825149 | orchestrator | changed: [testbed-node-0] => (item=/var/run/ceph) 2025-08-29 17:39:01.825159 | orchestrator | changed: [testbed-node-2] => (item=/var/log/ceph) 2025-08-29 17:39:01.825170 | orchestrator | changed: [testbed-node-1] => (item=/var/run/ceph) 2025-08-29 17:39:01.825181 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror) 2025-08-29 17:39:01.825191 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror) 2025-08-29 17:39:01.825201 | orchestrator | changed: [testbed-node-0] => (item=/var/log/ceph) 2025-08-29 17:39:01.825212 | orchestrator | changed: [testbed-node-1] => (item=/var/log/ceph) 2025-08-29 17:39:01.825223 | orchestrator | changed: [testbed-node-5] => (item=/var/run/ceph) 2025-08-29 17:39:01.825234 | orchestrator | changed: [testbed-node-4] => (item=/var/run/ceph) 2025-08-29 17:39:01.825244 | orchestrator | changed: [testbed-node-3] => (item=/var/log/ceph) 2025-08-29 17:39:01.825255 | orchestrator | changed: [testbed-node-5] => (item=/var/log/ceph) 2025-08-29 17:39:01.825267 | orchestrator | changed: [testbed-node-4] => (item=/var/log/ceph) 2025-08-29 17:39:01.825278 | orchestrator | 2025-08-29 17:39:01.825289 | orchestrator | TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************ 2025-08-29 17:39:01.825301 | orchestrator | Friday 29 August 2025 17:30:50 +0000 (0:00:06.262) 0:03:09.737 ********* 2025-08-29 17:39:01.825311 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.825331 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.825342 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.825354 | orchestrator | included: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.825365 | orchestrator | 2025-08-29 17:39:01.825375 | orchestrator | TASK [ceph-config : Create rados gateway instance directories] ***************** 2025-08-29 17:39:01.825407 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.802) 0:03:10.540 ********* 2025-08-29 17:39:01.825418 | orchestrator | changed: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.825429 | orchestrator | changed: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.825440 | orchestrator | changed: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.825449 | orchestrator | 2025-08-29 17:39:01.825459 | orchestrator | TASK [ceph-config : Generate environment file] ********************************* 2025-08-29 17:39:01.825469 | orchestrator | Friday 29 August 2025 17:30:51 +0000 (0:00:00.596) 0:03:11.137 ********* 2025-08-29 17:39:01.825480 | orchestrator | changed: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.825489 | orchestrator | changed: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.825499 | orchestrator | changed: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.825509 | orchestrator | 2025-08-29 17:39:01.825520 | orchestrator | TASK [ceph-config : Reset num_osds] ******************************************** 2025-08-29 17:39:01.825531 | orchestrator | Friday 29 August 2025 17:30:53 +0000 (0:00:01.213) 0:03:12.351 ********* 2025-08-29 17:39:01.825542 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.825552 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.825563 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.825575 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.825587 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.825597 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.825609 | orchestrator | 2025-08-29 17:39:01.825621 | orchestrator | TASK [ceph-config : Count number of osds for lvm scenario] ********************* 2025-08-29 17:39:01.825633 | orchestrator | Friday 29 August 2025 17:30:53 +0000 (0:00:00.695) 0:03:13.046 ********* 2025-08-29 17:39:01.825643 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.825655 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.825665 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.825675 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.825685 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.825696 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.825706 | orchestrator | 2025-08-29 17:39:01.825716 | orchestrator | TASK [ceph-config : Look up for ceph-volume rejected devices] ****************** 2025-08-29 17:39:01.825727 | orchestrator | Friday 29 August 2025 17:30:54 +0000 (0:00:00.524) 0:03:13.570 ********* 2025-08-29 17:39:01.825737 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.825747 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.825758 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.825768 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.825779 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.825789 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.825800 | orchestrator | 2025-08-29 17:39:01.825810 | orchestrator | TASK [ceph-config : Set_fact rejected_devices] ********************************* 2025-08-29 17:39:01.825821 | orchestrator | Friday 29 August 2025 17:30:55 +0000 (0:00:00.690) 0:03:14.261 ********* 2025-08-29 17:39:01.825839 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.825850 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.825881 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.825893 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.825903 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.825914 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.825925 | orchestrator | 2025-08-29 17:39:01.825937 | orchestrator | TASK [ceph-config : Set_fact _devices] ***************************************** 2025-08-29 17:39:01.825947 | orchestrator | Friday 29 August 2025 17:30:55 +0000 (0:00:00.617) 0:03:14.878 ********* 2025-08-29 17:39:01.825958 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.825968 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.825978 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.825990 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.825997 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826004 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826010 | orchestrator | 2025-08-29 17:39:01.826055 | orchestrator | TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] *** 2025-08-29 17:39:01.826065 | orchestrator | Friday 29 August 2025 17:30:56 +0000 (0:00:00.702) 0:03:15.581 ********* 2025-08-29 17:39:01.826071 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826078 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826085 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826091 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826098 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826104 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826111 | orchestrator | 2025-08-29 17:39:01.826117 | orchestrator | TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] *** 2025-08-29 17:39:01.826124 | orchestrator | Friday 29 August 2025 17:30:57 +0000 (0:00:00.693) 0:03:16.274 ********* 2025-08-29 17:39:01.826131 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826137 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826144 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826150 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826157 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826163 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826170 | orchestrator | 2025-08-29 17:39:01.826177 | orchestrator | TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] *** 2025-08-29 17:39:01.826184 | orchestrator | Friday 29 August 2025 17:30:57 +0000 (0:00:00.769) 0:03:17.044 ********* 2025-08-29 17:39:01.826191 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826197 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826204 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826210 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826217 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826223 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826230 | orchestrator | 2025-08-29 17:39:01.826236 | orchestrator | TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] *** 2025-08-29 17:39:01.826243 | orchestrator | Friday 29 August 2025 17:30:58 +0000 (0:00:00.632) 0:03:17.676 ********* 2025-08-29 17:39:01.826250 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826256 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826263 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826269 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.826276 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.826282 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.826289 | orchestrator | 2025-08-29 17:39:01.826295 | orchestrator | TASK [ceph-config : Set_fact num_osds (add existing osds)] ********************* 2025-08-29 17:39:01.826302 | orchestrator | Friday 29 August 2025 17:31:02 +0000 (0:00:03.948) 0:03:21.625 ********* 2025-08-29 17:39:01.826309 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826315 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826322 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826328 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.826342 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.826349 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.826355 | orchestrator | 2025-08-29 17:39:01.826362 | orchestrator | TASK [ceph-config : Set_fact _osd_memory_target] ******************************* 2025-08-29 17:39:01.826369 | orchestrator | Friday 29 August 2025 17:31:03 +0000 (0:00:00.810) 0:03:22.435 ********* 2025-08-29 17:39:01.826375 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826382 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826404 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826411 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.826417 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.826424 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.826430 | orchestrator | 2025-08-29 17:39:01.826437 | orchestrator | TASK [ceph-config : Set osd_memory_target to cluster host config] ************** 2025-08-29 17:39:01.826444 | orchestrator | Friday 29 August 2025 17:31:04 +0000 (0:00:01.104) 0:03:23.539 ********* 2025-08-29 17:39:01.826450 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826457 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826463 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826470 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826476 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826483 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826490 | orchestrator | 2025-08-29 17:39:01.826496 | orchestrator | TASK [ceph-config : Render rgw configs] **************************************** 2025-08-29 17:39:01.826503 | orchestrator | Friday 29 August 2025 17:31:04 +0000 (0:00:00.576) 0:03:24.116 ********* 2025-08-29 17:39:01.826509 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826516 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826523 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826529 | orchestrator | ok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.826536 | orchestrator | ok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.826544 | orchestrator | ok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.826550 | orchestrator | 2025-08-29 17:39:01.826562 | orchestrator | TASK [ceph-config : Set config to cluster] ************************************* 2025-08-29 17:39:01.826584 | orchestrator | Friday 29 August 2025 17:31:05 +0000 (0:00:00.991) 0:03:25.107 ********* 2025-08-29 17:39:01.826591 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826598 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826605 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826613 | orchestrator | skipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}])  2025-08-29 17:39:01.826623 | orchestrator | skipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}])  2025-08-29 17:39:01.826632 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826639 | orchestrator | skipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}])  2025-08-29 17:39:01.826646 | orchestrator | skipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}])  2025-08-29 17:39:01.826658 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826665 | orchestrator | skipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}])  2025-08-29 17:39:01.826672 | orchestrator | skipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}])  2025-08-29 17:39:01.826679 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826685 | orchestrator | 2025-08-29 17:39:01.826692 | orchestrator | TASK [ceph-config : Set rgw configs to file] *********************************** 2025-08-29 17:39:01.826699 | orchestrator | Friday 29 August 2025 17:31:06 +0000 (0:00:00.902) 0:03:26.010 ********* 2025-08-29 17:39:01.826705 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826712 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826719 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826725 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826732 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826738 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826745 | orchestrator | 2025-08-29 17:39:01.826752 | orchestrator | TASK [ceph-config : Create ceph conf directory] ******************************** 2025-08-29 17:39:01.826759 | orchestrator | Friday 29 August 2025 17:31:07 +0000 (0:00:00.933) 0:03:26.944 ********* 2025-08-29 17:39:01.826765 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826772 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826778 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826785 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826792 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826798 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826805 | orchestrator | 2025-08-29 17:39:01.826812 | orchestrator | TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface from node "{{ ceph_dashboard_call_item }}"] *** 2025-08-29 17:39:01.826818 | orchestrator | Friday 29 August 2025 17:31:08 +0000 (0:00:00.531) 0:03:27.475 ********* 2025-08-29 17:39:01.826825 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826832 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826838 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826845 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826851 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826858 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826865 | orchestrator | 2025-08-29 17:39:01.826871 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] **** 2025-08-29 17:39:01.826878 | orchestrator | Friday 29 August 2025 17:31:09 +0000 (0:00:00.865) 0:03:28.341 ********* 2025-08-29 17:39:01.826885 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826891 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826898 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826905 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826911 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826918 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826924 | orchestrator | 2025-08-29 17:39:01.826931 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] **** 2025-08-29 17:39:01.826938 | orchestrator | Friday 29 August 2025 17:31:09 +0000 (0:00:00.531) 0:03:28.873 ********* 2025-08-29 17:39:01.826945 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.826955 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.826962 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.826978 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.826985 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.826992 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.826998 | orchestrator | 2025-08-29 17:39:01.827005 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] *************** 2025-08-29 17:39:01.827012 | orchestrator | Friday 29 August 2025 17:31:10 +0000 (0:00:00.908) 0:03:29.781 ********* 2025-08-29 17:39:01.827018 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827025 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.827032 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.827038 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.827045 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.827052 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.827058 | orchestrator | 2025-08-29 17:39:01.827065 | orchestrator | TASK [ceph-facts : Set_fact _interface] **************************************** 2025-08-29 17:39:01.827072 | orchestrator | Friday 29 August 2025 17:31:11 +0000 (0:00:00.772) 0:03:30.554 ********* 2025-08-29 17:39:01.827079 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-3)  2025-08-29 17:39:01.827086 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-4)  2025-08-29 17:39:01.827092 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-5)  2025-08-29 17:39:01.827099 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827106 | orchestrator | 2025-08-29 17:39:01.827112 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ****** 2025-08-29 17:39:01.827119 | orchestrator | Friday 29 August 2025 17:31:12 +0000 (0:00:00.615) 0:03:31.170 ********* 2025-08-29 17:39:01.827126 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-3)  2025-08-29 17:39:01.827132 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-4)  2025-08-29 17:39:01.827139 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-5)  2025-08-29 17:39:01.827146 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827153 | orchestrator | 2025-08-29 17:39:01.827159 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ****** 2025-08-29 17:39:01.827166 | orchestrator | Friday 29 August 2025 17:31:12 +0000 (0:00:00.537) 0:03:31.707 ********* 2025-08-29 17:39:01.827173 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-3)  2025-08-29 17:39:01.827179 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-4)  2025-08-29 17:39:01.827186 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-5)  2025-08-29 17:39:01.827193 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827199 | orchestrator | 2025-08-29 17:39:01.827206 | orchestrator | TASK [ceph-facts : Reset rgw_instances (workaround)] *************************** 2025-08-29 17:39:01.827213 | orchestrator | Friday 29 August 2025 17:31:12 +0000 (0:00:00.363) 0:03:32.071 ********* 2025-08-29 17:39:01.827220 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827226 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.827233 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.827239 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.827246 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.827253 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.827259 | orchestrator | 2025-08-29 17:39:01.827266 | orchestrator | TASK [ceph-facts : Set_fact rgw_instances] ************************************* 2025-08-29 17:39:01.827273 | orchestrator | Friday 29 August 2025 17:31:13 +0000 (0:00:00.546) 0:03:32.618 ********* 2025-08-29 17:39:01.827279 | orchestrator | skipping: [testbed-node-0] => (item=0)  2025-08-29 17:39:01.827286 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827292 | orchestrator | skipping: [testbed-node-1] => (item=0)  2025-08-29 17:39:01.827299 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.827306 | orchestrator | skipping: [testbed-node-2] => (item=0)  2025-08-29 17:39:01.827312 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.827319 | orchestrator | ok: [testbed-node-3] => (item=0) 2025-08-29 17:39:01.827326 | orchestrator | ok: [testbed-node-4] => (item=0) 2025-08-29 17:39:01.827337 | orchestrator | ok: [testbed-node-5] => (item=0) 2025-08-29 17:39:01.827343 | orchestrator | 2025-08-29 17:39:01.827350 | orchestrator | TASK [ceph-config : Generate Ceph file] **************************************** 2025-08-29 17:39:01.827357 | orchestrator | Friday 29 August 2025 17:31:16 +0000 (0:00:02.752) 0:03:35.371 ********* 2025-08-29 17:39:01.827363 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.827370 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.827376 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.827383 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.827437 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.827444 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.827450 | orchestrator | 2025-08-29 17:39:01.827457 | orchestrator | RUNNING HANDLER [ceph-handler : Make tempdir for scripts] ********************** 2025-08-29 17:39:01.827464 | orchestrator | Friday 29 August 2025 17:31:19 +0000 (0:00:03.462) 0:03:38.833 ********* 2025-08-29 17:39:01.827471 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.827477 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.827484 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.827490 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.827497 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.827504 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.827510 | orchestrator | 2025-08-29 17:39:01.827517 | orchestrator | RUNNING HANDLER [ceph-handler : Mons handler] ********************************** 2025-08-29 17:39:01.827524 | orchestrator | Friday 29 August 2025 17:31:21 +0000 (0:00:01.325) 0:03:40.159 ********* 2025-08-29 17:39:01.827530 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.827537 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.827543 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.827550 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.827557 | orchestrator | 2025-08-29 17:39:01.827564 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ******** 2025-08-29 17:39:01.827570 | orchestrator | Friday 29 August 2025 17:31:21 +0000 (0:00:00.962) 0:03:41.121 ********* 2025-08-29 17:39:01.827577 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.827584 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.827590 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.827597 | orchestrator | 2025-08-29 17:39:01.827608 | orchestrator | RUNNING HANDLER [ceph-handler : Copy mon restart script] *********************** 2025-08-29 17:39:01.827620 | orchestrator | Friday 29 August 2025 17:31:22 +0000 (0:00:00.334) 0:03:41.455 ********* 2025-08-29 17:39:01.827627 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.827634 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.827640 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.827647 | orchestrator | 2025-08-29 17:39:01.827654 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ******************** 2025-08-29 17:39:01.827661 | orchestrator | Friday 29 August 2025 17:31:23 +0000 (0:00:01.146) 0:03:42.602 ********* 2025-08-29 17:39:01.827667 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-0)  2025-08-29 17:39:01.827674 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-1)  2025-08-29 17:39:01.827681 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-2)  2025-08-29 17:39:01.827687 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827694 | orchestrator | 2025-08-29 17:39:01.827701 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] ********* 2025-08-29 17:39:01.827707 | orchestrator | Friday 29 August 2025 17:31:24 +0000 (0:00:00.627) 0:03:43.230 ********* 2025-08-29 17:39:01.827714 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.827721 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.827727 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.827734 | orchestrator | 2025-08-29 17:39:01.827741 | orchestrator | RUNNING HANDLER [ceph-handler : Osds handler] ********************************** 2025-08-29 17:39:01.827747 | orchestrator | Friday 29 August 2025 17:31:24 +0000 (0:00:00.431) 0:03:43.661 ********* 2025-08-29 17:39:01.827759 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.827766 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.827772 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.827779 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.827786 | orchestrator | 2025-08-29 17:39:01.827792 | orchestrator | RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] ********************** 2025-08-29 17:39:01.827799 | orchestrator | Friday 29 August 2025 17:31:25 +0000 (0:00:00.704) 0:03:44.365 ********* 2025-08-29 17:39:01.827806 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.827812 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.827819 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.827826 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.827832 | orchestrator | 2025-08-29 17:39:01.827839 | orchestrator | RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ******** 2025-08-29 17:39:01.827846 | orchestrator | Friday 29 August 2025 17:31:25 +0000 (0:00:00.422) 0:03:44.788 ********* 2025-08-29 17:39:01.827852 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.827859 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.827866 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.827872 | orchestrator | 2025-08-29 17:39:01.827878 | orchestrator | RUNNING HANDLER [ceph-handler : Unset noup flag] ******************************* 2025-08-29 17:39:01.827884 | orchestrator | Friday 29 August 2025 17:31:26 +0000 (0:00:00.444) 0:03:45.233 ********* 2025-08-29 17:39:01.827891 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.827897 | orchestrator | 2025-08-29 17:39:01.827903 | orchestrator | RUNNING HANDLER [ceph-handler : Copy osd restart script] *********************** 2025-08-29 17:39:01.827909 | orchestrator | Friday 29 August 2025 17:31:26 +0000 (0:00:00.169) 0:03:45.402 ********* 2025-08-29 17:39:01.827915 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.827922 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.827928 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.827934 | orchestrator | 2025-08-29 17:39:01.827940 | orchestrator | RUNNING HANDLER [ceph-handler : Get pool list] ********************************* 2025-08-29 17:39:01.827946 | orchestrator | Friday 29 August 2025 17:31:26 +0000 (0:00:00.275) 0:03:45.678 ********* 2025-08-29 17:39:01.827953 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.827959 | orchestrator | 2025-08-29 17:39:01.827965 | orchestrator | RUNNING HANDLER [ceph-handler : Get balancer module status] ******************** 2025-08-29 17:39:01.827971 | orchestrator | Friday 29 August 2025 17:31:26 +0000 (0:00:00.179) 0:03:45.857 ********* 2025-08-29 17:39:01.827978 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.827984 | orchestrator | 2025-08-29 17:39:01.827990 | orchestrator | RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] ************** 2025-08-29 17:39:01.827996 | orchestrator | Friday 29 August 2025 17:31:26 +0000 (0:00:00.185) 0:03:46.042 ********* 2025-08-29 17:39:01.828002 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828008 | orchestrator | 2025-08-29 17:39:01.828015 | orchestrator | RUNNING HANDLER [ceph-handler : Disable balancer] ****************************** 2025-08-29 17:39:01.828021 | orchestrator | Friday 29 August 2025 17:31:27 +0000 (0:00:00.110) 0:03:46.153 ********* 2025-08-29 17:39:01.828027 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828033 | orchestrator | 2025-08-29 17:39:01.828039 | orchestrator | RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] ***************** 2025-08-29 17:39:01.828045 | orchestrator | Friday 29 August 2025 17:31:27 +0000 (0:00:00.181) 0:03:46.335 ********* 2025-08-29 17:39:01.828051 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828057 | orchestrator | 2025-08-29 17:39:01.828064 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] ******************* 2025-08-29 17:39:01.828070 | orchestrator | Friday 29 August 2025 17:31:27 +0000 (0:00:00.182) 0:03:46.517 ********* 2025-08-29 17:39:01.828076 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.828087 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.828093 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.828099 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828105 | orchestrator | 2025-08-29 17:39:01.828111 | orchestrator | RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] ********* 2025-08-29 17:39:01.828117 | orchestrator | Friday 29 August 2025 17:31:27 +0000 (0:00:00.549) 0:03:47.066 ********* 2025-08-29 17:39:01.828123 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828130 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.828139 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.828146 | orchestrator | 2025-08-29 17:39:01.828156 | orchestrator | RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] *************** 2025-08-29 17:39:01.828163 | orchestrator | Friday 29 August 2025 17:31:28 +0000 (0:00:00.221) 0:03:47.288 ********* 2025-08-29 17:39:01.828169 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828175 | orchestrator | 2025-08-29 17:39:01.828181 | orchestrator | RUNNING HANDLER [ceph-handler : Re-enable balancer] **************************** 2025-08-29 17:39:01.828187 | orchestrator | Friday 29 August 2025 17:31:28 +0000 (0:00:00.201) 0:03:47.490 ********* 2025-08-29 17:39:01.828194 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828200 | orchestrator | 2025-08-29 17:39:01.828206 | orchestrator | RUNNING HANDLER [ceph-handler : Mdss handler] ********************************** 2025-08-29 17:39:01.828212 | orchestrator | Friday 29 August 2025 17:31:28 +0000 (0:00:00.195) 0:03:47.686 ********* 2025-08-29 17:39:01.828218 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.828224 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.828231 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.828237 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.828243 | orchestrator | 2025-08-29 17:39:01.828249 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ******** 2025-08-29 17:39:01.828256 | orchestrator | Friday 29 August 2025 17:31:29 +0000 (0:00:00.786) 0:03:48.472 ********* 2025-08-29 17:39:01.828262 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.828268 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.828274 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.828280 | orchestrator | 2025-08-29 17:39:01.828287 | orchestrator | RUNNING HANDLER [ceph-handler : Copy mds restart script] *********************** 2025-08-29 17:39:01.828293 | orchestrator | Friday 29 August 2025 17:31:29 +0000 (0:00:00.282) 0:03:48.755 ********* 2025-08-29 17:39:01.828299 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.828306 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.828312 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.828318 | orchestrator | 2025-08-29 17:39:01.828324 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ******************** 2025-08-29 17:39:01.828331 | orchestrator | Friday 29 August 2025 17:31:30 +0000 (0:00:01.230) 0:03:49.985 ********* 2025-08-29 17:39:01.828337 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.828343 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.828349 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.828355 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828362 | orchestrator | 2025-08-29 17:39:01.828368 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] ********* 2025-08-29 17:39:01.828374 | orchestrator | Friday 29 August 2025 17:31:31 +0000 (0:00:00.668) 0:03:50.654 ********* 2025-08-29 17:39:01.828380 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.828397 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.828404 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.828410 | orchestrator | 2025-08-29 17:39:01.828416 | orchestrator | RUNNING HANDLER [ceph-handler : Rgws handler] ********************************** 2025-08-29 17:39:01.828422 | orchestrator | Friday 29 August 2025 17:31:31 +0000 (0:00:00.299) 0:03:50.954 ********* 2025-08-29 17:39:01.828436 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.828443 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.828449 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.828455 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.828461 | orchestrator | 2025-08-29 17:39:01.828468 | orchestrator | RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ******** 2025-08-29 17:39:01.828474 | orchestrator | Friday 29 August 2025 17:31:32 +0000 (0:00:00.912) 0:03:51.867 ********* 2025-08-29 17:39:01.828480 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.828486 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.828492 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.828499 | orchestrator | 2025-08-29 17:39:01.828505 | orchestrator | RUNNING HANDLER [ceph-handler : Copy rgw restart script] *********************** 2025-08-29 17:39:01.828511 | orchestrator | Friday 29 August 2025 17:31:33 +0000 (0:00:00.285) 0:03:52.153 ********* 2025-08-29 17:39:01.828517 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.828523 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.828529 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.828536 | orchestrator | 2025-08-29 17:39:01.828542 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ******************** 2025-08-29 17:39:01.828548 | orchestrator | Friday 29 August 2025 17:31:34 +0000 (0:00:01.442) 0:03:53.595 ********* 2025-08-29 17:39:01.828554 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.828560 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.828566 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.828573 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828579 | orchestrator | 2025-08-29 17:39:01.828585 | orchestrator | RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] ********* 2025-08-29 17:39:01.828591 | orchestrator | Friday 29 August 2025 17:31:35 +0000 (0:00:00.633) 0:03:54.229 ********* 2025-08-29 17:39:01.828597 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.828603 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.828610 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.828616 | orchestrator | 2025-08-29 17:39:01.828622 | orchestrator | RUNNING HANDLER [ceph-handler : Rbdmirrors handler] **************************** 2025-08-29 17:39:01.828628 | orchestrator | Friday 29 August 2025 17:31:35 +0000 (0:00:00.330) 0:03:54.560 ********* 2025-08-29 17:39:01.828634 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.828641 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.828647 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.828653 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.828659 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.828665 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.828671 | orchestrator | 2025-08-29 17:39:01.828677 | orchestrator | RUNNING HANDLER [ceph-handler : Mgrs handler] ********************************** 2025-08-29 17:39:01.828687 | orchestrator | Friday 29 August 2025 17:31:36 +0000 (0:00:00.754) 0:03:55.314 ********* 2025-08-29 17:39:01.828698 | orchestrator | skipping: [te2025-08-29 17:39:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:01.828704 | orchestrator | stbed-node-3] 2025-08-29 17:39:01.828711 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.828717 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.828723 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.828729 | orchestrator | 2025-08-29 17:39:01.828736 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ******** 2025-08-29 17:39:01.828742 | orchestrator | Friday 29 August 2025 17:31:37 +0000 (0:00:00.911) 0:03:56.226 ********* 2025-08-29 17:39:01.828748 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.828755 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.828761 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.828770 | orchestrator | 2025-08-29 17:39:01.828777 | orchestrator | RUNNING HANDLER [ceph-handler : Copy mgr restart script] *********************** 2025-08-29 17:39:01.828783 | orchestrator | Friday 29 August 2025 17:31:37 +0000 (0:00:00.373) 0:03:56.599 ********* 2025-08-29 17:39:01.828789 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.828796 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.828802 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.828808 | orchestrator | 2025-08-29 17:39:01.828814 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ******************** 2025-08-29 17:39:01.828820 | orchestrator | Friday 29 August 2025 17:31:38 +0000 (0:00:01.189) 0:03:57.788 ********* 2025-08-29 17:39:01.828827 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-0)  2025-08-29 17:39:01.828833 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-1)  2025-08-29 17:39:01.828839 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-2)  2025-08-29 17:39:01.828845 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.828852 | orchestrator | 2025-08-29 17:39:01.828858 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] ********* 2025-08-29 17:39:01.828865 | orchestrator | Friday 29 August 2025 17:31:39 +0000 (0:00:00.471) 0:03:58.260 ********* 2025-08-29 17:39:01.828871 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.828877 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.828883 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.828890 | orchestrator | 2025-08-29 17:39:01.828896 | orchestrator | PLAY [Apply role ceph-mon] ***************************************************** 2025-08-29 17:39:01.828902 | orchestrator | 2025-08-29 17:39:01.828909 | orchestrator | TASK [ceph-handler : Include check_running_cluster.yml] ************************ 2025-08-29 17:39:01.828915 | orchestrator | Friday 29 August 2025 17:31:39 +0000 (0:00:00.390) 0:03:58.651 ********* 2025-08-29 17:39:01.828921 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.828927 | orchestrator | 2025-08-29 17:39:01.828934 | orchestrator | TASK [ceph-handler : Include check_running_containers.yml] ********************* 2025-08-29 17:39:01.828940 | orchestrator | Friday 29 August 2025 17:31:40 +0000 (0:00:00.510) 0:03:59.161 ********* 2025-08-29 17:39:01.828946 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.828953 | orchestrator | 2025-08-29 17:39:01.828959 | orchestrator | TASK [ceph-handler : Check for a mon container] ******************************** 2025-08-29 17:39:01.828965 | orchestrator | Friday 29 August 2025 17:31:40 +0000 (0:00:00.406) 0:03:59.567 ********* 2025-08-29 17:39:01.828971 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.828977 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.828983 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.828990 | orchestrator | 2025-08-29 17:39:01.828996 | orchestrator | TASK [ceph-handler : Check for an osd container] ******************************* 2025-08-29 17:39:01.829002 | orchestrator | Friday 29 August 2025 17:31:41 +0000 (0:00:00.794) 0:04:00.361 ********* 2025-08-29 17:39:01.829008 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829015 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829021 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829027 | orchestrator | 2025-08-29 17:39:01.829033 | orchestrator | TASK [ceph-handler : Check for a mds container] ******************************** 2025-08-29 17:39:01.829039 | orchestrator | Friday 29 August 2025 17:31:41 +0000 (0:00:00.225) 0:04:00.587 ********* 2025-08-29 17:39:01.829045 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829051 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829058 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829064 | orchestrator | 2025-08-29 17:39:01.829070 | orchestrator | TASK [ceph-handler : Check for a rgw container] ******************************** 2025-08-29 17:39:01.829076 | orchestrator | Friday 29 August 2025 17:31:41 +0000 (0:00:00.257) 0:04:00.845 ********* 2025-08-29 17:39:01.829082 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829092 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829099 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829105 | orchestrator | 2025-08-29 17:39:01.829111 | orchestrator | TASK [ceph-handler : Check for a mgr container] ******************************** 2025-08-29 17:39:01.829117 | orchestrator | Friday 29 August 2025 17:31:41 +0000 (0:00:00.271) 0:04:01.116 ********* 2025-08-29 17:39:01.829123 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829130 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829136 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829142 | orchestrator | 2025-08-29 17:39:01.829148 | orchestrator | TASK [ceph-handler : Check for a rbd mirror container] ************************* 2025-08-29 17:39:01.829154 | orchestrator | Friday 29 August 2025 17:31:42 +0000 (0:00:00.769) 0:04:01.886 ********* 2025-08-29 17:39:01.829161 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829167 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829173 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829179 | orchestrator | 2025-08-29 17:39:01.829185 | orchestrator | TASK [ceph-handler : Check for a nfs container] ******************************** 2025-08-29 17:39:01.829192 | orchestrator | Friday 29 August 2025 17:31:42 +0000 (0:00:00.230) 0:04:02.116 ********* 2025-08-29 17:39:01.829198 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829209 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829215 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829221 | orchestrator | 2025-08-29 17:39:01.829231 | orchestrator | TASK [ceph-handler : Check for a ceph-crash container] ************************* 2025-08-29 17:39:01.829238 | orchestrator | Friday 29 August 2025 17:31:43 +0000 (0:00:00.263) 0:04:02.380 ********* 2025-08-29 17:39:01.829244 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829250 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829257 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829263 | orchestrator | 2025-08-29 17:39:01.829269 | orchestrator | TASK [ceph-handler : Check for a ceph-exporter container] ********************** 2025-08-29 17:39:01.829275 | orchestrator | Friday 29 August 2025 17:31:43 +0000 (0:00:00.636) 0:04:03.016 ********* 2025-08-29 17:39:01.829281 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829288 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829294 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829300 | orchestrator | 2025-08-29 17:39:01.829306 | orchestrator | TASK [ceph-handler : Include check_socket_non_container.yml] ******************* 2025-08-29 17:39:01.829312 | orchestrator | Friday 29 August 2025 17:31:44 +0000 (0:00:00.888) 0:04:03.905 ********* 2025-08-29 17:39:01.829319 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829325 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829331 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829337 | orchestrator | 2025-08-29 17:39:01.829343 | orchestrator | TASK [ceph-handler : Set_fact handler_mon_status] ****************************** 2025-08-29 17:39:01.829349 | orchestrator | Friday 29 August 2025 17:31:45 +0000 (0:00:00.279) 0:04:04.184 ********* 2025-08-29 17:39:01.829356 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829362 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829368 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829374 | orchestrator | 2025-08-29 17:39:01.829380 | orchestrator | TASK [ceph-handler : Set_fact handler_osd_status] ****************************** 2025-08-29 17:39:01.829398 | orchestrator | Friday 29 August 2025 17:31:45 +0000 (0:00:00.264) 0:04:04.448 ********* 2025-08-29 17:39:01.829405 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829411 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829418 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829424 | orchestrator | 2025-08-29 17:39:01.829430 | orchestrator | TASK [ceph-handler : Set_fact handler_mds_status] ****************************** 2025-08-29 17:39:01.829437 | orchestrator | Friday 29 August 2025 17:31:45 +0000 (0:00:00.260) 0:04:04.709 ********* 2025-08-29 17:39:01.829443 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829449 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829460 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829466 | orchestrator | 2025-08-29 17:39:01.829473 | orchestrator | TASK [ceph-handler : Set_fact handler_rgw_status] ****************************** 2025-08-29 17:39:01.829479 | orchestrator | Friday 29 August 2025 17:31:46 +0000 (0:00:00.484) 0:04:05.193 ********* 2025-08-29 17:39:01.829485 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829491 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829497 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829503 | orchestrator | 2025-08-29 17:39:01.829509 | orchestrator | TASK [ceph-handler : Set_fact handler_nfs_status] ****************************** 2025-08-29 17:39:01.829516 | orchestrator | Friday 29 August 2025 17:31:46 +0000 (0:00:00.305) 0:04:05.499 ********* 2025-08-29 17:39:01.829522 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829528 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829534 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829540 | orchestrator | 2025-08-29 17:39:01.829547 | orchestrator | TASK [ceph-handler : Set_fact handler_rbd_status] ****************************** 2025-08-29 17:39:01.829553 | orchestrator | Friday 29 August 2025 17:31:46 +0000 (0:00:00.301) 0:04:05.801 ********* 2025-08-29 17:39:01.829559 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829565 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.829571 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.829577 | orchestrator | 2025-08-29 17:39:01.829584 | orchestrator | TASK [ceph-handler : Set_fact handler_mgr_status] ****************************** 2025-08-29 17:39:01.829590 | orchestrator | Friday 29 August 2025 17:31:47 +0000 (0:00:00.345) 0:04:06.146 ********* 2025-08-29 17:39:01.829596 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829602 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829608 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829614 | orchestrator | 2025-08-29 17:39:01.829621 | orchestrator | TASK [ceph-handler : Set_fact handler_crash_status] **************************** 2025-08-29 17:39:01.829627 | orchestrator | Friday 29 August 2025 17:31:47 +0000 (0:00:00.534) 0:04:06.681 ********* 2025-08-29 17:39:01.829633 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829639 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829645 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829652 | orchestrator | 2025-08-29 17:39:01.829658 | orchestrator | TASK [ceph-handler : Set_fact handler_exporter_status] ************************* 2025-08-29 17:39:01.829664 | orchestrator | Friday 29 August 2025 17:31:48 +0000 (0:00:00.537) 0:04:07.218 ********* 2025-08-29 17:39:01.829670 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829676 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829682 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829688 | orchestrator | 2025-08-29 17:39:01.829695 | orchestrator | TASK [ceph-mon : Set_fact container_exec_cmd] ********************************** 2025-08-29 17:39:01.829701 | orchestrator | Friday 29 August 2025 17:31:48 +0000 (0:00:00.427) 0:04:07.646 ********* 2025-08-29 17:39:01.829707 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829713 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829719 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829725 | orchestrator | 2025-08-29 17:39:01.829731 | orchestrator | TASK [ceph-mon : Include deploy_monitors.yml] ********************************** 2025-08-29 17:39:01.829738 | orchestrator | Friday 29 August 2025 17:31:48 +0000 (0:00:00.271) 0:04:07.917 ********* 2025-08-29 17:39:01.829744 | orchestrator | included: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-2, testbed-node-1 2025-08-29 17:39:01.829750 | orchestrator | 2025-08-29 17:39:01.829756 | orchestrator | TASK [ceph-mon : Check if monitor initial keyring already exists] ************** 2025-08-29 17:39:01.829763 | orchestrator | Friday 29 August 2025 17:31:49 +0000 (0:00:00.664) 0:04:08.582 ********* 2025-08-29 17:39:01.829769 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.829775 | orchestrator | 2025-08-29 17:39:01.829785 | orchestrator | TASK [ceph-mon : Generate monitor initial keyring] ***************************** 2025-08-29 17:39:01.829795 | orchestrator | Friday 29 August 2025 17:31:49 +0000 (0:00:00.109) 0:04:08.691 ********* 2025-08-29 17:39:01.829806 | orchestrator | changed: [testbed-node-0 -> localhost] 2025-08-29 17:39:01.829812 | orchestrator | 2025-08-29 17:39:01.829818 | orchestrator | TASK [ceph-mon : Set_fact _initial_mon_key_success] **************************** 2025-08-29 17:39:01.829824 | orchestrator | Friday 29 August 2025 17:31:50 +0000 (0:00:00.895) 0:04:09.587 ********* 2025-08-29 17:39:01.829830 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829837 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829843 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829849 | orchestrator | 2025-08-29 17:39:01.829855 | orchestrator | TASK [ceph-mon : Get initial keyring when it already exists] ******************* 2025-08-29 17:39:01.829862 | orchestrator | Friday 29 August 2025 17:31:50 +0000 (0:00:00.276) 0:04:09.863 ********* 2025-08-29 17:39:01.829868 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.829874 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.829880 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.829886 | orchestrator | 2025-08-29 17:39:01.829892 | orchestrator | TASK [ceph-mon : Create monitor initial keyring] ******************************* 2025-08-29 17:39:01.829899 | orchestrator | Friday 29 August 2025 17:31:51 +0000 (0:00:00.408) 0:04:10.271 ********* 2025-08-29 17:39:01.829905 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.829911 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.829917 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.829923 | orchestrator | 2025-08-29 17:39:01.829930 | orchestrator | TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] *********** 2025-08-29 17:39:01.829936 | orchestrator | Friday 29 August 2025 17:31:52 +0000 (0:00:01.205) 0:04:11.476 ********* 2025-08-29 17:39:01.829942 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.829948 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.829954 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.829961 | orchestrator | 2025-08-29 17:39:01.829967 | orchestrator | TASK [ceph-mon : Create monitor directory] ************************************* 2025-08-29 17:39:01.829973 | orchestrator | Friday 29 August 2025 17:31:53 +0000 (0:00:00.687) 0:04:12.164 ********* 2025-08-29 17:39:01.829979 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.829985 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.829991 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.829997 | orchestrator | 2025-08-29 17:39:01.830004 | orchestrator | TASK [ceph-mon : Recursively fix ownership of monitor directory] *************** 2025-08-29 17:39:01.830010 | orchestrator | Friday 29 August 2025 17:31:53 +0000 (0:00:00.708) 0:04:12.872 ********* 2025-08-29 17:39:01.830081 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.830090 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.830097 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.830103 | orchestrator | 2025-08-29 17:39:01.830109 | orchestrator | TASK [ceph-mon : Create admin keyring] ***************************************** 2025-08-29 17:39:01.830115 | orchestrator | Friday 29 August 2025 17:31:54 +0000 (0:00:00.787) 0:04:13.660 ********* 2025-08-29 17:39:01.830122 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830128 | orchestrator | 2025-08-29 17:39:01.830134 | orchestrator | TASK [ceph-mon : Slurp admin keyring] ****************************************** 2025-08-29 17:39:01.830141 | orchestrator | Friday 29 August 2025 17:31:55 +0000 (0:00:01.162) 0:04:14.822 ********* 2025-08-29 17:39:01.830147 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.830153 | orchestrator | 2025-08-29 17:39:01.830159 | orchestrator | TASK [ceph-mon : Copy admin keyring over to mons] ****************************** 2025-08-29 17:39:01.830165 | orchestrator | Friday 29 August 2025 17:31:56 +0000 (0:00:00.641) 0:04:15.464 ********* 2025-08-29 17:39:01.830172 | orchestrator | changed: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.830178 | orchestrator | ok: [testbed-node-0] => (item=None) 2025-08-29 17:39:01.830184 | orchestrator | ok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.830190 | orchestrator | changed: [testbed-node-1] => (item=None) 2025-08-29 17:39:01.830196 | orchestrator | ok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:39:01.830207 | orchestrator | ok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:39:01.830214 | orchestrator | changed: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:39:01.830220 | orchestrator | changed: [testbed-node-1 -> {{ item }}] 2025-08-29 17:39:01.830226 | orchestrator | ok: [testbed-node-2] => (item=None) 2025-08-29 17:39:01.830232 | orchestrator | ok: [testbed-node-2 -> {{ item }}] 2025-08-29 17:39:01.830239 | orchestrator | ok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:39:01.830245 | orchestrator | ok: [testbed-node-0 -> {{ item }}] 2025-08-29 17:39:01.830251 | orchestrator | 2025-08-29 17:39:01.830257 | orchestrator | TASK [ceph-mon : Import admin keyring into mon keyring] ************************ 2025-08-29 17:39:01.830263 | orchestrator | Friday 29 August 2025 17:31:59 +0000 (0:00:02.912) 0:04:18.376 ********* 2025-08-29 17:39:01.830270 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830276 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.830282 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.830288 | orchestrator | 2025-08-29 17:39:01.830294 | orchestrator | TASK [ceph-mon : Set_fact ceph-mon container command] ************************** 2025-08-29 17:39:01.830300 | orchestrator | Friday 29 August 2025 17:32:00 +0000 (0:00:01.205) 0:04:19.582 ********* 2025-08-29 17:39:01.830307 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.830313 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.830319 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.830325 | orchestrator | 2025-08-29 17:39:01.830332 | orchestrator | TASK [ceph-mon : Set_fact monmaptool container command] ************************ 2025-08-29 17:39:01.830338 | orchestrator | Friday 29 August 2025 17:32:00 +0000 (0:00:00.512) 0:04:20.095 ********* 2025-08-29 17:39:01.830344 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.830350 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.830356 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.830363 | orchestrator | 2025-08-29 17:39:01.830369 | orchestrator | TASK [ceph-mon : Generate initial monmap] ************************************** 2025-08-29 17:39:01.830375 | orchestrator | Friday 29 August 2025 17:32:01 +0000 (0:00:00.315) 0:04:20.411 ********* 2025-08-29 17:39:01.830381 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830401 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.830408 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.830414 | orchestrator | 2025-08-29 17:39:01.830439 | orchestrator | TASK [ceph-mon : Ceph monitor mkfs with keyring] ******************************* 2025-08-29 17:39:01.830447 | orchestrator | Friday 29 August 2025 17:32:02 +0000 (0:00:01.523) 0:04:21.935 ********* 2025-08-29 17:39:01.830453 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830459 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.830465 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.830471 | orchestrator | 2025-08-29 17:39:01.830477 | orchestrator | TASK [ceph-mon : Ceph monitor mkfs without keyring] **************************** 2025-08-29 17:39:01.830483 | orchestrator | Friday 29 August 2025 17:32:04 +0000 (0:00:01.390) 0:04:23.325 ********* 2025-08-29 17:39:01.830490 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.830496 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.830502 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.830508 | orchestrator | 2025-08-29 17:39:01.830514 | orchestrator | TASK [ceph-mon : Include start_monitor.yml] ************************************ 2025-08-29 17:39:01.830520 | orchestrator | Friday 29 August 2025 17:32:04 +0000 (0:00:00.294) 0:04:23.619 ********* 2025-08-29 17:39:01.830526 | orchestrator | included: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.830533 | orchestrator | 2025-08-29 17:39:01.830539 | orchestrator | TASK [ceph-mon : Ensure systemd service override directory exists] ************* 2025-08-29 17:39:01.830545 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.772) 0:04:24.392 ********* 2025-08-29 17:39:01.830551 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.830558 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.830568 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.830574 | orchestrator | 2025-08-29 17:39:01.830581 | orchestrator | TASK [ceph-mon : Add ceph-mon systemd service overrides] *********************** 2025-08-29 17:39:01.830587 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.294) 0:04:24.686 ********* 2025-08-29 17:39:01.830593 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.830599 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.830605 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.830611 | orchestrator | 2025-08-29 17:39:01.830618 | orchestrator | TASK [ceph-mon : Include_tasks systemd.yml] ************************************ 2025-08-29 17:39:01.830624 | orchestrator | Friday 29 August 2025 17:32:05 +0000 (0:00:00.270) 0:04:24.957 ********* 2025-08-29 17:39:01.830630 | orchestrator | included: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.830636 | orchestrator | 2025-08-29 17:39:01.830642 | orchestrator | TASK [ceph-mon : Generate systemd unit file for mon container] ***************** 2025-08-29 17:39:01.830649 | orchestrator | Friday 29 August 2025 17:32:06 +0000 (0:00:00.826) 0:04:25.784 ********* 2025-08-29 17:39:01.830655 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.830661 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.830667 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830673 | orchestrator | 2025-08-29 17:39:01.830679 | orchestrator | TASK [ceph-mon : Generate systemd ceph-mon target file] ************************ 2025-08-29 17:39:01.830685 | orchestrator | Friday 29 August 2025 17:32:08 +0000 (0:00:01.792) 0:04:27.576 ********* 2025-08-29 17:39:01.830719 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830726 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.830732 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.830738 | orchestrator | 2025-08-29 17:39:01.830745 | orchestrator | TASK [ceph-mon : Enable ceph-mon.target] *************************************** 2025-08-29 17:39:01.830751 | orchestrator | Friday 29 August 2025 17:32:09 +0000 (0:00:01.314) 0:04:28.890 ********* 2025-08-29 17:39:01.830757 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830763 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.830769 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.830776 | orchestrator | 2025-08-29 17:39:01.830782 | orchestrator | TASK [ceph-mon : Start the monitor service] ************************************ 2025-08-29 17:39:01.830788 | orchestrator | Friday 29 August 2025 17:32:11 +0000 (0:00:02.109) 0:04:31.000 ********* 2025-08-29 17:39:01.830794 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.830800 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.830807 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.830813 | orchestrator | 2025-08-29 17:39:01.830819 | orchestrator | TASK [ceph-mon : Include_tasks ceph_keys.yml] ********************************** 2025-08-29 17:39:01.830825 | orchestrator | Friday 29 August 2025 17:32:13 +0000 (0:00:02.026) 0:04:33.027 ********* 2025-08-29 17:39:01.830832 | orchestrator | included: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.830838 | orchestrator | 2025-08-29 17:39:01.830844 | orchestrator | TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] ************* 2025-08-29 17:39:01.830850 | orchestrator | Friday 29 August 2025 17:32:14 +0000 (0:00:00.517) 0:04:33.544 ********* 2025-08-29 17:39:01.830856 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left). 2025-08-29 17:39:01.830864 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.830874 | orchestrator | 2025-08-29 17:39:01.830885 | orchestrator | TASK [ceph-mon : Fetch ceph initial keys] ************************************** 2025-08-29 17:39:01.830895 | orchestrator | Friday 29 August 2025 17:32:36 +0000 (0:00:22.144) 0:04:55.688 ********* 2025-08-29 17:39:01.830905 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.830914 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.830924 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.830934 | orchestrator | 2025-08-29 17:39:01.830945 | orchestrator | TASK [ceph-mon : Include secure_cluster.yml] *********************************** 2025-08-29 17:39:01.830965 | orchestrator | Friday 29 August 2025 17:32:46 +0000 (0:00:10.389) 0:05:06.077 ********* 2025-08-29 17:39:01.830975 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.830984 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.830991 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.830997 | orchestrator | 2025-08-29 17:39:01.831003 | orchestrator | TASK [ceph-mon : Set cluster configs] ****************************************** 2025-08-29 17:39:01.831009 | orchestrator | Friday 29 August 2025 17:32:47 +0000 (0:00:00.272) 0:05:06.350 ********* 2025-08-29 17:39:01.831042 | orchestrator | changed: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__65d9b83f8cbfd46afac26ddfb64b73af1711ff8c'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}]) 2025-08-29 17:39:01.831052 | orchestrator | changed: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__65d9b83f8cbfd46afac26ddfb64b73af1711ff8c'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}]) 2025-08-29 17:39:01.831060 | orchestrator | changed: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__65d9b83f8cbfd46afac26ddfb64b73af1711ff8c'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}]) 2025-08-29 17:39:01.831067 | orchestrator | changed: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__65d9b83f8cbfd46afac26ddfb64b73af1711ff8c'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}]) 2025-08-29 17:39:01.831075 | orchestrator | changed: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__65d9b83f8cbfd46afac26ddfb64b73af1711ff8c'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}]) 2025-08-29 17:39:01.831082 | orchestrator | skipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__65d9b83f8cbfd46afac26ddfb64b73af1711ff8c'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__65d9b83f8cbfd46afac26ddfb64b73af1711ff8c'}])  2025-08-29 17:39:01.831090 | orchestrator | 2025-08-29 17:39:01.831096 | orchestrator | RUNNING HANDLER [ceph-handler : Make tempdir for scripts] ********************** 2025-08-29 17:39:01.831102 | orchestrator | Friday 29 August 2025 17:33:02 +0000 (0:00:15.078) 0:05:21.429 ********* 2025-08-29 17:39:01.831108 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831114 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831121 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831127 | orchestrator | 2025-08-29 17:39:01.831133 | orchestrator | RUNNING HANDLER [ceph-handler : Mons handler] ********************************** 2025-08-29 17:39:01.831139 | orchestrator | Friday 29 August 2025 17:33:02 +0000 (0:00:00.338) 0:05:21.768 ********* 2025-08-29 17:39:01.831145 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.831152 | orchestrator | 2025-08-29 17:39:01.831158 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ******** 2025-08-29 17:39:01.831169 | orchestrator | Friday 29 August 2025 17:33:03 +0000 (0:00:00.503) 0:05:22.271 ********* 2025-08-29 17:39:01.831176 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.831182 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.831188 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.831194 | orchestrator | 2025-08-29 17:39:01.831200 | orchestrator | RUNNING HANDLER [ceph-handler : Copy mon restart script] *********************** 2025-08-29 17:39:01.831206 | orchestrator | Friday 29 August 2025 17:33:03 +0000 (0:00:00.486) 0:05:22.757 ********* 2025-08-29 17:39:01.831213 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831219 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831225 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831231 | orchestrator | 2025-08-29 17:39:01.831237 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ******************** 2025-08-29 17:39:01.831244 | orchestrator | Friday 29 August 2025 17:33:03 +0000 (0:00:00.327) 0:05:23.084 ********* 2025-08-29 17:39:01.831250 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-0)  2025-08-29 17:39:01.831256 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-1)  2025-08-29 17:39:01.831262 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-2)  2025-08-29 17:39:01.831268 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831274 | orchestrator | 2025-08-29 17:39:01.831281 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] ********* 2025-08-29 17:39:01.831287 | orchestrator | Friday 29 August 2025 17:33:04 +0000 (0:00:00.661) 0:05:23.745 ********* 2025-08-29 17:39:01.831293 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.831299 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.831309 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.831320 | orchestrator | 2025-08-29 17:39:01.831354 | orchestrator | PLAY [Apply role ceph-mgr] ***************************************************** 2025-08-29 17:39:01.831364 | orchestrator | 2025-08-29 17:39:01.831373 | orchestrator | TASK [ceph-handler : Include check_running_cluster.yml] ************************ 2025-08-29 17:39:01.831383 | orchestrator | Friday 29 August 2025 17:33:05 +0000 (0:00:00.589) 0:05:24.335 ********* 2025-08-29 17:39:01.831524 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.831537 | orchestrator | 2025-08-29 17:39:01.831544 | orchestrator | TASK [ceph-handler : Include check_running_containers.yml] ********************* 2025-08-29 17:39:01.831550 | orchestrator | Friday 29 August 2025 17:33:05 +0000 (0:00:00.802) 0:05:25.137 ********* 2025-08-29 17:39:01.831556 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.831562 | orchestrator | 2025-08-29 17:39:01.831568 | orchestrator | TASK [ceph-handler : Check for a mon container] ******************************** 2025-08-29 17:39:01.831574 | orchestrator | Friday 29 August 2025 17:33:06 +0000 (0:00:00.522) 0:05:25.660 ********* 2025-08-29 17:39:01.831580 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.831587 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.831593 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.831599 | orchestrator | 2025-08-29 17:39:01.831605 | orchestrator | TASK [ceph-handler : Check for an osd container] ******************************* 2025-08-29 17:39:01.831611 | orchestrator | Friday 29 August 2025 17:33:07 +0000 (0:00:01.045) 0:05:26.705 ********* 2025-08-29 17:39:01.831617 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831624 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831630 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831636 | orchestrator | 2025-08-29 17:39:01.831642 | orchestrator | TASK [ceph-handler : Check for a mds container] ******************************** 2025-08-29 17:39:01.831648 | orchestrator | Friday 29 August 2025 17:33:07 +0000 (0:00:00.308) 0:05:27.014 ********* 2025-08-29 17:39:01.831654 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831660 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831666 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831684 | orchestrator | 2025-08-29 17:39:01.831691 | orchestrator | TASK [ceph-handler : Check for a rgw container] ******************************** 2025-08-29 17:39:01.831697 | orchestrator | Friday 29 August 2025 17:33:08 +0000 (0:00:00.260) 0:05:27.274 ********* 2025-08-29 17:39:01.831703 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831709 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831715 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831721 | orchestrator | 2025-08-29 17:39:01.831727 | orchestrator | TASK [ceph-handler : Check for a mgr container] ******************************** 2025-08-29 17:39:01.831733 | orchestrator | Friday 29 August 2025 17:33:08 +0000 (0:00:00.250) 0:05:27.525 ********* 2025-08-29 17:39:01.831740 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.831746 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.831752 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.831758 | orchestrator | 2025-08-29 17:39:01.831764 | orchestrator | TASK [ceph-handler : Check for a rbd mirror container] ************************* 2025-08-29 17:39:01.831770 | orchestrator | Friday 29 August 2025 17:33:09 +0000 (0:00:00.814) 0:05:28.339 ********* 2025-08-29 17:39:01.831776 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831782 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831788 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831795 | orchestrator | 2025-08-29 17:39:01.831801 | orchestrator | TASK [ceph-handler : Check for a nfs container] ******************************** 2025-08-29 17:39:01.831807 | orchestrator | Friday 29 August 2025 17:33:09 +0000 (0:00:00.251) 0:05:28.590 ********* 2025-08-29 17:39:01.831813 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831819 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831825 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831831 | orchestrator | 2025-08-29 17:39:01.831837 | orchestrator | TASK [ceph-handler : Check for a ceph-crash container] ************************* 2025-08-29 17:39:01.831843 | orchestrator | Friday 29 August 2025 17:33:09 +0000 (0:00:00.298) 0:05:28.889 ********* 2025-08-29 17:39:01.831849 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.831855 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.831861 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.831867 | orchestrator | 2025-08-29 17:39:01.831873 | orchestrator | TASK [ceph-handler : Check for a ceph-exporter container] ********************** 2025-08-29 17:39:01.831880 | orchestrator | Friday 29 August 2025 17:33:10 +0000 (0:00:00.676) 0:05:29.565 ********* 2025-08-29 17:39:01.831886 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.831892 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.831898 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.831903 | orchestrator | 2025-08-29 17:39:01.831908 | orchestrator | TASK [ceph-handler : Include check_socket_non_container.yml] ******************* 2025-08-29 17:39:01.831914 | orchestrator | Friday 29 August 2025 17:33:11 +0000 (0:00:00.928) 0:05:30.493 ********* 2025-08-29 17:39:01.831919 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831924 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831930 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831935 | orchestrator | 2025-08-29 17:39:01.831940 | orchestrator | TASK [ceph-handler : Set_fact handler_mon_status] ****************************** 2025-08-29 17:39:01.831946 | orchestrator | Friday 29 August 2025 17:33:11 +0000 (0:00:00.297) 0:05:30.790 ********* 2025-08-29 17:39:01.831951 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.831956 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.831962 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.831967 | orchestrator | 2025-08-29 17:39:01.831972 | orchestrator | TASK [ceph-handler : Set_fact handler_osd_status] ****************************** 2025-08-29 17:39:01.831978 | orchestrator | Friday 29 August 2025 17:33:11 +0000 (0:00:00.307) 0:05:31.097 ********* 2025-08-29 17:39:01.831983 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.831989 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.831994 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.831999 | orchestrator | 2025-08-29 17:39:01.832004 | orchestrator | TASK [ceph-handler : Set_fact handler_mds_status] ****************************** 2025-08-29 17:39:01.832015 | orchestrator | Friday 29 August 2025 17:33:12 +0000 (0:00:00.322) 0:05:31.420 ********* 2025-08-29 17:39:01.832024 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832030 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832064 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832070 | orchestrator | 2025-08-29 17:39:01.832076 | orchestrator | TASK [ceph-handler : Set_fact handler_rgw_status] ****************************** 2025-08-29 17:39:01.832081 | orchestrator | Friday 29 August 2025 17:33:12 +0000 (0:00:00.399) 0:05:31.820 ********* 2025-08-29 17:39:01.832087 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832092 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832098 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832103 | orchestrator | 2025-08-29 17:39:01.832108 | orchestrator | TASK [ceph-handler : Set_fact handler_nfs_status] ****************************** 2025-08-29 17:39:01.832114 | orchestrator | Friday 29 August 2025 17:33:12 +0000 (0:00:00.271) 0:05:32.091 ********* 2025-08-29 17:39:01.832119 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832125 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832130 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832135 | orchestrator | 2025-08-29 17:39:01.832141 | orchestrator | TASK [ceph-handler : Set_fact handler_rbd_status] ****************************** 2025-08-29 17:39:01.832146 | orchestrator | Friday 29 August 2025 17:33:13 +0000 (0:00:00.243) 0:05:32.334 ********* 2025-08-29 17:39:01.832152 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832157 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832163 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832168 | orchestrator | 2025-08-29 17:39:01.832173 | orchestrator | TASK [ceph-handler : Set_fact handler_mgr_status] ****************************** 2025-08-29 17:39:01.832179 | orchestrator | Friday 29 August 2025 17:33:13 +0000 (0:00:00.275) 0:05:32.610 ********* 2025-08-29 17:39:01.832184 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.832190 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.832195 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.832201 | orchestrator | 2025-08-29 17:39:01.832206 | orchestrator | TASK [ceph-handler : Set_fact handler_crash_status] **************************** 2025-08-29 17:39:01.832211 | orchestrator | Friday 29 August 2025 17:33:14 +0000 (0:00:00.534) 0:05:33.144 ********* 2025-08-29 17:39:01.832217 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.832222 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.832228 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.832233 | orchestrator | 2025-08-29 17:39:01.832238 | orchestrator | TASK [ceph-handler : Set_fact handler_exporter_status] ************************* 2025-08-29 17:39:01.832244 | orchestrator | Friday 29 August 2025 17:33:14 +0000 (0:00:00.338) 0:05:33.483 ********* 2025-08-29 17:39:01.832249 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.832254 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.832260 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.832265 | orchestrator | 2025-08-29 17:39:01.832270 | orchestrator | TASK [ceph-mgr : Set_fact container_exec_cmd] ********************************** 2025-08-29 17:39:01.832276 | orchestrator | Friday 29 August 2025 17:33:14 +0000 (0:00:00.575) 0:05:34.058 ********* 2025-08-29 17:39:01.832281 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:39:01.832287 | orchestrator | ok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:39:01.832292 | orchestrator | ok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:39:01.832298 | orchestrator | 2025-08-29 17:39:01.832303 | orchestrator | TASK [ceph-mgr : Include common.yml] ******************************************* 2025-08-29 17:39:01.832308 | orchestrator | Friday 29 August 2025 17:33:15 +0000 (0:00:00.836) 0:05:34.894 ********* 2025-08-29 17:39:01.832314 | orchestrator | included: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.832319 | orchestrator | 2025-08-29 17:39:01.832325 | orchestrator | TASK [ceph-mgr : Create mgr directory] ***************************************** 2025-08-29 17:39:01.832334 | orchestrator | Friday 29 August 2025 17:33:16 +0000 (0:00:00.788) 0:05:35.683 ********* 2025-08-29 17:39:01.832340 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.832345 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.832350 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.832356 | orchestrator | 2025-08-29 17:39:01.832372 | orchestrator | TASK [ceph-mgr : Fetch ceph mgr keyring] *************************************** 2025-08-29 17:39:01.832377 | orchestrator | Friday 29 August 2025 17:33:17 +0000 (0:00:00.676) 0:05:36.360 ********* 2025-08-29 17:39:01.832383 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832401 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832407 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832412 | orchestrator | 2025-08-29 17:39:01.832418 | orchestrator | TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] ********************* 2025-08-29 17:39:01.832423 | orchestrator | Friday 29 August 2025 17:33:17 +0000 (0:00:00.329) 0:05:36.689 ********* 2025-08-29 17:39:01.832429 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:39:01.832434 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:39:01.832440 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:39:01.832445 | orchestrator | changed: [testbed-node-0 -> {{ groups[mon_group_name][0] }}] 2025-08-29 17:39:01.832451 | orchestrator | 2025-08-29 17:39:01.832456 | orchestrator | TASK [ceph-mgr : Set_fact _mgr_keys] ******************************************* 2025-08-29 17:39:01.832462 | orchestrator | Friday 29 August 2025 17:33:27 +0000 (0:00:09.910) 0:05:46.600 ********* 2025-08-29 17:39:01.832467 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.832473 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.832478 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.832484 | orchestrator | 2025-08-29 17:39:01.832489 | orchestrator | TASK [ceph-mgr : Get keys from monitors] *************************************** 2025-08-29 17:39:01.832495 | orchestrator | Friday 29 August 2025 17:33:28 +0000 (0:00:00.858) 0:05:47.459 ********* 2025-08-29 17:39:01.832500 | orchestrator | skipping: [testbed-node-0] => (item=None)  2025-08-29 17:39:01.832505 | orchestrator | skipping: [testbed-node-1] => (item=None)  2025-08-29 17:39:01.832511 | orchestrator | skipping: [testbed-node-2] => (item=None)  2025-08-29 17:39:01.832516 | orchestrator | ok: [testbed-node-0] => (item=None) 2025-08-29 17:39:01.832522 | orchestrator | ok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.832530 | orchestrator | ok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.832536 | orchestrator | 2025-08-29 17:39:01.832556 | orchestrator | TASK [ceph-mgr : Copy ceph key(s) if needed] *********************************** 2025-08-29 17:39:01.832562 | orchestrator | Friday 29 August 2025 17:33:30 +0000 (0:00:02.109) 0:05:49.568 ********* 2025-08-29 17:39:01.832568 | orchestrator | skipping: [testbed-node-0] => (item=None)  2025-08-29 17:39:01.832573 | orchestrator | skipping: [testbed-node-1] => (item=None)  2025-08-29 17:39:01.832578 | orchestrator | skipping: [testbed-node-2] => (item=None)  2025-08-29 17:39:01.832584 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:39:01.832589 | orchestrator | changed: [testbed-node-1] => (item=None) 2025-08-29 17:39:01.832595 | orchestrator | changed: [testbed-node-2] => (item=None) 2025-08-29 17:39:01.832600 | orchestrator | 2025-08-29 17:39:01.832605 | orchestrator | TASK [ceph-mgr : Set mgr key permissions] ************************************** 2025-08-29 17:39:01.832611 | orchestrator | Friday 29 August 2025 17:33:31 +0000 (0:00:01.267) 0:05:50.835 ********* 2025-08-29 17:39:01.832616 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.832622 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.832627 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.832632 | orchestrator | 2025-08-29 17:39:01.832638 | orchestrator | TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] ***************** 2025-08-29 17:39:01.832643 | orchestrator | Friday 29 August 2025 17:33:32 +0000 (0:00:00.686) 0:05:51.522 ********* 2025-08-29 17:39:01.832649 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832654 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832664 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832669 | orchestrator | 2025-08-29 17:39:01.832675 | orchestrator | TASK [ceph-mgr : Include pre_requisite.yml] ************************************ 2025-08-29 17:39:01.832680 | orchestrator | Friday 29 August 2025 17:33:33 +0000 (0:00:00.653) 0:05:52.176 ********* 2025-08-29 17:39:01.832685 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832691 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832696 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832702 | orchestrator | 2025-08-29 17:39:01.832707 | orchestrator | TASK [ceph-mgr : Include start_mgr.yml] **************************************** 2025-08-29 17:39:01.832712 | orchestrator | Friday 29 August 2025 17:33:33 +0000 (0:00:00.323) 0:05:52.499 ********* 2025-08-29 17:39:01.832718 | orchestrator | included: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.832723 | orchestrator | 2025-08-29 17:39:01.832729 | orchestrator | TASK [ceph-mgr : Ensure systemd service override directory exists] ************* 2025-08-29 17:39:01.832734 | orchestrator | Friday 29 August 2025 17:33:33 +0000 (0:00:00.522) 0:05:53.021 ********* 2025-08-29 17:39:01.832740 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832745 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832751 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832756 | orchestrator | 2025-08-29 17:39:01.832761 | orchestrator | TASK [ceph-mgr : Add ceph-mgr systemd service overrides] *********************** 2025-08-29 17:39:01.832767 | orchestrator | Friday 29 August 2025 17:33:34 +0000 (0:00:00.751) 0:05:53.773 ********* 2025-08-29 17:39:01.832772 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832778 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832783 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.832788 | orchestrator | 2025-08-29 17:39:01.832794 | orchestrator | TASK [ceph-mgr : Include_tasks systemd.yml] ************************************ 2025-08-29 17:39:01.832799 | orchestrator | Friday 29 August 2025 17:33:35 +0000 (0:00:00.371) 0:05:54.144 ********* 2025-08-29 17:39:01.832805 | orchestrator | included: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.832810 | orchestrator | 2025-08-29 17:39:01.832815 | orchestrator | TASK [ceph-mgr : Generate systemd unit file] *********************************** 2025-08-29 17:39:01.832821 | orchestrator | Friday 29 August 2025 17:33:35 +0000 (0:00:00.631) 0:05:54.775 ********* 2025-08-29 17:39:01.832826 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.832832 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.832837 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.832842 | orchestrator | 2025-08-29 17:39:01.832848 | orchestrator | TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************ 2025-08-29 17:39:01.832853 | orchestrator | Friday 29 August 2025 17:33:37 +0000 (0:00:01.718) 0:05:56.494 ********* 2025-08-29 17:39:01.832859 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.832864 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.832869 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.832875 | orchestrator | 2025-08-29 17:39:01.832880 | orchestrator | TASK [ceph-mgr : Enable ceph-mgr.target] *************************************** 2025-08-29 17:39:01.832886 | orchestrator | Friday 29 August 2025 17:33:38 +0000 (0:00:01.304) 0:05:57.798 ********* 2025-08-29 17:39:01.832891 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.832896 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.832902 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.832907 | orchestrator | 2025-08-29 17:39:01.832913 | orchestrator | TASK [ceph-mgr : Systemd start mgr] ******************************************** 2025-08-29 17:39:01.832918 | orchestrator | Friday 29 August 2025 17:33:40 +0000 (0:00:01.839) 0:05:59.637 ********* 2025-08-29 17:39:01.832923 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.832929 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.832934 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.832940 | orchestrator | 2025-08-29 17:39:01.832945 | orchestrator | TASK [ceph-mgr : Include mgr_modules.yml] ************************************** 2025-08-29 17:39:01.832955 | orchestrator | Friday 29 August 2025 17:33:42 +0000 (0:00:01.903) 0:06:01.541 ********* 2025-08-29 17:39:01.832960 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.832966 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.832971 | orchestrator | included: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2 2025-08-29 17:39:01.832977 | orchestrator | 2025-08-29 17:39:01.832982 | orchestrator | TASK [ceph-mgr : Wait for all mgr to be up] ************************************ 2025-08-29 17:39:01.832988 | orchestrator | Friday 29 August 2025 17:33:43 +0000 (0:00:00.717) 0:06:02.259 ********* 2025-08-29 17:39:01.832996 | orchestrator | FAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left). 2025-08-29 17:39:01.833015 | orchestrator | FAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left). 2025-08-29 17:39:01.833021 | orchestrator | FAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left). 2025-08-29 17:39:01.833027 | orchestrator | FAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left). 2025-08-29 17:39:01.833032 | orchestrator | FAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left). 2025-08-29 17:39:01.833038 | orchestrator | ok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:39:01.833043 | orchestrator | 2025-08-29 17:39:01.833048 | orchestrator | TASK [ceph-mgr : Get enabled modules from ceph-mgr] **************************** 2025-08-29 17:39:01.833054 | orchestrator | Friday 29 August 2025 17:34:13 +0000 (0:00:29.935) 0:06:32.194 ********* 2025-08-29 17:39:01.833059 | orchestrator | ok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:39:01.833065 | orchestrator | 2025-08-29 17:39:01.833070 | orchestrator | TASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] *** 2025-08-29 17:39:01.833075 | orchestrator | Friday 29 August 2025 17:34:14 +0000 (0:00:01.308) 0:06:33.503 ********* 2025-08-29 17:39:01.833081 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.833086 | orchestrator | 2025-08-29 17:39:01.833091 | orchestrator | TASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] ************************** 2025-08-29 17:39:01.833097 | orchestrator | Friday 29 August 2025 17:34:14 +0000 (0:00:00.320) 0:06:33.823 ********* 2025-08-29 17:39:01.833102 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.833107 | orchestrator | 2025-08-29 17:39:01.833113 | orchestrator | TASK [ceph-mgr : Disable ceph mgr enabled modules] ***************************** 2025-08-29 17:39:01.833118 | orchestrator | Friday 29 August 2025 17:34:14 +0000 (0:00:00.170) 0:06:33.994 ********* 2025-08-29 17:39:01.833123 | orchestrator | changed: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat) 2025-08-29 17:39:01.833129 | orchestrator | changed: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs) 2025-08-29 17:39:01.833134 | orchestrator | changed: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful) 2025-08-29 17:39:01.833140 | orchestrator | 2025-08-29 17:39:01.833145 | orchestrator | TASK [ceph-mgr : Add modules to ceph-mgr] ************************************** 2025-08-29 17:39:01.833150 | orchestrator | Friday 29 August 2025 17:34:21 +0000 (0:00:06.292) 0:06:40.286 ********* 2025-08-29 17:39:01.833156 | orchestrator | skipping: [testbed-node-2] => (item=balancer)  2025-08-29 17:39:01.833161 | orchestrator | changed: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard) 2025-08-29 17:39:01.833166 | orchestrator | changed: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus) 2025-08-29 17:39:01.833172 | orchestrator | skipping: [testbed-node-2] => (item=status)  2025-08-29 17:39:01.833177 | orchestrator | 2025-08-29 17:39:01.833183 | orchestrator | RUNNING HANDLER [ceph-handler : Make tempdir for scripts] ********************** 2025-08-29 17:39:01.833188 | orchestrator | Friday 29 August 2025 17:34:26 +0000 (0:00:05.327) 0:06:45.614 ********* 2025-08-29 17:39:01.833194 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.833199 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.833209 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.833214 | orchestrator | 2025-08-29 17:39:01.833219 | orchestrator | RUNNING HANDLER [ceph-handler : Mgrs handler] ********************************** 2025-08-29 17:39:01.833225 | orchestrator | Friday 29 August 2025 17:34:27 +0000 (0:00:00.755) 0:06:46.370 ********* 2025-08-29 17:39:01.833230 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:01.833236 | orchestrator | 2025-08-29 17:39:01.833241 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ******** 2025-08-29 17:39:01.833246 | orchestrator | Friday 29 August 2025 17:34:27 +0000 (0:00:00.513) 0:06:46.883 ********* 2025-08-29 17:39:01.833252 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.833257 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.833263 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.833268 | orchestrator | 2025-08-29 17:39:01.833273 | orchestrator | RUNNING HANDLER [ceph-handler : Copy mgr restart script] *********************** 2025-08-29 17:39:01.833279 | orchestrator | Friday 29 August 2025 17:34:28 +0000 (0:00:00.606) 0:06:47.490 ********* 2025-08-29 17:39:01.833284 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.833290 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.833295 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.833300 | orchestrator | 2025-08-29 17:39:01.833306 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ******************** 2025-08-29 17:39:01.833311 | orchestrator | Friday 29 August 2025 17:34:29 +0000 (0:00:01.324) 0:06:48.814 ********* 2025-08-29 17:39:01.833317 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-0)  2025-08-29 17:39:01.833322 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-1)  2025-08-29 17:39:01.833327 | orchestrator | skipping: [testbed-node-0] => (item=testbed-node-2)  2025-08-29 17:39:01.833333 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.833338 | orchestrator | 2025-08-29 17:39:01.833343 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] ********* 2025-08-29 17:39:01.833349 | orchestrator | Friday 29 August 2025 17:34:30 +0000 (0:00:00.604) 0:06:49.419 ********* 2025-08-29 17:39:01.833354 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.833360 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.833365 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.833371 | orchestrator | 2025-08-29 17:39:01.833376 | orchestrator | PLAY [Apply role ceph-osd] ***************************************************** 2025-08-29 17:39:01.833381 | orchestrator | 2025-08-29 17:39:01.833429 | orchestrator | TASK [ceph-handler : Include check_running_cluster.yml] ************************ 2025-08-29 17:39:01.833436 | orchestrator | Friday 29 August 2025 17:34:31 +0000 (0:00:00.902) 0:06:50.322 ********* 2025-08-29 17:39:01.833445 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.833450 | orchestrator | 2025-08-29 17:39:01.833471 | orchestrator | TASK [ceph-handler : Include check_running_containers.yml] ********************* 2025-08-29 17:39:01.833478 | orchestrator | Friday 29 August 2025 17:34:31 +0000 (0:00:00.555) 0:06:50.877 ********* 2025-08-29 17:39:01.833484 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.833489 | orchestrator | 2025-08-29 17:39:01.833495 | orchestrator | TASK [ceph-handler : Check for a mon container] ******************************** 2025-08-29 17:39:01.833500 | orchestrator | Friday 29 August 2025 17:34:32 +0000 (0:00:00.785) 0:06:51.662 ********* 2025-08-29 17:39:01.833505 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.833511 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.833516 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.833522 | orchestrator | 2025-08-29 17:39:01.833527 | orchestrator | TASK [ceph-handler : Check for an osd container] ******************************* 2025-08-29 17:39:01.833532 | orchestrator | Friday 29 August 2025 17:34:32 +0000 (0:00:00.311) 0:06:51.974 ********* 2025-08-29 17:39:01.833538 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833547 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833553 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833558 | orchestrator | 2025-08-29 17:39:01.833564 | orchestrator | TASK [ceph-handler : Check for a mds container] ******************************** 2025-08-29 17:39:01.833569 | orchestrator | Friday 29 August 2025 17:34:33 +0000 (0:00:00.742) 0:06:52.716 ********* 2025-08-29 17:39:01.833575 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833580 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833585 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833591 | orchestrator | 2025-08-29 17:39:01.833596 | orchestrator | TASK [ceph-handler : Check for a rgw container] ******************************** 2025-08-29 17:39:01.833601 | orchestrator | Friday 29 August 2025 17:34:34 +0000 (0:00:00.744) 0:06:53.460 ********* 2025-08-29 17:39:01.833607 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833612 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833618 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833623 | orchestrator | 2025-08-29 17:39:01.833628 | orchestrator | TASK [ceph-handler : Check for a mgr container] ******************************** 2025-08-29 17:39:01.833634 | orchestrator | Friday 29 August 2025 17:34:35 +0000 (0:00:01.075) 0:06:54.536 ********* 2025-08-29 17:39:01.833639 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.833644 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.833650 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.833655 | orchestrator | 2025-08-29 17:39:01.833661 | orchestrator | TASK [ceph-handler : Check for a rbd mirror container] ************************* 2025-08-29 17:39:01.833666 | orchestrator | Friday 29 August 2025 17:34:35 +0000 (0:00:00.391) 0:06:54.927 ********* 2025-08-29 17:39:01.833671 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.833677 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.833682 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.833688 | orchestrator | 2025-08-29 17:39:01.833693 | orchestrator | TASK [ceph-handler : Check for a nfs container] ******************************** 2025-08-29 17:39:01.833698 | orchestrator | Friday 29 August 2025 17:34:36 +0000 (0:00:00.297) 0:06:55.224 ********* 2025-08-29 17:39:01.833704 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.833709 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.833714 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.833720 | orchestrator | 2025-08-29 17:39:01.833725 | orchestrator | TASK [ceph-handler : Check for a ceph-crash container] ************************* 2025-08-29 17:39:01.833730 | orchestrator | Friday 29 August 2025 17:34:36 +0000 (0:00:00.335) 0:06:55.560 ********* 2025-08-29 17:39:01.833736 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833741 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833747 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833752 | orchestrator | 2025-08-29 17:39:01.833757 | orchestrator | TASK [ceph-handler : Check for a ceph-exporter container] ********************** 2025-08-29 17:39:01.833763 | orchestrator | Friday 29 August 2025 17:34:37 +0000 (0:00:01.019) 0:06:56.580 ********* 2025-08-29 17:39:01.833768 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833774 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833779 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833784 | orchestrator | 2025-08-29 17:39:01.833790 | orchestrator | TASK [ceph-handler : Include check_socket_non_container.yml] ******************* 2025-08-29 17:39:01.833795 | orchestrator | Friday 29 August 2025 17:34:38 +0000 (0:00:00.732) 0:06:57.312 ********* 2025-08-29 17:39:01.833800 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.833806 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.833811 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.833817 | orchestrator | 2025-08-29 17:39:01.833822 | orchestrator | TASK [ceph-handler : Set_fact handler_mon_status] ****************************** 2025-08-29 17:39:01.833827 | orchestrator | Friday 29 August 2025 17:34:38 +0000 (0:00:00.404) 0:06:57.716 ********* 2025-08-29 17:39:01.833833 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.833838 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.833843 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.833852 | orchestrator | 2025-08-29 17:39:01.833857 | orchestrator | TASK [ceph-handler : Set_fact handler_osd_status] ****************************** 2025-08-29 17:39:01.833863 | orchestrator | Friday 29 August 2025 17:34:38 +0000 (0:00:00.320) 0:06:58.037 ********* 2025-08-29 17:39:01.833868 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833874 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833879 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833884 | orchestrator | 2025-08-29 17:39:01.833890 | orchestrator | TASK [ceph-handler : Set_fact handler_mds_status] ****************************** 2025-08-29 17:39:01.833895 | orchestrator | Friday 29 August 2025 17:34:39 +0000 (0:00:00.584) 0:06:58.622 ********* 2025-08-29 17:39:01.833901 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833906 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833911 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833917 | orchestrator | 2025-08-29 17:39:01.833922 | orchestrator | TASK [ceph-handler : Set_fact handler_rgw_status] ****************************** 2025-08-29 17:39:01.833928 | orchestrator | Friday 29 August 2025 17:34:39 +0000 (0:00:00.315) 0:06:58.937 ********* 2025-08-29 17:39:01.833933 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.833938 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.833944 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.833949 | orchestrator | 2025-08-29 17:39:01.833958 | orchestrator | TASK [ceph-handler : Set_fact handler_nfs_status] ****************************** 2025-08-29 17:39:01.833967 | orchestrator | Friday 29 August 2025 17:34:40 +0000 (0:00:00.335) 0:06:59.272 ********* 2025-08-29 17:39:01.833972 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.833978 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.833983 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.833989 | orchestrator | 2025-08-29 17:39:01.833994 | orchestrator | TASK [ceph-handler : Set_fact handler_rbd_status] ****************************** 2025-08-29 17:39:01.834000 | orchestrator | Friday 29 August 2025 17:34:40 +0000 (0:00:00.329) 0:06:59.602 ********* 2025-08-29 17:39:01.834005 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.834011 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.834036 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.834043 | orchestrator | 2025-08-29 17:39:01.834049 | orchestrator | TASK [ceph-handler : Set_fact handler_mgr_status] ****************************** 2025-08-29 17:39:01.834054 | orchestrator | Friday 29 August 2025 17:34:40 +0000 (0:00:00.534) 0:07:00.136 ********* 2025-08-29 17:39:01.834060 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.834065 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.834071 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.834076 | orchestrator | 2025-08-29 17:39:01.834082 | orchestrator | TASK [ceph-handler : Set_fact handler_crash_status] **************************** 2025-08-29 17:39:01.834087 | orchestrator | Friday 29 August 2025 17:34:41 +0000 (0:00:00.296) 0:07:00.433 ********* 2025-08-29 17:39:01.834093 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.834098 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.834104 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.834109 | orchestrator | 2025-08-29 17:39:01.834115 | orchestrator | TASK [ceph-handler : Set_fact handler_exporter_status] ************************* 2025-08-29 17:39:01.834120 | orchestrator | Friday 29 August 2025 17:34:41 +0000 (0:00:00.381) 0:07:00.814 ********* 2025-08-29 17:39:01.834125 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.834131 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.834136 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.834142 | orchestrator | 2025-08-29 17:39:01.834147 | orchestrator | TASK [ceph-osd : Set_fact add_osd] ********************************************* 2025-08-29 17:39:01.834153 | orchestrator | Friday 29 August 2025 17:34:42 +0000 (0:00:00.521) 0:07:01.336 ********* 2025-08-29 17:39:01.834158 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.834164 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.834169 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.834175 | orchestrator | 2025-08-29 17:39:01.834180 | orchestrator | TASK [ceph-osd : Set_fact container_exec_cmd] ********************************** 2025-08-29 17:39:01.834190 | orchestrator | Friday 29 August 2025 17:34:42 +0000 (0:00:00.580) 0:07:01.917 ********* 2025-08-29 17:39:01.834195 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0) 2025-08-29 17:39:01.834201 | orchestrator | ok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:39:01.834206 | orchestrator | ok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:39:01.834212 | orchestrator | 2025-08-29 17:39:01.834217 | orchestrator | TASK [ceph-osd : Include_tasks system_tuning.yml] ****************************** 2025-08-29 17:39:01.834223 | orchestrator | Friday 29 August 2025 17:34:43 +0000 (0:00:00.610) 0:07:02.528 ********* 2025-08-29 17:39:01.834228 | orchestrator | included: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.834234 | orchestrator | 2025-08-29 17:39:01.834239 | orchestrator | TASK [ceph-osd : Create tmpfiles.d directory] ********************************** 2025-08-29 17:39:01.834245 | orchestrator | Friday 29 August 2025 17:34:43 +0000 (0:00:00.522) 0:07:03.050 ********* 2025-08-29 17:39:01.834250 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.834256 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.834261 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.834266 | orchestrator | 2025-08-29 17:39:01.834272 | orchestrator | TASK [ceph-osd : Disable transparent hugepage] ********************************* 2025-08-29 17:39:01.834277 | orchestrator | Friday 29 August 2025 17:34:44 +0000 (0:00:00.647) 0:07:03.698 ********* 2025-08-29 17:39:01.834283 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.834288 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.834294 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.834299 | orchestrator | 2025-08-29 17:39:01.834305 | orchestrator | TASK [ceph-osd : Get default vm.min_free_kbytes] ******************************* 2025-08-29 17:39:01.834310 | orchestrator | Friday 29 August 2025 17:34:44 +0000 (0:00:00.325) 0:07:04.024 ********* 2025-08-29 17:39:01.834316 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.834321 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.834326 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.834332 | orchestrator | 2025-08-29 17:39:01.834337 | orchestrator | TASK [ceph-osd : Set_fact vm_min_free_kbytes] ********************************** 2025-08-29 17:39:01.834343 | orchestrator | Friday 29 August 2025 17:34:45 +0000 (0:00:00.588) 0:07:04.612 ********* 2025-08-29 17:39:01.834348 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.834354 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.834359 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.834365 | orchestrator | 2025-08-29 17:39:01.834370 | orchestrator | TASK [ceph-osd : Apply operating system tuning] ******************************** 2025-08-29 17:39:01.834376 | orchestrator | Friday 29 August 2025 17:34:45 +0000 (0:00:00.376) 0:07:04.988 ********* 2025-08-29 17:39:01.834381 | orchestrator | changed: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True}) 2025-08-29 17:39:01.834399 | orchestrator | changed: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True}) 2025-08-29 17:39:01.834405 | orchestrator | changed: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True}) 2025-08-29 17:39:01.834410 | orchestrator | changed: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859}) 2025-08-29 17:39:01.834416 | orchestrator | changed: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859}) 2025-08-29 17:39:01.834424 | orchestrator | changed: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859}) 2025-08-29 17:39:01.834436 | orchestrator | changed: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0}) 2025-08-29 17:39:01.834441 | orchestrator | changed: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0}) 2025-08-29 17:39:01.834447 | orchestrator | changed: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0}) 2025-08-29 17:39:01.834452 | orchestrator | changed: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10}) 2025-08-29 17:39:01.834465 | orchestrator | changed: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10}) 2025-08-29 17:39:01.834470 | orchestrator | changed: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10}) 2025-08-29 17:39:01.834476 | orchestrator | changed: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'}) 2025-08-29 17:39:01.834481 | orchestrator | changed: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'}) 2025-08-29 17:39:01.834487 | orchestrator | changed: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'}) 2025-08-29 17:39:01.834492 | orchestrator | 2025-08-29 17:39:01.834498 | orchestrator | TASK [ceph-osd : Install dependencies] ***************************************** 2025-08-29 17:39:01.834503 | orchestrator | Friday 29 August 2025 17:34:48 +0000 (0:00:02.413) 0:07:07.402 ********* 2025-08-29 17:39:01.834509 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.834515 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.834520 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.834525 | orchestrator | 2025-08-29 17:39:01.834531 | orchestrator | TASK [ceph-osd : Include_tasks common.yml] ************************************* 2025-08-29 17:39:01.834536 | orchestrator | Friday 29 August 2025 17:34:48 +0000 (0:00:00.307) 0:07:07.709 ********* 2025-08-29 17:39:01.834542 | orchestrator | included: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.834548 | orchestrator | 2025-08-29 17:39:01.834553 | orchestrator | TASK [ceph-osd : Create bootstrap-osd and osd directories] ********************* 2025-08-29 17:39:01.834559 | orchestrator | Friday 29 August 2025 17:34:49 +0000 (0:00:00.501) 0:07:08.211 ********* 2025-08-29 17:39:01.834564 | orchestrator | ok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/) 2025-08-29 17:39:01.834570 | orchestrator | ok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/) 2025-08-29 17:39:01.834575 | orchestrator | ok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/) 2025-08-29 17:39:01.834580 | orchestrator | ok: [testbed-node-3] => (item=/var/lib/ceph/osd/) 2025-08-29 17:39:01.834586 | orchestrator | ok: [testbed-node-5] => (item=/var/lib/ceph/osd/) 2025-08-29 17:39:01.834591 | orchestrator | ok: [testbed-node-4] => (item=/var/lib/ceph/osd/) 2025-08-29 17:39:01.834597 | orchestrator | 2025-08-29 17:39:01.834602 | orchestrator | TASK [ceph-osd : Get keys from monitors] *************************************** 2025-08-29 17:39:01.834608 | orchestrator | Friday 29 August 2025 17:34:50 +0000 (0:00:01.216) 0:07:09.427 ********* 2025-08-29 17:39:01.834613 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.834619 | orchestrator | skipping: [testbed-node-3] => (item=None)  2025-08-29 17:39:01.834624 | orchestrator | ok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}] 2025-08-29 17:39:01.834630 | orchestrator | 2025-08-29 17:39:01.834635 | orchestrator | TASK [ceph-osd : Copy ceph key(s) if needed] *********************************** 2025-08-29 17:39:01.834641 | orchestrator | Friday 29 August 2025 17:34:52 +0000 (0:00:02.346) 0:07:11.773 ********* 2025-08-29 17:39:01.834646 | orchestrator | changed: [testbed-node-3] => (item=None) 2025-08-29 17:39:01.834652 | orchestrator | skipping: [testbed-node-3] => (item=None)  2025-08-29 17:39:01.834657 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.834663 | orchestrator | changed: [testbed-node-4] => (item=None) 2025-08-29 17:39:01.834668 | orchestrator | skipping: [testbed-node-4] => (item=None)  2025-08-29 17:39:01.834674 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.834679 | orchestrator | changed: [testbed-node-5] => (item=None) 2025-08-29 17:39:01.834684 | orchestrator | skipping: [testbed-node-5] => (item=None)  2025-08-29 17:39:01.834690 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.834695 | orchestrator | 2025-08-29 17:39:01.834701 | orchestrator | TASK [ceph-osd : Set noup flag] ************************************************ 2025-08-29 17:39:01.834706 | orchestrator | Friday 29 August 2025 17:34:53 +0000 (0:00:01.224) 0:07:12.998 ********* 2025-08-29 17:39:01.834712 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:39:01.834722 | orchestrator | 2025-08-29 17:39:01.834728 | orchestrator | TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ****************************** 2025-08-29 17:39:01.834734 | orchestrator | Friday 29 August 2025 17:34:56 +0000 (0:00:02.354) 0:07:15.353 ********* 2025-08-29 17:39:01.834739 | orchestrator | included: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.834745 | orchestrator | 2025-08-29 17:39:01.834750 | orchestrator | TASK [ceph-osd : Use ceph-volume to create osds] ******************************* 2025-08-29 17:39:01.834756 | orchestrator | Friday 29 August 2025 17:34:56 +0000 (0:00:00.598) 0:07:15.951 ********* 2025-08-29 17:39:01.834761 | orchestrator | changed: [testbed-node-4] => (item={'data': 'osd-block-b89384ec-5219-5f2a-8735-84f78c8179d2', 'data_vg': 'ceph-b89384ec-5219-5f2a-8735-84f78c8179d2'}) 2025-08-29 17:39:01.834768 | orchestrator | changed: [testbed-node-5] => (item={'data': 'osd-block-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe', 'data_vg': 'ceph-e5f8acb9-1955-50e0-bba9-ecbc3a5da5fe'}) 2025-08-29 17:39:01.834777 | orchestrator | changed: [testbed-node-3] => (item={'data': 'osd-block-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b', 'data_vg': 'ceph-25b2a0ab-1f04-5c39-a4c5-8c8c13772d9b'}) 2025-08-29 17:39:01.834786 | orchestrator | changed: [testbed-node-5] => (item={'data': 'osd-block-21b183ef-757d-561c-bead-7bb3aee28288', 'data_vg': 'ceph-21b183ef-757d-561c-bead-7bb3aee28288'}) 2025-08-29 17:39:01.834792 | orchestrator | changed: [testbed-node-4] => (item={'data': 'osd-block-860f9296-4089-56f8-8238-0f24b03dbca2', 'data_vg': 'ceph-860f9296-4089-56f8-8238-0f24b03dbca2'}) 2025-08-29 17:39:01.834797 | orchestrator | changed: [testbed-node-3] => (item={'data': 'osd-block-67ff47d9-d75a-55af-b095-c4dbbf8f796a', 'data_vg': 'ceph-67ff47d9-d75a-55af-b095-c4dbbf8f796a'}) 2025-08-29 17:39:01.834803 | orchestrator | 2025-08-29 17:39:01.834808 | orchestrator | TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************ 2025-08-29 17:39:01.834814 | orchestrator | Friday 29 August 2025 17:35:38 +0000 (0:00:41.385) 0:07:57.337 ********* 2025-08-29 17:39:01.834819 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.834825 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.834830 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.834835 | orchestrator | 2025-08-29 17:39:01.834841 | orchestrator | TASK [ceph-osd : Include_tasks start_osds.yml] ********************************* 2025-08-29 17:39:01.834846 | orchestrator | Friday 29 August 2025 17:35:38 +0000 (0:00:00.432) 0:07:57.769 ********* 2025-08-29 17:39:01.834852 | orchestrator | included: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.834857 | orchestrator | 2025-08-29 17:39:01.834863 | orchestrator | TASK [ceph-osd : Get osd ids] ************************************************** 2025-08-29 17:39:01.834868 | orchestrator | Friday 29 August 2025 17:35:39 +0000 (0:00:00.519) 0:07:58.289 ********* 2025-08-29 17:39:01.834874 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.834879 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.834884 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.834890 | orchestrator | 2025-08-29 17:39:01.834895 | orchestrator | TASK [ceph-osd : Collect osd ids] ********************************************** 2025-08-29 17:39:01.834901 | orchestrator | Friday 29 August 2025 17:35:40 +0000 (0:00:00.984) 0:07:59.273 ********* 2025-08-29 17:39:01.834906 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.834911 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.834917 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.834922 | orchestrator | 2025-08-29 17:39:01.834928 | orchestrator | TASK [ceph-osd : Include_tasks systemd.yml] ************************************ 2025-08-29 17:39:01.834933 | orchestrator | Friday 29 August 2025 17:35:43 +0000 (0:00:02.994) 0:08:02.268 ********* 2025-08-29 17:39:01.834939 | orchestrator | included: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.834944 | orchestrator | 2025-08-29 17:39:01.834950 | orchestrator | TASK [ceph-osd : Generate systemd unit file] *********************************** 2025-08-29 17:39:01.834969 | orchestrator | Friday 29 August 2025 17:35:43 +0000 (0:00:00.509) 0:08:02.777 ********* 2025-08-29 17:39:01.834974 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.834980 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.834985 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.834991 | orchestrator | 2025-08-29 17:39:01.834996 | orchestrator | TASK [ceph-osd : Generate systemd ceph-osd target file] ************************ 2025-08-29 17:39:01.835002 | orchestrator | Friday 29 August 2025 17:35:45 +0000 (0:00:01.529) 0:08:04.307 ********* 2025-08-29 17:39:01.835007 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.835013 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.835018 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.835023 | orchestrator | 2025-08-29 17:39:01.835029 | orchestrator | TASK [ceph-osd : Enable ceph-osd.target] *************************************** 2025-08-29 17:39:01.835034 | orchestrator | Friday 29 August 2025 17:35:46 +0000 (0:00:01.202) 0:08:05.509 ********* 2025-08-29 17:39:01.835040 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.835045 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.835051 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.835056 | orchestrator | 2025-08-29 17:39:01.835061 | orchestrator | TASK [ceph-osd : Ensure systemd service override directory exists] ************* 2025-08-29 17:39:01.835067 | orchestrator | Friday 29 August 2025 17:35:48 +0000 (0:00:01.759) 0:08:07.268 ********* 2025-08-29 17:39:01.835072 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835078 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835083 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835088 | orchestrator | 2025-08-29 17:39:01.835094 | orchestrator | TASK [ceph-osd : Add ceph-osd systemd service overrides] *********************** 2025-08-29 17:39:01.835099 | orchestrator | Friday 29 August 2025 17:35:48 +0000 (0:00:00.336) 0:08:07.605 ********* 2025-08-29 17:39:01.835105 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835110 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835116 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835121 | orchestrator | 2025-08-29 17:39:01.835126 | orchestrator | TASK [ceph-osd : Ensure /var/lib/ceph/osd/- is present] ********* 2025-08-29 17:39:01.835132 | orchestrator | Friday 29 August 2025 17:35:49 +0000 (0:00:00.631) 0:08:08.236 ********* 2025-08-29 17:39:01.835137 | orchestrator | ok: [testbed-node-3] => (item=0) 2025-08-29 17:39:01.835143 | orchestrator | ok: [testbed-node-4] => (item=2) 2025-08-29 17:39:01.835148 | orchestrator | ok: [testbed-node-5] => (item=4) 2025-08-29 17:39:01.835154 | orchestrator | ok: [testbed-node-3] => (item=5) 2025-08-29 17:39:01.835159 | orchestrator | ok: [testbed-node-4] => (item=3) 2025-08-29 17:39:01.835164 | orchestrator | ok: [testbed-node-5] => (item=1) 2025-08-29 17:39:01.835170 | orchestrator | 2025-08-29 17:39:01.835175 | orchestrator | TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] ***************** 2025-08-29 17:39:01.835181 | orchestrator | Friday 29 August 2025 17:35:50 +0000 (0:00:01.045) 0:08:09.282 ********* 2025-08-29 17:39:01.835186 | orchestrator | changed: [testbed-node-3] => (item=0) 2025-08-29 17:39:01.835192 | orchestrator | changed: [testbed-node-4] => (item=2) 2025-08-29 17:39:01.835197 | orchestrator | changed: [testbed-node-5] => (item=4) 2025-08-29 17:39:01.835203 | orchestrator | changed: [testbed-node-3] => (item=5) 2025-08-29 17:39:01.835208 | orchestrator | changed: [testbed-node-4] => (item=3) 2025-08-29 17:39:01.835216 | orchestrator | changed: [testbed-node-5] => (item=1) 2025-08-29 17:39:01.835222 | orchestrator | 2025-08-29 17:39:01.835230 | orchestrator | TASK [ceph-osd : Systemd start osd] ******************************************** 2025-08-29 17:39:01.835236 | orchestrator | Friday 29 August 2025 17:35:52 +0000 (0:00:02.279) 0:08:11.562 ********* 2025-08-29 17:39:01.835242 | orchestrator | changed: [testbed-node-5] => (item=4) 2025-08-29 17:39:01.835247 | orchestrator | changed: [testbed-node-3] => (item=0) 2025-08-29 17:39:01.835253 | orchestrator | changed: [testbed-node-4] => (item=2) 2025-08-29 17:39:01.835258 | orchestrator | changed: [testbed-node-5] => (item=1) 2025-08-29 17:39:01.835267 | orchestrator | changed: [testbed-node-3] => (item=5) 2025-08-29 17:39:01.835272 | orchestrator | changed: [testbed-node-4] => (item=3) 2025-08-29 17:39:01.835278 | orchestrator | 2025-08-29 17:39:01.835283 | orchestrator | TASK [ceph-osd : Unset noup flag] ********************************************** 2025-08-29 17:39:01.835289 | orchestrator | Friday 29 August 2025 17:35:56 +0000 (0:00:03.810) 0:08:15.373 ********* 2025-08-29 17:39:01.835294 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835300 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835305 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:39:01.835310 | orchestrator | 2025-08-29 17:39:01.835316 | orchestrator | TASK [ceph-osd : Wait for all osd to be up] ************************************ 2025-08-29 17:39:01.835321 | orchestrator | Friday 29 August 2025 17:35:59 +0000 (0:00:03.601) 0:08:18.975 ********* 2025-08-29 17:39:01.835327 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835332 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835338 | orchestrator | FAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left). 2025-08-29 17:39:01.835343 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:39:01.835349 | orchestrator | 2025-08-29 17:39:01.835355 | orchestrator | TASK [ceph-osd : Include crush_rules.yml] ************************************** 2025-08-29 17:39:01.835360 | orchestrator | Friday 29 August 2025 17:36:12 +0000 (0:00:12.708) 0:08:31.683 ********* 2025-08-29 17:39:01.835366 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835371 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835376 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835382 | orchestrator | 2025-08-29 17:39:01.835398 | orchestrator | RUNNING HANDLER [ceph-handler : Make tempdir for scripts] ********************** 2025-08-29 17:39:01.835404 | orchestrator | Friday 29 August 2025 17:36:13 +0000 (0:00:01.080) 0:08:32.764 ********* 2025-08-29 17:39:01.835410 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835415 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835421 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835426 | orchestrator | 2025-08-29 17:39:01.835431 | orchestrator | RUNNING HANDLER [ceph-handler : Osds handler] ********************************** 2025-08-29 17:39:01.835437 | orchestrator | Friday 29 August 2025 17:36:13 +0000 (0:00:00.354) 0:08:33.118 ********* 2025-08-29 17:39:01.835443 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.835448 | orchestrator | 2025-08-29 17:39:01.835454 | orchestrator | RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] ********************** 2025-08-29 17:39:01.835459 | orchestrator | Friday 29 August 2025 17:36:14 +0000 (0:00:00.519) 0:08:33.638 ********* 2025-08-29 17:39:01.835465 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.835470 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.835476 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.835481 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835486 | orchestrator | 2025-08-29 17:39:01.835492 | orchestrator | RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ******** 2025-08-29 17:39:01.835497 | orchestrator | Friday 29 August 2025 17:36:15 +0000 (0:00:00.838) 0:08:34.477 ********* 2025-08-29 17:39:01.835503 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835508 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835514 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835519 | orchestrator | 2025-08-29 17:39:01.835524 | orchestrator | RUNNING HANDLER [ceph-handler : Unset noup flag] ******************************* 2025-08-29 17:39:01.835530 | orchestrator | Friday 29 August 2025 17:36:15 +0000 (0:00:00.308) 0:08:34.786 ********* 2025-08-29 17:39:01.835535 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835541 | orchestrator | 2025-08-29 17:39:01.835546 | orchestrator | RUNNING HANDLER [ceph-handler : Copy osd restart script] *********************** 2025-08-29 17:39:01.835552 | orchestrator | Friday 29 August 2025 17:36:15 +0000 (0:00:00.225) 0:08:35.011 ********* 2025-08-29 17:39:01.835561 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835566 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835572 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835577 | orchestrator | 2025-08-29 17:39:01.835583 | orchestrator | RUNNING HANDLER [ceph-handler : Get pool list] ********************************* 2025-08-29 17:39:01.835588 | orchestrator | Friday 29 August 2025 17:36:16 +0000 (0:00:00.299) 0:08:35.310 ********* 2025-08-29 17:39:01.835594 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835599 | orchestrator | 2025-08-29 17:39:01.835605 | orchestrator | RUNNING HANDLER [ceph-handler : Get balancer module status] ******************** 2025-08-29 17:39:01.835610 | orchestrator | Friday 29 August 2025 17:36:16 +0000 (0:00:00.212) 0:08:35.523 ********* 2025-08-29 17:39:01.835616 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835621 | orchestrator | 2025-08-29 17:39:01.835627 | orchestrator | RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] ************** 2025-08-29 17:39:01.835632 | orchestrator | Friday 29 August 2025 17:36:16 +0000 (0:00:00.236) 0:08:35.760 ********* 2025-08-29 17:39:01.835638 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835643 | orchestrator | 2025-08-29 17:39:01.835649 | orchestrator | RUNNING HANDLER [ceph-handler : Disable balancer] ****************************** 2025-08-29 17:39:01.835654 | orchestrator | Friday 29 August 2025 17:36:16 +0000 (0:00:00.123) 0:08:35.884 ********* 2025-08-29 17:39:01.835660 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835665 | orchestrator | 2025-08-29 17:39:01.835671 | orchestrator | RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] ***************** 2025-08-29 17:39:01.835679 | orchestrator | Friday 29 August 2025 17:36:16 +0000 (0:00:00.219) 0:08:36.103 ********* 2025-08-29 17:39:01.835688 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835694 | orchestrator | 2025-08-29 17:39:01.835699 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] ******************* 2025-08-29 17:39:01.835705 | orchestrator | Friday 29 August 2025 17:36:17 +0000 (0:00:00.783) 0:08:36.887 ********* 2025-08-29 17:39:01.835710 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.835716 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.835721 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.835726 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835732 | orchestrator | 2025-08-29 17:39:01.835737 | orchestrator | RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] ********* 2025-08-29 17:39:01.835743 | orchestrator | Friday 29 August 2025 17:36:18 +0000 (0:00:00.449) 0:08:37.337 ********* 2025-08-29 17:39:01.835748 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835754 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835759 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835764 | orchestrator | 2025-08-29 17:39:01.835770 | orchestrator | RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] *************** 2025-08-29 17:39:01.835775 | orchestrator | Friday 29 August 2025 17:36:18 +0000 (0:00:00.303) 0:08:37.640 ********* 2025-08-29 17:39:01.835781 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835786 | orchestrator | 2025-08-29 17:39:01.835791 | orchestrator | RUNNING HANDLER [ceph-handler : Re-enable balancer] **************************** 2025-08-29 17:39:01.835797 | orchestrator | Friday 29 August 2025 17:36:18 +0000 (0:00:00.216) 0:08:37.857 ********* 2025-08-29 17:39:01.835802 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835808 | orchestrator | 2025-08-29 17:39:01.835813 | orchestrator | PLAY [Apply role ceph-crash] *************************************************** 2025-08-29 17:39:01.835818 | orchestrator | 2025-08-29 17:39:01.835824 | orchestrator | TASK [ceph-handler : Include check_running_cluster.yml] ************************ 2025-08-29 17:39:01.835829 | orchestrator | Friday 29 August 2025 17:36:19 +0000 (0:00:00.649) 0:08:38.506 ********* 2025-08-29 17:39:01.835835 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.835846 | orchestrator | 2025-08-29 17:39:01.835851 | orchestrator | TASK [ceph-handler : Include check_running_containers.yml] ********************* 2025-08-29 17:39:01.835857 | orchestrator | Friday 29 August 2025 17:36:20 +0000 (0:00:01.167) 0:08:39.673 ********* 2025-08-29 17:39:01.835862 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.835868 | orchestrator | 2025-08-29 17:39:01.835873 | orchestrator | TASK [ceph-handler : Check for a mon container] ******************************** 2025-08-29 17:39:01.835879 | orchestrator | Friday 29 August 2025 17:36:21 +0000 (0:00:01.141) 0:08:40.815 ********* 2025-08-29 17:39:01.835884 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.835890 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.835895 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.835901 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.835906 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.835912 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.835917 | orchestrator | 2025-08-29 17:39:01.835923 | orchestrator | TASK [ceph-handler : Check for an osd container] ******************************* 2025-08-29 17:39:01.835928 | orchestrator | Friday 29 August 2025 17:36:22 +0000 (0:00:00.907) 0:08:41.722 ********* 2025-08-29 17:39:01.835934 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.835939 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.835945 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.835950 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.835955 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.835961 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.835966 | orchestrator | 2025-08-29 17:39:01.835972 | orchestrator | TASK [ceph-handler : Check for a mds container] ******************************** 2025-08-29 17:39:01.835977 | orchestrator | Friday 29 August 2025 17:36:23 +0000 (0:00:00.932) 0:08:42.655 ********* 2025-08-29 17:39:01.835983 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.835988 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.835994 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.835999 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836004 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836010 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836015 | orchestrator | 2025-08-29 17:39:01.836021 | orchestrator | TASK [ceph-handler : Check for a rgw container] ******************************** 2025-08-29 17:39:01.836026 | orchestrator | Friday 29 August 2025 17:36:24 +0000 (0:00:01.308) 0:08:43.963 ********* 2025-08-29 17:39:01.836032 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836037 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836042 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836048 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836053 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836059 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836064 | orchestrator | 2025-08-29 17:39:01.836070 | orchestrator | TASK [ceph-handler : Check for a mgr container] ******************************** 2025-08-29 17:39:01.836075 | orchestrator | Friday 29 August 2025 17:36:25 +0000 (0:00:00.955) 0:08:44.919 ********* 2025-08-29 17:39:01.836081 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836086 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836092 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836097 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836103 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.836108 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.836113 | orchestrator | 2025-08-29 17:39:01.836119 | orchestrator | TASK [ceph-handler : Check for a rbd mirror container] ************************* 2025-08-29 17:39:01.836125 | orchestrator | Friday 29 August 2025 17:36:26 +0000 (0:00:00.973) 0:08:45.892 ********* 2025-08-29 17:39:01.836130 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836135 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836141 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836153 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836158 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836167 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836173 | orchestrator | 2025-08-29 17:39:01.836178 | orchestrator | TASK [ceph-handler : Check for a nfs container] ******************************** 2025-08-29 17:39:01.836184 | orchestrator | Friday 29 August 2025 17:36:27 +0000 (0:00:00.596) 0:08:46.489 ********* 2025-08-29 17:39:01.836189 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836195 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836200 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836206 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836211 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836216 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836222 | orchestrator | 2025-08-29 17:39:01.836227 | orchestrator | TASK [ceph-handler : Check for a ceph-crash container] ************************* 2025-08-29 17:39:01.836233 | orchestrator | Friday 29 August 2025 17:36:28 +0000 (0:00:00.806) 0:08:47.295 ********* 2025-08-29 17:39:01.836238 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836244 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.836249 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.836255 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836260 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836265 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836271 | orchestrator | 2025-08-29 17:39:01.836276 | orchestrator | TASK [ceph-handler : Check for a ceph-exporter container] ********************** 2025-08-29 17:39:01.836282 | orchestrator | Friday 29 August 2025 17:36:29 +0000 (0:00:00.978) 0:08:48.274 ********* 2025-08-29 17:39:01.836287 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836293 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.836298 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.836304 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836309 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836315 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836320 | orchestrator | 2025-08-29 17:39:01.836326 | orchestrator | TASK [ceph-handler : Include check_socket_non_container.yml] ******************* 2025-08-29 17:39:01.836331 | orchestrator | Friday 29 August 2025 17:36:30 +0000 (0:00:01.266) 0:08:49.541 ********* 2025-08-29 17:39:01.836337 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836342 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836348 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836353 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836359 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836364 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836370 | orchestrator | 2025-08-29 17:39:01.836375 | orchestrator | TASK [ceph-handler : Set_fact handler_mon_status] ****************************** 2025-08-29 17:39:01.836381 | orchestrator | Friday 29 August 2025 17:36:30 +0000 (0:00:00.597) 0:08:50.139 ********* 2025-08-29 17:39:01.836398 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836404 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.836410 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.836415 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836421 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836426 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836432 | orchestrator | 2025-08-29 17:39:01.836437 | orchestrator | TASK [ceph-handler : Set_fact handler_osd_status] ****************************** 2025-08-29 17:39:01.836443 | orchestrator | Friday 29 August 2025 17:36:31 +0000 (0:00:00.824) 0:08:50.963 ********* 2025-08-29 17:39:01.836448 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836454 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836459 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836465 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836470 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836475 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836481 | orchestrator | 2025-08-29 17:39:01.836487 | orchestrator | TASK [ceph-handler : Set_fact handler_mds_status] ****************************** 2025-08-29 17:39:01.836496 | orchestrator | Friday 29 August 2025 17:36:32 +0000 (0:00:00.613) 0:08:51.577 ********* 2025-08-29 17:39:01.836502 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836507 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836513 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836518 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836524 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836529 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836534 | orchestrator | 2025-08-29 17:39:01.836540 | orchestrator | TASK [ceph-handler : Set_fact handler_rgw_status] ****************************** 2025-08-29 17:39:01.836545 | orchestrator | Friday 29 August 2025 17:36:33 +0000 (0:00:00.806) 0:08:52.384 ********* 2025-08-29 17:39:01.836551 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836556 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836562 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836567 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836573 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836578 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836583 | orchestrator | 2025-08-29 17:39:01.836589 | orchestrator | TASK [ceph-handler : Set_fact handler_nfs_status] ****************************** 2025-08-29 17:39:01.836594 | orchestrator | Friday 29 August 2025 17:36:33 +0000 (0:00:00.632) 0:08:53.016 ********* 2025-08-29 17:39:01.836600 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836605 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836611 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836616 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836622 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836627 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836632 | orchestrator | 2025-08-29 17:39:01.836638 | orchestrator | TASK [ceph-handler : Set_fact handler_rbd_status] ****************************** 2025-08-29 17:39:01.836643 | orchestrator | Friday 29 August 2025 17:36:34 +0000 (0:00:00.881) 0:08:53.898 ********* 2025-08-29 17:39:01.836649 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:01.836654 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:01.836660 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:01.836665 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836671 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836676 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836681 | orchestrator | 2025-08-29 17:39:01.836687 | orchestrator | TASK [ceph-handler : Set_fact handler_mgr_status] ****************************** 2025-08-29 17:39:01.836692 | orchestrator | Friday 29 August 2025 17:36:35 +0000 (0:00:00.561) 0:08:54.460 ********* 2025-08-29 17:39:01.836698 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836703 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.836711 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.836717 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.836726 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.836732 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.836737 | orchestrator | 2025-08-29 17:39:01.836742 | orchestrator | TASK [ceph-handler : Set_fact handler_crash_status] **************************** 2025-08-29 17:39:01.836748 | orchestrator | Friday 29 August 2025 17:36:36 +0000 (0:00:00.878) 0:08:55.338 ********* 2025-08-29 17:39:01.836753 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836759 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.836764 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.836770 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836775 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836781 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836786 | orchestrator | 2025-08-29 17:39:01.836791 | orchestrator | TASK [ceph-handler : Set_fact handler_exporter_status] ************************* 2025-08-29 17:39:01.836797 | orchestrator | Friday 29 August 2025 17:36:36 +0000 (0:00:00.608) 0:08:55.947 ********* 2025-08-29 17:39:01.836802 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836808 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.836817 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.836823 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.836828 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.836834 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.836839 | orchestrator | 2025-08-29 17:39:01.836844 | orchestrator | TASK [ceph-crash : Create client.crash keyring] ******************************** 2025-08-29 17:39:01.836850 | orchestrator | Friday 29 August 2025 17:36:38 +0000 (0:00:01.386) 0:08:57.333 ********* 2025-08-29 17:39:01.836855 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.836861 | orchestrator | 2025-08-29 17:39:01.836866 | orchestrator | TASK [ceph-crash : Get keys from monitors] ************************************* 2025-08-29 17:39:01.836872 | orchestrator | Friday 29 August 2025 17:36:42 +0000 (0:00:04.174) 0:09:01.508 ********* 2025-08-29 17:39:01.836877 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836883 | orchestrator | 2025-08-29 17:39:01.836888 | orchestrator | TASK [ceph-crash : Copy ceph key(s) if needed] ********************************* 2025-08-29 17:39:01.836894 | orchestrator | Friday 29 August 2025 17:36:44 +0000 (0:00:02.611) 0:09:04.119 ********* 2025-08-29 17:39:01.836899 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.836905 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.836910 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.836916 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.836921 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.836926 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.836932 | orchestrator | 2025-08-29 17:39:01.836937 | orchestrator | TASK [ceph-crash : Create /var/lib/ceph/crash/posted] ************************** 2025-08-29 17:39:01.836943 | orchestrator | Friday 29 August 2025 17:36:46 +0000 (0:00:01.805) 0:09:05.925 ********* 2025-08-29 17:39:01.836948 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.836953 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.836959 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.836964 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.836970 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.836975 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.836980 | orchestrator | 2025-08-29 17:39:01.836986 | orchestrator | TASK [ceph-crash : Include_tasks systemd.yml] ********************************** 2025-08-29 17:39:01.836991 | orchestrator | Friday 29 August 2025 17:36:48 +0000 (0:00:01.242) 0:09:07.168 ********* 2025-08-29 17:39:01.836997 | orchestrator | included: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.837003 | orchestrator | 2025-08-29 17:39:01.837009 | orchestrator | TASK [ceph-crash : Generate systemd unit file for ceph-crash container] ******** 2025-08-29 17:39:01.837014 | orchestrator | Friday 29 August 2025 17:36:49 +0000 (0:00:01.213) 0:09:08.382 ********* 2025-08-29 17:39:01.837020 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.837025 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.837031 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.837036 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.837041 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.837047 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.837052 | orchestrator | 2025-08-29 17:39:01.837058 | orchestrator | TASK [ceph-crash : Start the ceph-crash service] ******************************* 2025-08-29 17:39:01.837063 | orchestrator | Friday 29 August 2025 17:36:50 +0000 (0:00:01.688) 0:09:10.070 ********* 2025-08-29 17:39:01.837068 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.837074 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.837079 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.837085 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.837090 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.837095 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.837101 | orchestrator | 2025-08-29 17:39:01.837106 | orchestrator | RUNNING HANDLER [ceph-handler : Ceph crash handler] **************************** 2025-08-29 17:39:01.837112 | orchestrator | Friday 29 August 2025 17:36:53 +0000 (0:00:03.013) 0:09:13.084 ********* 2025-08-29 17:39:01.837121 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.837126 | orchestrator | 2025-08-29 17:39:01.837132 | orchestrator | RUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ****** 2025-08-29 17:39:01.837137 | orchestrator | Friday 29 August 2025 17:36:55 +0000 (0:00:01.138) 0:09:14.222 ********* 2025-08-29 17:39:01.837143 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.837148 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.837154 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.837159 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837164 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837170 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837175 | orchestrator | 2025-08-29 17:39:01.837181 | orchestrator | RUNNING HANDLER [ceph-handler : Restart the ceph-crash service] **************** 2025-08-29 17:39:01.837186 | orchestrator | Friday 29 August 2025 17:36:55 +0000 (0:00:00.876) 0:09:15.098 ********* 2025-08-29 17:39:01.837192 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:01.837197 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:01.837203 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.837213 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:01.837218 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.837226 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.837232 | orchestrator | 2025-08-29 17:39:01.837238 | orchestrator | RUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] ******* 2025-08-29 17:39:01.837243 | orchestrator | Friday 29 August 2025 17:36:58 +0000 (0:00:02.106) 0:09:17.204 ********* 2025-08-29 17:39:01.837249 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:01.837254 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:01.837259 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:01.837265 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837270 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837276 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837281 | orchestrator | 2025-08-29 17:39:01.837287 | orchestrator | PLAY [Apply role ceph-mds] ***************************************************** 2025-08-29 17:39:01.837292 | orchestrator | 2025-08-29 17:39:01.837298 | orchestrator | TASK [ceph-handler : Include check_running_cluster.yml] ************************ 2025-08-29 17:39:01.837303 | orchestrator | Friday 29 August 2025 17:36:59 +0000 (0:00:01.238) 0:09:18.443 ********* 2025-08-29 17:39:01.837309 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.837314 | orchestrator | 2025-08-29 17:39:01.837320 | orchestrator | TASK [ceph-handler : Include check_running_containers.yml] ********************* 2025-08-29 17:39:01.837325 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:00.721) 0:09:19.164 ********* 2025-08-29 17:39:01.837331 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.837336 | orchestrator | 2025-08-29 17:39:01.837342 | orchestrator | TASK [ceph-handler : Check for a mon container] ******************************** 2025-08-29 17:39:01.837347 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:00.541) 0:09:19.706 ********* 2025-08-29 17:39:01.837353 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837358 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837364 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837369 | orchestrator | 2025-08-29 17:39:01.837375 | orchestrator | TASK [ceph-handler : Check for an osd container] ******************************* 2025-08-29 17:39:01.837380 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:00.302) 0:09:20.009 ********* 2025-08-29 17:39:01.837401 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837407 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837413 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837418 | orchestrator | 2025-08-29 17:39:01.837424 | orchestrator | TASK [ceph-handler : Check for a mds container] ******************************** 2025-08-29 17:39:01.837433 | orchestrator | Friday 29 August 2025 17:37:01 +0000 (0:00:01.000) 0:09:21.009 ********* 2025-08-29 17:39:01.837438 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837444 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837449 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837455 | orchestrator | 2025-08-29 17:39:01.837460 | orchestrator | TASK [ceph-handler : Check for a rgw container] ******************************** 2025-08-29 17:39:01.837466 | orchestrator | Friday 29 August 2025 17:37:02 +0000 (0:00:00.734) 0:09:21.743 ********* 2025-08-29 17:39:01.837471 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837477 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837483 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837488 | orchestrator | 2025-08-29 17:39:01.837494 | orchestrator | TASK [ceph-handler : Check for a mgr container] ******************************** 2025-08-29 17:39:01.837499 | orchestrator | Friday 29 August 2025 17:37:03 +0000 (0:00:01.127) 0:09:22.870 ********* 2025-08-29 17:39:01.837505 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837510 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837516 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837521 | orchestrator | 2025-08-29 17:39:01.837527 | orchestrator | TASK [ceph-handler : Check for a rbd mirror container] ************************* 2025-08-29 17:39:01.837533 | orchestrator | Friday 29 August 2025 17:37:04 +0000 (0:00:00.321) 0:09:23.192 ********* 2025-08-29 17:39:01.837538 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837544 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837549 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837555 | orchestrator | 2025-08-29 17:39:01.837560 | orchestrator | TASK [ceph-handler : Check for a nfs container] ******************************** 2025-08-29 17:39:01.837566 | orchestrator | Friday 29 August 2025 17:37:04 +0000 (0:00:00.577) 0:09:23.770 ********* 2025-08-29 17:39:01.837571 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837577 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837582 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837588 | orchestrator | 2025-08-29 17:39:01.837593 | orchestrator | TASK [ceph-handler : Check for a ceph-crash container] ************************* 2025-08-29 17:39:01.837599 | orchestrator | Friday 29 August 2025 17:37:04 +0000 (0:00:00.318) 0:09:24.088 ********* 2025-08-29 17:39:01.837604 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837610 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837615 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837620 | orchestrator | 2025-08-29 17:39:01.837626 | orchestrator | TASK [ceph-handler : Check for a ceph-exporter container] ********************** 2025-08-29 17:39:01.837631 | orchestrator | Friday 29 August 2025 17:37:05 +0000 (0:00:00.743) 0:09:24.832 ********* 2025-08-29 17:39:01.837637 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837642 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837648 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837653 | orchestrator | 2025-08-29 17:39:01.837659 | orchestrator | TASK [ceph-handler : Include check_socket_non_container.yml] ******************* 2025-08-29 17:39:01.837664 | orchestrator | Friday 29 August 2025 17:37:06 +0000 (0:00:00.685) 0:09:25.518 ********* 2025-08-29 17:39:01.837670 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837675 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837681 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837686 | orchestrator | 2025-08-29 17:39:01.837692 | orchestrator | TASK [ceph-handler : Set_fact handler_mon_status] ****************************** 2025-08-29 17:39:01.837697 | orchestrator | Friday 29 August 2025 17:37:07 +0000 (0:00:00.643) 0:09:26.161 ********* 2025-08-29 17:39:01.837703 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837708 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837714 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837719 | orchestrator | 2025-08-29 17:39:01.837728 | orchestrator | TASK [ceph-handler : Set_fact handler_osd_status] ****************************** 2025-08-29 17:39:01.837736 | orchestrator | Friday 29 August 2025 17:37:07 +0000 (0:00:00.519) 0:09:26.681 ********* 2025-08-29 17:39:01.837745 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837751 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837756 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837762 | orchestrator | 2025-08-29 17:39:01.837767 | orchestrator | TASK [ceph-handler : Set_fact handler_mds_status] ****************************** 2025-08-29 17:39:01.837773 | orchestrator | Friday 29 August 2025 17:37:07 +0000 (0:00:00.431) 0:09:27.112 ********* 2025-08-29 17:39:01.837778 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837784 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837789 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837794 | orchestrator | 2025-08-29 17:39:01.837800 | orchestrator | TASK [ceph-handler : Set_fact handler_rgw_status] ****************************** 2025-08-29 17:39:01.837805 | orchestrator | Friday 29 August 2025 17:37:08 +0000 (0:00:00.350) 0:09:27.463 ********* 2025-08-29 17:39:01.837811 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837816 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837822 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837827 | orchestrator | 2025-08-29 17:39:01.837833 | orchestrator | TASK [ceph-handler : Set_fact handler_nfs_status] ****************************** 2025-08-29 17:39:01.837838 | orchestrator | Friday 29 August 2025 17:37:09 +0000 (0:00:00.688) 0:09:28.152 ********* 2025-08-29 17:39:01.837844 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837849 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837855 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837860 | orchestrator | 2025-08-29 17:39:01.837865 | orchestrator | TASK [ceph-handler : Set_fact handler_rbd_status] ****************************** 2025-08-29 17:39:01.837871 | orchestrator | Friday 29 August 2025 17:37:09 +0000 (0:00:00.313) 0:09:28.465 ********* 2025-08-29 17:39:01.837876 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837882 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837887 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837893 | orchestrator | 2025-08-29 17:39:01.837898 | orchestrator | TASK [ceph-handler : Set_fact handler_mgr_status] ****************************** 2025-08-29 17:39:01.837903 | orchestrator | Friday 29 August 2025 17:37:09 +0000 (0:00:00.451) 0:09:28.917 ********* 2025-08-29 17:39:01.837909 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.837914 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.837920 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.837925 | orchestrator | 2025-08-29 17:39:01.837930 | orchestrator | TASK [ceph-handler : Set_fact handler_crash_status] **************************** 2025-08-29 17:39:01.837936 | orchestrator | Friday 29 August 2025 17:37:10 +0000 (0:00:00.494) 0:09:29.412 ********* 2025-08-29 17:39:01.837941 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837947 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837952 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837957 | orchestrator | 2025-08-29 17:39:01.837963 | orchestrator | TASK [ceph-handler : Set_fact handler_exporter_status] ************************* 2025-08-29 17:39:01.837968 | orchestrator | Friday 29 August 2025 17:37:10 +0000 (0:00:00.684) 0:09:30.096 ********* 2025-08-29 17:39:01.837974 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.837979 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.837985 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.837990 | orchestrator | 2025-08-29 17:39:01.837995 | orchestrator | TASK [ceph-mds : Include create_mds_filesystems.yml] *************************** 2025-08-29 17:39:01.838001 | orchestrator | Friday 29 August 2025 17:37:11 +0000 (0:00:00.491) 0:09:30.587 ********* 2025-08-29 17:39:01.838006 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.838012 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.838049 | orchestrator | included: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3 2025-08-29 17:39:01.838055 | orchestrator | 2025-08-29 17:39:01.838060 | orchestrator | TASK [ceph-facts : Get current default crush rule details] ********************* 2025-08-29 17:39:01.838066 | orchestrator | Friday 29 August 2025 17:37:11 +0000 (0:00:00.486) 0:09:31.073 ********* 2025-08-29 17:39:01.838071 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:39:01.838081 | orchestrator | 2025-08-29 17:39:01.838087 | orchestrator | TASK [ceph-facts : Get current default crush rule name] ************************ 2025-08-29 17:39:01.838092 | orchestrator | Friday 29 August 2025 17:37:13 +0000 (0:00:02.046) 0:09:33.119 ********* 2025-08-29 17:39:01.838099 | orchestrator | skipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]})  2025-08-29 17:39:01.838107 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.838112 | orchestrator | 2025-08-29 17:39:01.838118 | orchestrator | TASK [ceph-mds : Create filesystem pools] ************************************** 2025-08-29 17:39:01.838123 | orchestrator | Friday 29 August 2025 17:37:14 +0000 (0:00:00.293) 0:09:33.413 ********* 2025-08-29 17:39:01.838130 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1}) 2025-08-29 17:39:01.838140 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1}) 2025-08-29 17:39:01.838146 | orchestrator | 2025-08-29 17:39:01.838152 | orchestrator | TASK [ceph-mds : Create ceph filesystem] *************************************** 2025-08-29 17:39:01.838157 | orchestrator | Friday 29 August 2025 17:37:23 +0000 (0:00:08.847) 0:09:42.261 ********* 2025-08-29 17:39:01.838162 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:39:01.838168 | orchestrator | 2025-08-29 17:39:01.838177 | orchestrator | TASK [ceph-mds : Include common.yml] ******************************************* 2025-08-29 17:39:01.838186 | orchestrator | Friday 29 August 2025 17:37:26 +0000 (0:00:03.293) 0:09:45.554 ********* 2025-08-29 17:39:01.838192 | orchestrator | included: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.838197 | orchestrator | 2025-08-29 17:39:01.838203 | orchestrator | TASK [ceph-mds : Create bootstrap-mds and mds directories] ********************* 2025-08-29 17:39:01.838208 | orchestrator | Friday 29 August 2025 17:37:26 +0000 (0:00:00.514) 0:09:46.068 ********* 2025-08-29 17:39:01.838214 | orchestrator | ok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/) 2025-08-29 17:39:01.838219 | orchestrator | changed: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3) 2025-08-29 17:39:01.838224 | orchestrator | ok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/) 2025-08-29 17:39:01.838230 | orchestrator | ok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/) 2025-08-29 17:39:01.838235 | orchestrator | changed: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4) 2025-08-29 17:39:01.838241 | orchestrator | changed: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5) 2025-08-29 17:39:01.838246 | orchestrator | 2025-08-29 17:39:01.838251 | orchestrator | TASK [ceph-mds : Get keys from monitors] *************************************** 2025-08-29 17:39:01.838257 | orchestrator | Friday 29 August 2025 17:37:28 +0000 (0:00:01.333) 0:09:47.401 ********* 2025-08-29 17:39:01.838262 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.838268 | orchestrator | skipping: [testbed-node-3] => (item=None)  2025-08-29 17:39:01.838273 | orchestrator | ok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}] 2025-08-29 17:39:01.838279 | orchestrator | 2025-08-29 17:39:01.838284 | orchestrator | TASK [ceph-mds : Copy ceph key(s) if needed] *********************************** 2025-08-29 17:39:01.838290 | orchestrator | Friday 29 August 2025 17:37:30 +0000 (0:00:02.027) 0:09:49.429 ********* 2025-08-29 17:39:01.838295 | orchestrator | changed: [testbed-node-3] => (item=None) 2025-08-29 17:39:01.838300 | orchestrator | skipping: [testbed-node-3] => (item=None)  2025-08-29 17:39:01.838310 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838316 | orchestrator | changed: [testbed-node-4] => (item=None) 2025-08-29 17:39:01.838321 | orchestrator | skipping: [testbed-node-4] => (item=None)  2025-08-29 17:39:01.838327 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838332 | orchestrator | changed: [testbed-node-5] => (item=None) 2025-08-29 17:39:01.838337 | orchestrator | skipping: [testbed-node-5] => (item=None)  2025-08-29 17:39:01.838343 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838348 | orchestrator | 2025-08-29 17:39:01.838354 | orchestrator | TASK [ceph-mds : Create mds keyring] ******************************************* 2025-08-29 17:39:01.838359 | orchestrator | Friday 29 August 2025 17:37:31 +0000 (0:00:01.097) 0:09:50.527 ********* 2025-08-29 17:39:01.838365 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838370 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838375 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838381 | orchestrator | 2025-08-29 17:39:01.838398 | orchestrator | TASK [ceph-mds : Non_containerized.yml] **************************************** 2025-08-29 17:39:01.838403 | orchestrator | Friday 29 August 2025 17:37:34 +0000 (0:00:02.625) 0:09:53.152 ********* 2025-08-29 17:39:01.838409 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.838414 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.838419 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.838425 | orchestrator | 2025-08-29 17:39:01.838430 | orchestrator | TASK [ceph-mds : Containerized.yml] ******************************************** 2025-08-29 17:39:01.838436 | orchestrator | Friday 29 August 2025 17:37:34 +0000 (0:00:00.579) 0:09:53.732 ********* 2025-08-29 17:39:01.838441 | orchestrator | included: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.838447 | orchestrator | 2025-08-29 17:39:01.838452 | orchestrator | TASK [ceph-mds : Include_tasks systemd.yml] ************************************ 2025-08-29 17:39:01.838458 | orchestrator | Friday 29 August 2025 17:37:35 +0000 (0:00:00.540) 0:09:54.272 ********* 2025-08-29 17:39:01.838463 | orchestrator | included: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.838469 | orchestrator | 2025-08-29 17:39:01.838474 | orchestrator | TASK [ceph-mds : Generate systemd unit file] *********************************** 2025-08-29 17:39:01.838480 | orchestrator | Friday 29 August 2025 17:37:35 +0000 (0:00:00.717) 0:09:54.989 ********* 2025-08-29 17:39:01.838485 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838491 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838496 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838502 | orchestrator | 2025-08-29 17:39:01.838507 | orchestrator | TASK [ceph-mds : Generate systemd ceph-mds target file] ************************ 2025-08-29 17:39:01.838512 | orchestrator | Friday 29 August 2025 17:37:37 +0000 (0:00:01.308) 0:09:56.298 ********* 2025-08-29 17:39:01.838518 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838523 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838529 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838534 | orchestrator | 2025-08-29 17:39:01.838540 | orchestrator | TASK [ceph-mds : Enable ceph-mds.target] *************************************** 2025-08-29 17:39:01.838545 | orchestrator | Friday 29 August 2025 17:37:38 +0000 (0:00:01.186) 0:09:57.485 ********* 2025-08-29 17:39:01.838550 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838556 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838561 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838567 | orchestrator | 2025-08-29 17:39:01.838572 | orchestrator | TASK [ceph-mds : Systemd start mds container] ********************************** 2025-08-29 17:39:01.838577 | orchestrator | Friday 29 August 2025 17:37:40 +0000 (0:00:01.851) 0:09:59.336 ********* 2025-08-29 17:39:01.838583 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838588 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838593 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838599 | orchestrator | 2025-08-29 17:39:01.838607 | orchestrator | TASK [ceph-mds : Wait for mds socket to exist] ********************************* 2025-08-29 17:39:01.838620 | orchestrator | Friday 29 August 2025 17:37:42 +0000 (0:00:02.230) 0:10:01.566 ********* 2025-08-29 17:39:01.838626 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.838631 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.838636 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.838642 | orchestrator | 2025-08-29 17:39:01.838647 | orchestrator | RUNNING HANDLER [ceph-handler : Make tempdir for scripts] ********************** 2025-08-29 17:39:01.838653 | orchestrator | Friday 29 August 2025 17:37:43 +0000 (0:00:01.233) 0:10:02.799 ********* 2025-08-29 17:39:01.838658 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838663 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838669 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838674 | orchestrator | 2025-08-29 17:39:01.838679 | orchestrator | RUNNING HANDLER [ceph-handler : Mdss handler] ********************************** 2025-08-29 17:39:01.838685 | orchestrator | Friday 29 August 2025 17:37:44 +0000 (0:00:00.638) 0:10:03.438 ********* 2025-08-29 17:39:01.838690 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.838696 | orchestrator | 2025-08-29 17:39:01.838701 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ******** 2025-08-29 17:39:01.838707 | orchestrator | Friday 29 August 2025 17:37:45 +0000 (0:00:00.827) 0:10:04.266 ********* 2025-08-29 17:39:01.838712 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.838717 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.838723 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.838728 | orchestrator | 2025-08-29 17:39:01.838734 | orchestrator | RUNNING HANDLER [ceph-handler : Copy mds restart script] *********************** 2025-08-29 17:39:01.838739 | orchestrator | Friday 29 August 2025 17:37:45 +0000 (0:00:00.327) 0:10:04.594 ********* 2025-08-29 17:39:01.838744 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.838750 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.838755 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.838761 | orchestrator | 2025-08-29 17:39:01.838766 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ******************** 2025-08-29 17:39:01.838771 | orchestrator | Friday 29 August 2025 17:37:46 +0000 (0:00:01.194) 0:10:05.789 ********* 2025-08-29 17:39:01.838777 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.838782 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.838788 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.838793 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.838798 | orchestrator | 2025-08-29 17:39:01.838804 | orchestrator | RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] ********* 2025-08-29 17:39:01.838809 | orchestrator | Friday 29 August 2025 17:37:47 +0000 (0:00:01.180) 0:10:06.969 ********* 2025-08-29 17:39:01.838815 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.838820 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.838826 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.838831 | orchestrator | 2025-08-29 17:39:01.838837 | orchestrator | PLAY [Apply role ceph-rgw] ***************************************************** 2025-08-29 17:39:01.838842 | orchestrator | 2025-08-29 17:39:01.838848 | orchestrator | TASK [ceph-handler : Include check_running_cluster.yml] ************************ 2025-08-29 17:39:01.838853 | orchestrator | Friday 29 August 2025 17:37:48 +0000 (0:00:00.545) 0:10:07.514 ********* 2025-08-29 17:39:01.838859 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.838864 | orchestrator | 2025-08-29 17:39:01.838869 | orchestrator | TASK [ceph-handler : Include check_running_containers.yml] ********************* 2025-08-29 17:39:01.838875 | orchestrator | Friday 29 August 2025 17:37:49 +0000 (0:00:00.690) 0:10:08.205 ********* 2025-08-29 17:39:01.838880 | orchestrator | included: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.838889 | orchestrator | 2025-08-29 17:39:01.838895 | orchestrator | TASK [ceph-handler : Check for a mon container] ******************************** 2025-08-29 17:39:01.838900 | orchestrator | Friday 29 August 2025 17:37:49 +0000 (0:00:00.526) 0:10:08.731 ********* 2025-08-29 17:39:01.838906 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.838911 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.838916 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.838922 | orchestrator | 2025-08-29 17:39:01.838927 | orchestrator | TASK [ceph-handler : Check for an osd container] ******************************* 2025-08-29 17:39:01.838932 | orchestrator | Friday 29 August 2025 17:37:49 +0000 (0:00:00.287) 0:10:09.019 ********* 2025-08-29 17:39:01.838938 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.838943 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.838949 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.838954 | orchestrator | 2025-08-29 17:39:01.838959 | orchestrator | TASK [ceph-handler : Check for a mds container] ******************************** 2025-08-29 17:39:01.838965 | orchestrator | Friday 29 August 2025 17:37:50 +0000 (0:00:00.964) 0:10:09.983 ********* 2025-08-29 17:39:01.838970 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.838976 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.838981 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.838986 | orchestrator | 2025-08-29 17:39:01.838992 | orchestrator | TASK [ceph-handler : Check for a rgw container] ******************************** 2025-08-29 17:39:01.838997 | orchestrator | Friday 29 August 2025 17:37:51 +0000 (0:00:00.788) 0:10:10.772 ********* 2025-08-29 17:39:01.839003 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839008 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839014 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839019 | orchestrator | 2025-08-29 17:39:01.839024 | orchestrator | TASK [ceph-handler : Check for a mgr container] ******************************** 2025-08-29 17:39:01.839030 | orchestrator | Friday 29 August 2025 17:37:52 +0000 (0:00:00.766) 0:10:11.538 ********* 2025-08-29 17:39:01.839035 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839041 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839046 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839052 | orchestrator | 2025-08-29 17:39:01.839057 | orchestrator | TASK [ceph-handler : Check for a rbd mirror container] ************************* 2025-08-29 17:39:01.839066 | orchestrator | Friday 29 August 2025 17:37:52 +0000 (0:00:00.292) 0:10:11.831 ********* 2025-08-29 17:39:01.839071 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839079 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839085 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839090 | orchestrator | 2025-08-29 17:39:01.839096 | orchestrator | TASK [ceph-handler : Check for a nfs container] ******************************** 2025-08-29 17:39:01.839101 | orchestrator | Friday 29 August 2025 17:37:53 +0000 (0:00:00.539) 0:10:12.370 ********* 2025-08-29 17:39:01.839107 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839112 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839117 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839123 | orchestrator | 2025-08-29 17:39:01.839128 | orchestrator | TASK [ceph-handler : Check for a ceph-crash container] ************************* 2025-08-29 17:39:01.839134 | orchestrator | Friday 29 August 2025 17:37:53 +0000 (0:00:00.303) 0:10:12.674 ********* 2025-08-29 17:39:01.839139 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839145 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839150 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839156 | orchestrator | 2025-08-29 17:39:01.839161 | orchestrator | TASK [ceph-handler : Check for a ceph-exporter container] ********************** 2025-08-29 17:39:01.839167 | orchestrator | Friday 29 August 2025 17:37:54 +0000 (0:00:00.722) 0:10:13.397 ********* 2025-08-29 17:39:01.839172 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839178 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839183 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839189 | orchestrator | 2025-08-29 17:39:01.839194 | orchestrator | TASK [ceph-handler : Include check_socket_non_container.yml] ******************* 2025-08-29 17:39:01.839203 | orchestrator | Friday 29 August 2025 17:37:54 +0000 (0:00:00.720) 0:10:14.117 ********* 2025-08-29 17:39:01.839209 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839214 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839220 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839225 | orchestrator | 2025-08-29 17:39:01.839231 | orchestrator | TASK [ceph-handler : Set_fact handler_mon_status] ****************************** 2025-08-29 17:39:01.839236 | orchestrator | Friday 29 August 2025 17:37:55 +0000 (0:00:00.599) 0:10:14.717 ********* 2025-08-29 17:39:01.839242 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839247 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839252 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839258 | orchestrator | 2025-08-29 17:39:01.839263 | orchestrator | TASK [ceph-handler : Set_fact handler_osd_status] ****************************** 2025-08-29 17:39:01.839269 | orchestrator | Friday 29 August 2025 17:37:55 +0000 (0:00:00.345) 0:10:15.062 ********* 2025-08-29 17:39:01.839274 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839280 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839285 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839290 | orchestrator | 2025-08-29 17:39:01.839296 | orchestrator | TASK [ceph-handler : Set_fact handler_mds_status] ****************************** 2025-08-29 17:39:01.839301 | orchestrator | Friday 29 August 2025 17:37:56 +0000 (0:00:00.326) 0:10:15.389 ********* 2025-08-29 17:39:01.839307 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839312 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839318 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839323 | orchestrator | 2025-08-29 17:39:01.839329 | orchestrator | TASK [ceph-handler : Set_fact handler_rgw_status] ****************************** 2025-08-29 17:39:01.839334 | orchestrator | Friday 29 August 2025 17:37:56 +0000 (0:00:00.342) 0:10:15.731 ********* 2025-08-29 17:39:01.839340 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839345 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839350 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839356 | orchestrator | 2025-08-29 17:39:01.839361 | orchestrator | TASK [ceph-handler : Set_fact handler_nfs_status] ****************************** 2025-08-29 17:39:01.839367 | orchestrator | Friday 29 August 2025 17:37:57 +0000 (0:00:00.585) 0:10:16.316 ********* 2025-08-29 17:39:01.839372 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839378 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839383 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839401 | orchestrator | 2025-08-29 17:39:01.839406 | orchestrator | TASK [ceph-handler : Set_fact handler_rbd_status] ****************************** 2025-08-29 17:39:01.839412 | orchestrator | Friday 29 August 2025 17:37:57 +0000 (0:00:00.318) 0:10:16.635 ********* 2025-08-29 17:39:01.839417 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839423 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839428 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839433 | orchestrator | 2025-08-29 17:39:01.839439 | orchestrator | TASK [ceph-handler : Set_fact handler_mgr_status] ****************************** 2025-08-29 17:39:01.839444 | orchestrator | Friday 29 August 2025 17:37:57 +0000 (0:00:00.309) 0:10:16.944 ********* 2025-08-29 17:39:01.839450 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839455 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839460 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839466 | orchestrator | 2025-08-29 17:39:01.839471 | orchestrator | TASK [ceph-handler : Set_fact handler_crash_status] **************************** 2025-08-29 17:39:01.839477 | orchestrator | Friday 29 August 2025 17:37:58 +0000 (0:00:00.309) 0:10:17.254 ********* 2025-08-29 17:39:01.839482 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839488 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839493 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839499 | orchestrator | 2025-08-29 17:39:01.839504 | orchestrator | TASK [ceph-handler : Set_fact handler_exporter_status] ************************* 2025-08-29 17:39:01.839509 | orchestrator | Friday 29 August 2025 17:37:58 +0000 (0:00:00.549) 0:10:17.803 ********* 2025-08-29 17:39:01.839519 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.839524 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.839530 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.839535 | orchestrator | 2025-08-29 17:39:01.839540 | orchestrator | TASK [ceph-rgw : Include common.yml] ******************************************* 2025-08-29 17:39:01.839546 | orchestrator | Friday 29 August 2025 17:37:59 +0000 (0:00:00.542) 0:10:18.346 ********* 2025-08-29 17:39:01.839551 | orchestrator | included: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.839557 | orchestrator | 2025-08-29 17:39:01.839562 | orchestrator | TASK [ceph-rgw : Get keys from monitors] *************************************** 2025-08-29 17:39:01.839568 | orchestrator | Friday 29 August 2025 17:38:00 +0000 (0:00:00.840) 0:10:19.187 ********* 2025-08-29 17:39:01.839576 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.839584 | orchestrator | skipping: [testbed-node-3] => (item=None)  2025-08-29 17:39:01.839590 | orchestrator | ok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}] 2025-08-29 17:39:01.839596 | orchestrator | 2025-08-29 17:39:01.839601 | orchestrator | TASK [ceph-rgw : Copy ceph key(s) if needed] *********************************** 2025-08-29 17:39:01.839606 | orchestrator | Friday 29 August 2025 17:38:02 +0000 (0:00:02.320) 0:10:21.507 ********* 2025-08-29 17:39:01.839612 | orchestrator | changed: [testbed-node-3] => (item=None) 2025-08-29 17:39:01.839617 | orchestrator | skipping: [testbed-node-3] => (item=None)  2025-08-29 17:39:01.839623 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.839628 | orchestrator | changed: [testbed-node-4] => (item=None) 2025-08-29 17:39:01.839634 | orchestrator | skipping: [testbed-node-4] => (item=None)  2025-08-29 17:39:01.839639 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.839644 | orchestrator | changed: [testbed-node-5] => (item=None) 2025-08-29 17:39:01.839650 | orchestrator | skipping: [testbed-node-5] => (item=None)  2025-08-29 17:39:01.839655 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.839661 | orchestrator | 2025-08-29 17:39:01.839666 | orchestrator | TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] ********** 2025-08-29 17:39:01.839671 | orchestrator | Friday 29 August 2025 17:38:03 +0000 (0:00:01.201) 0:10:22.708 ********* 2025-08-29 17:39:01.839677 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839682 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.839688 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.839693 | orchestrator | 2025-08-29 17:39:01.839698 | orchestrator | TASK [ceph-rgw : Include_tasks pre_requisite.yml] ****************************** 2025-08-29 17:39:01.839704 | orchestrator | Friday 29 August 2025 17:38:03 +0000 (0:00:00.332) 0:10:23.041 ********* 2025-08-29 17:39:01.839709 | orchestrator | included: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.839715 | orchestrator | 2025-08-29 17:39:01.839720 | orchestrator | TASK [ceph-rgw : Create rados gateway directories] ***************************** 2025-08-29 17:39:01.839725 | orchestrator | Friday 29 August 2025 17:38:04 +0000 (0:00:00.813) 0:10:23.854 ********* 2025-08-29 17:39:01.839731 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.839737 | orchestrator | changed: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.839742 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.839748 | orchestrator | 2025-08-29 17:39:01.839753 | orchestrator | TASK [ceph-rgw : Create rgw keyrings] ****************************************** 2025-08-29 17:39:01.839759 | orchestrator | Friday 29 August 2025 17:38:05 +0000 (0:00:00.912) 0:10:24.767 ********* 2025-08-29 17:39:01.839764 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.839773 | orchestrator | changed: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}] 2025-08-29 17:39:01.839778 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.839784 | orchestrator | changed: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}] 2025-08-29 17:39:01.839789 | orchestrator | changed: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.839795 | orchestrator | changed: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}] 2025-08-29 17:39:01.839800 | orchestrator | 2025-08-29 17:39:01.839806 | orchestrator | TASK [ceph-rgw : Get keys from monitors] *************************************** 2025-08-29 17:39:01.839811 | orchestrator | Friday 29 August 2025 17:38:10 +0000 (0:00:04.678) 0:10:29.445 ********* 2025-08-29 17:39:01.839817 | orchestrator | ok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.839822 | orchestrator | ok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}] 2025-08-29 17:39:01.839827 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.839833 | orchestrator | ok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}] 2025-08-29 17:39:01.839838 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:39:01.839844 | orchestrator | ok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}] 2025-08-29 17:39:01.839849 | orchestrator | 2025-08-29 17:39:01.839854 | orchestrator | TASK [ceph-rgw : Copy ceph key(s) if needed] *********************************** 2025-08-29 17:39:01.839860 | orchestrator | Friday 29 August 2025 17:38:12 +0000 (0:00:02.520) 0:10:31.965 ********* 2025-08-29 17:39:01.839865 | orchestrator | changed: [testbed-node-3] => (item=None) 2025-08-29 17:39:01.839871 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.839876 | orchestrator | changed: [testbed-node-4] => (item=None) 2025-08-29 17:39:01.839882 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.839887 | orchestrator | changed: [testbed-node-5] => (item=None) 2025-08-29 17:39:01.839892 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.839898 | orchestrator | 2025-08-29 17:39:01.839903 | orchestrator | TASK [ceph-rgw : Rgw pool creation tasks] ************************************** 2025-08-29 17:39:01.839909 | orchestrator | Friday 29 August 2025 17:38:14 +0000 (0:00:01.882) 0:10:33.847 ********* 2025-08-29 17:39:01.839919 | orchestrator | included: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3 2025-08-29 17:39:01.839925 | orchestrator | 2025-08-29 17:39:01.839933 | orchestrator | TASK [ceph-rgw : Create ec profile] ******************************************** 2025-08-29 17:39:01.839939 | orchestrator | Friday 29 August 2025 17:38:14 +0000 (0:00:00.295) 0:10:34.143 ********* 2025-08-29 17:39:01.839944 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.839950 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.839956 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.839961 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.839967 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.839972 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.839978 | orchestrator | 2025-08-29 17:39:01.839983 | orchestrator | TASK [ceph-rgw : Set crush rule] *********************************************** 2025-08-29 17:39:01.839989 | orchestrator | Friday 29 August 2025 17:38:15 +0000 (0:00:00.744) 0:10:34.887 ********* 2025-08-29 17:39:01.839999 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.840005 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.840010 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.840016 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.840021 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})  2025-08-29 17:39:01.840027 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.840032 | orchestrator | 2025-08-29 17:39:01.840038 | orchestrator | TASK [ceph-rgw : Create rgw pools] ********************************************* 2025-08-29 17:39:01.840043 | orchestrator | Friday 29 August 2025 17:38:16 +0000 (0:00:00.606) 0:10:35.493 ********* 2025-08-29 17:39:01.840049 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) 2025-08-29 17:39:01.840054 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) 2025-08-29 17:39:01.840060 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) 2025-08-29 17:39:01.840066 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) 2025-08-29 17:39:01.840071 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) 2025-08-29 17:39:01.840076 | orchestrator | 2025-08-29 17:39:01.840082 | orchestrator | TASK [ceph-rgw : Include_tasks openstack-keystone.yml] ************************* 2025-08-29 17:39:01.840087 | orchestrator | Friday 29 August 2025 17:38:47 +0000 (0:00:31.482) 0:11:06.976 ********* 2025-08-29 17:39:01.840093 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.840098 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.840104 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.840109 | orchestrator | 2025-08-29 17:39:01.840114 | orchestrator | TASK [ceph-rgw : Include_tasks start_radosgw.yml] ****************************** 2025-08-29 17:39:01.840120 | orchestrator | Friday 29 August 2025 17:38:48 +0000 (0:00:00.298) 0:11:07.275 ********* 2025-08-29 17:39:01.840125 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.840131 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.840136 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.840141 | orchestrator | 2025-08-29 17:39:01.840147 | orchestrator | TASK [ceph-rgw : Include start_docker_rgw.yml] ********************************* 2025-08-29 17:39:01.840152 | orchestrator | Friday 29 August 2025 17:38:48 +0000 (0:00:00.305) 0:11:07.581 ********* 2025-08-29 17:39:01.840158 | orchestrator | included: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.840163 | orchestrator | 2025-08-29 17:39:01.840168 | orchestrator | TASK [ceph-rgw : Include_task systemd.yml] ************************************* 2025-08-29 17:39:01.840174 | orchestrator | Friday 29 August 2025 17:38:49 +0000 (0:00:00.872) 0:11:08.453 ********* 2025-08-29 17:39:01.840179 | orchestrator | included: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.840185 | orchestrator | 2025-08-29 17:39:01.840191 | orchestrator | TASK [ceph-rgw : Generate systemd unit file] *********************************** 2025-08-29 17:39:01.840196 | orchestrator | Friday 29 August 2025 17:38:49 +0000 (0:00:00.555) 0:11:09.008 ********* 2025-08-29 17:39:01.840208 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.840213 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.840221 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.840227 | orchestrator | 2025-08-29 17:39:01.840232 | orchestrator | TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ******************** 2025-08-29 17:39:01.840238 | orchestrator | Friday 29 August 2025 17:38:51 +0000 (0:00:01.546) 0:11:10.555 ********* 2025-08-29 17:39:01.840243 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.840248 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.840254 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.840259 | orchestrator | 2025-08-29 17:39:01.840265 | orchestrator | TASK [ceph-rgw : Enable ceph-radosgw.target] *********************************** 2025-08-29 17:39:01.840270 | orchestrator | Friday 29 August 2025 17:38:52 +0000 (0:00:01.148) 0:11:11.704 ********* 2025-08-29 17:39:01.840275 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:39:01.840281 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:39:01.840286 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:39:01.840292 | orchestrator | 2025-08-29 17:39:01.840297 | orchestrator | TASK [ceph-rgw : Systemd start rgw container] ********************************** 2025-08-29 17:39:01.840303 | orchestrator | Friday 29 August 2025 17:38:54 +0000 (0:00:01.724) 0:11:13.428 ********* 2025-08-29 17:39:01.840308 | orchestrator | changed: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.840313 | orchestrator | changed: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.840319 | orchestrator | changed: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081}) 2025-08-29 17:39:01.840324 | orchestrator | 2025-08-29 17:39:01.840330 | orchestrator | RUNNING HANDLER [ceph-handler : Make tempdir for scripts] ********************** 2025-08-29 17:39:01.840335 | orchestrator | Friday 29 August 2025 17:38:56 +0000 (0:00:02.612) 0:11:16.041 ********* 2025-08-29 17:39:01.840341 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.840346 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.840351 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.840357 | orchestrator | 2025-08-29 17:39:01.840362 | orchestrator | RUNNING HANDLER [ceph-handler : Rgws handler] ********************************** 2025-08-29 17:39:01.840368 | orchestrator | Friday 29 August 2025 17:38:57 +0000 (0:00:00.334) 0:11:16.376 ********* 2025-08-29 17:39:01.840373 | orchestrator | included: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:39:01.840379 | orchestrator | 2025-08-29 17:39:01.840384 | orchestrator | RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ******** 2025-08-29 17:39:01.840402 | orchestrator | Friday 29 August 2025 17:38:58 +0000 (0:00:00.780) 0:11:17.156 ********* 2025-08-29 17:39:01.840408 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.840413 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.840419 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.840424 | orchestrator | 2025-08-29 17:39:01.840429 | orchestrator | RUNNING HANDLER [ceph-handler : Copy rgw restart script] *********************** 2025-08-29 17:39:01.840435 | orchestrator | Friday 29 August 2025 17:38:58 +0000 (0:00:00.349) 0:11:17.505 ********* 2025-08-29 17:39:01.840440 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.840446 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:39:01.840451 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:39:01.840457 | orchestrator | 2025-08-29 17:39:01.840462 | orchestrator | RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ******************** 2025-08-29 17:39:01.840468 | orchestrator | Friday 29 August 2025 17:38:58 +0000 (0:00:00.342) 0:11:17.848 ********* 2025-08-29 17:39:01.840473 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:39:01.840479 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:39:01.840484 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:39:01.840493 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:39:01.840499 | orchestrator | 2025-08-29 17:39:01.840504 | orchestrator | RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] ********* 2025-08-29 17:39:01.840510 | orchestrator | Friday 29 August 2025 17:38:59 +0000 (0:00:00.853) 0:11:18.701 ********* 2025-08-29 17:39:01.840515 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:39:01.840521 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:39:01.840526 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:39:01.840531 | orchestrator | 2025-08-29 17:39:01.840537 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:39:01.840543 | orchestrator | testbed-node-0 : ok=141  changed=35  unreachable=0 failed=0 skipped=135  rescued=0 ignored=0 2025-08-29 17:39:01.840548 | orchestrator | testbed-node-1 : ok=127  changed=32  unreachable=0 failed=0 skipped=120  rescued=0 ignored=0 2025-08-29 17:39:01.840554 | orchestrator | testbed-node-2 : ok=134  changed=33  unreachable=0 failed=0 skipped=119  rescued=0 ignored=0 2025-08-29 17:39:01.840559 | orchestrator | testbed-node-3 : ok=186  changed=44  unreachable=0 failed=0 skipped=152  rescued=0 ignored=0 2025-08-29 17:39:01.840565 | orchestrator | testbed-node-4 : ok=175  changed=40  unreachable=0 failed=0 skipped=123  rescued=0 ignored=0 2025-08-29 17:39:01.840570 | orchestrator | testbed-node-5 : ok=177  changed=41  unreachable=0 failed=0 skipped=121  rescued=0 ignored=0 2025-08-29 17:39:01.840576 | orchestrator | 2025-08-29 17:39:01.840581 | orchestrator | 2025-08-29 17:39:01.840587 | orchestrator | 2025-08-29 17:39:01.840595 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:39:01.840604 | orchestrator | Friday 29 August 2025 17:38:59 +0000 (0:00:00.235) 0:11:18.937 ********* 2025-08-29 17:39:01.840609 | orchestrator | =============================================================================== 2025-08-29 17:39:01.840615 | orchestrator | ceph-container-common : Pulling Ceph container image ------------------- 79.35s 2025-08-29 17:39:01.840620 | orchestrator | ceph-osd : Use ceph-volume to create osds ------------------------------ 41.39s 2025-08-29 17:39:01.840626 | orchestrator | ceph-rgw : Create rgw pools -------------------------------------------- 31.48s 2025-08-29 17:39:01.840631 | orchestrator | ceph-mgr : Wait for all mgr to be up ----------------------------------- 29.94s 2025-08-29 17:39:01.840637 | orchestrator | ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 22.14s 2025-08-29 17:39:01.840642 | orchestrator | ceph-mon : Set cluster configs ----------------------------------------- 15.08s 2025-08-29 17:39:01.840647 | orchestrator | ceph-osd : Wait for all osd to be up ----------------------------------- 12.71s 2025-08-29 17:39:01.840653 | orchestrator | ceph-mon : Fetch ceph initial keys ------------------------------------- 10.39s 2025-08-29 17:39:01.840658 | orchestrator | ceph-mgr : Create ceph mgr keyring(s) on a mon node --------------------- 9.91s 2025-08-29 17:39:01.840664 | orchestrator | ceph-mds : Create filesystem pools -------------------------------------- 8.85s 2025-08-29 17:39:01.840669 | orchestrator | ceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.29s 2025-08-29 17:39:01.840674 | orchestrator | ceph-config : Create ceph initial directories --------------------------- 6.26s 2025-08-29 17:39:01.840680 | orchestrator | ceph-mgr : Add modules to ceph-mgr -------------------------------------- 5.33s 2025-08-29 17:39:01.840685 | orchestrator | ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 4.81s 2025-08-29 17:39:01.840690 | orchestrator | ceph-rgw : Create rgw keyrings ------------------------------------------ 4.68s 2025-08-29 17:39:01.840696 | orchestrator | ceph-crash : Create client.crash keyring -------------------------------- 4.17s 2025-08-29 17:39:01.840701 | orchestrator | ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created --- 3.95s 2025-08-29 17:39:01.840710 | orchestrator | ceph-osd : Systemd start osd -------------------------------------------- 3.81s 2025-08-29 17:39:01.840715 | orchestrator | ceph-osd : Unset noup flag ---------------------------------------------- 3.60s 2025-08-29 17:39:01.840721 | orchestrator | ceph-config : Generate Ceph file ---------------------------------------- 3.46s 2025-08-29 17:39:04.866178 | orchestrator | 2025-08-29 17:39:04 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:04.866809 | orchestrator | 2025-08-29 17:39:04 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:04.869684 | orchestrator | 2025-08-29 17:39:04 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:04.869721 | orchestrator | 2025-08-29 17:39:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:07.924125 | orchestrator | 2025-08-29 17:39:07 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:07.924840 | orchestrator | 2025-08-29 17:39:07 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:07.925608 | orchestrator | 2025-08-29 17:39:07 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:07.925638 | orchestrator | 2025-08-29 17:39:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:10.974678 | orchestrator | 2025-08-29 17:39:10 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:10.976123 | orchestrator | 2025-08-29 17:39:10 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:10.977673 | orchestrator | 2025-08-29 17:39:10 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:10.977714 | orchestrator | 2025-08-29 17:39:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:14.028966 | orchestrator | 2025-08-29 17:39:14 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:14.030137 | orchestrator | 2025-08-29 17:39:14 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:14.031505 | orchestrator | 2025-08-29 17:39:14 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:14.031557 | orchestrator | 2025-08-29 17:39:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:17.066436 | orchestrator | 2025-08-29 17:39:17 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:17.066879 | orchestrator | 2025-08-29 17:39:17 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:17.068169 | orchestrator | 2025-08-29 17:39:17 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:17.068254 | orchestrator | 2025-08-29 17:39:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:20.114578 | orchestrator | 2025-08-29 17:39:20 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:20.115900 | orchestrator | 2025-08-29 17:39:20 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:20.117600 | orchestrator | 2025-08-29 17:39:20 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:20.117912 | orchestrator | 2025-08-29 17:39:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:23.157184 | orchestrator | 2025-08-29 17:39:23 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:23.159546 | orchestrator | 2025-08-29 17:39:23 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:23.162113 | orchestrator | 2025-08-29 17:39:23 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:23.162579 | orchestrator | 2025-08-29 17:39:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:26.210420 | orchestrator | 2025-08-29 17:39:26 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:26.211563 | orchestrator | 2025-08-29 17:39:26 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:26.214093 | orchestrator | 2025-08-29 17:39:26 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:26.214415 | orchestrator | 2025-08-29 17:39:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:29.258849 | orchestrator | 2025-08-29 17:39:29 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:29.258964 | orchestrator | 2025-08-29 17:39:29 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:29.259561 | orchestrator | 2025-08-29 17:39:29 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:29.260181 | orchestrator | 2025-08-29 17:39:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:32.313633 | orchestrator | 2025-08-29 17:39:32 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:32.317052 | orchestrator | 2025-08-29 17:39:32 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:32.318824 | orchestrator | 2025-08-29 17:39:32 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:32.319447 | orchestrator | 2025-08-29 17:39:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:35.363235 | orchestrator | 2025-08-29 17:39:35 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:35.365072 | orchestrator | 2025-08-29 17:39:35 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:35.367243 | orchestrator | 2025-08-29 17:39:35 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:35.367284 | orchestrator | 2025-08-29 17:39:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:38.416940 | orchestrator | 2025-08-29 17:39:38 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:38.418274 | orchestrator | 2025-08-29 17:39:38 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state STARTED 2025-08-29 17:39:38.421115 | orchestrator | 2025-08-29 17:39:38 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:38.421343 | orchestrator | 2025-08-29 17:39:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:41.467785 | orchestrator | 2025-08-29 17:39:41 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:41.469904 | orchestrator | 2025-08-29 17:39:41 | INFO  | Task de28bb5a-9694-4275-aa73-ed37658d35e7 is in state SUCCESS 2025-08-29 17:39:41.470580 | orchestrator | 2025-08-29 17:39:41.472328 | orchestrator | 2025-08-29 17:39:41.472368 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:39:41.472381 | orchestrator | 2025-08-29 17:39:41.472457 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:39:41.472472 | orchestrator | Friday 29 August 2025 17:36:41 +0000 (0:00:00.261) 0:00:00.261 ********* 2025-08-29 17:39:41.472484 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:41.472497 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:41.472508 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:41.472519 | orchestrator | 2025-08-29 17:39:41.472530 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:39:41.472568 | orchestrator | Friday 29 August 2025 17:36:41 +0000 (0:00:00.293) 0:00:00.555 ********* 2025-08-29 17:39:41.472581 | orchestrator | ok: [testbed-node-0] => (item=enable_opensearch_True) 2025-08-29 17:39:41.472593 | orchestrator | ok: [testbed-node-1] => (item=enable_opensearch_True) 2025-08-29 17:39:41.472618 | orchestrator | ok: [testbed-node-2] => (item=enable_opensearch_True) 2025-08-29 17:39:41.472630 | orchestrator | 2025-08-29 17:39:41.472640 | orchestrator | PLAY [Apply role opensearch] *************************************************** 2025-08-29 17:39:41.472651 | orchestrator | 2025-08-29 17:39:41.472662 | orchestrator | TASK [opensearch : include_tasks] ********************************************** 2025-08-29 17:39:41.472673 | orchestrator | Friday 29 August 2025 17:36:41 +0000 (0:00:00.418) 0:00:00.974 ********* 2025-08-29 17:39:41.472684 | orchestrator | included: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:41.472695 | orchestrator | 2025-08-29 17:39:41.472706 | orchestrator | TASK [opensearch : Setting sysctl values] ************************************** 2025-08-29 17:39:41.472717 | orchestrator | Friday 29 August 2025 17:36:42 +0000 (0:00:00.498) 0:00:01.473 ********* 2025-08-29 17:39:41.472728 | orchestrator | changed: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144}) 2025-08-29 17:39:41.472738 | orchestrator | changed: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144}) 2025-08-29 17:39:41.472749 | orchestrator | changed: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144}) 2025-08-29 17:39:41.472760 | orchestrator | 2025-08-29 17:39:41.472771 | orchestrator | TASK [opensearch : Ensuring config directories exist] ************************** 2025-08-29 17:39:41.472781 | orchestrator | Friday 29 August 2025 17:36:43 +0000 (0:00:00.768) 0:00:02.242 ********* 2025-08-29 17:39:41.472796 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.472812 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.472839 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.472866 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.472881 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.472895 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.472907 | orchestrator | 2025-08-29 17:39:41.472920 | orchestrator | TASK [opensearch : include_tasks] ********************************************** 2025-08-29 17:39:41.472934 | orchestrator | Friday 29 August 2025 17:36:44 +0000 (0:00:01.654) 0:00:03.896 ********* 2025-08-29 17:39:41.472946 | orchestrator | included: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:41.472958 | orchestrator | 2025-08-29 17:39:41.472990 | orchestrator | TASK [service-cert-copy : opensearch | Copying over extra CA certificates] ***** 2025-08-29 17:39:41.473003 | orchestrator | Friday 29 August 2025 17:36:45 +0000 (0:00:00.491) 0:00:04.388 ********* 2025-08-29 17:39:41.473026 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.473045 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.473059 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.473073 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.473096 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.473121 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.473133 | orchestrator | 2025-08-29 17:39:41.473144 | orchestrator | TASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] *** 2025-08-29 17:39:41.473155 | orchestrator | Friday 29 August 2025 17:36:47 +0000 (0:00:02.725) 0:00:07.113 ********* 2025-08-29 17:39:41.473167 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:39:41.473179 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:39:41.473198 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:41.473216 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:39:41.473233 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:39:41.473245 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:41.473256 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:39:41.473268 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:39:41.473349 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:41.473500 | orchestrator | 2025-08-29 17:39:41.473512 | orchestrator | TASK [service-cert-copy : opensearch | Copying over backend internal TLS key] *** 2025-08-29 17:39:41.473523 | orchestrator | Friday 29 August 2025 17:36:49 +0000 (0:00:01.394) 0:00:08.508 ********* 2025-08-29 17:39:41.473544 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:39:41.473564 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:39:41.473577 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:41.473589 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:39:41.473601 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:39:41.473621 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:41.473638 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})  2025-08-29 17:39:41.473655 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})  2025-08-29 17:39:41.473667 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:41.473678 | orchestrator | 2025-08-29 17:39:41.473689 | orchestrator | TASK [opensearch : Copying over config.json files for services] **************** 2025-08-29 17:39:41.473700 | orchestrator | Friday 29 August 2025 17:36:50 +0000 (0:00:01.028) 0:00:09.536 ********* 2025-08-29 17:39:41.473712 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.473724 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.473750 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.473774 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.473788 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.473800 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.473821 | orchestrator | 2025-08-29 17:39:41.473833 | orchestrator | TASK [opensearch : Copying over opensearch service config file] **************** 2025-08-29 17:39:41.473844 | orchestrator | Friday 29 August 2025 17:36:52 +0000 (0:00:02.408) 0:00:11.945 ********* 2025-08-29 17:39:41.473855 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:41.473866 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:41.473877 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:41.473888 | orchestrator | 2025-08-29 17:39:41.473899 | orchestrator | TASK [opensearch : Copying over opensearch-dashboards config file] ************* 2025-08-29 17:39:41.473910 | orchestrator | Friday 29 August 2025 17:36:56 +0000 (0:00:03.379) 0:00:15.324 ********* 2025-08-29 17:39:41.473921 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:41.473932 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:41.473942 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:41.473953 | orchestrator | 2025-08-29 17:39:41.473964 | orchestrator | TASK [opensearch : Check opensearch containers] ******************************** 2025-08-29 17:39:41.473975 | orchestrator | Friday 29 August 2025 17:36:57 +0000 (0:00:01.733) 0:00:17.058 ********* 2025-08-29 17:39:41.473995 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.474012 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.474073 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/opensearch:2.19.2.20250711', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) 2025-08-29 17:39:41.474093 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.474113 | orchestrator | changed: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.474132 | orchestrator | changed: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) 2025-08-29 17:39:41.474145 | orchestrator | 2025-08-29 17:39:41.474157 | orchestrator | TASK [opensearch : include_tasks] ********************************************** 2025-08-29 17:39:41.474169 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:02.328) 0:00:19.387 ********* 2025-08-29 17:39:41.474181 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:41.474194 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:41.474206 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:41.474218 | orchestrator | 2025-08-29 17:39:41.474230 | orchestrator | TASK [opensearch : Flush handlers] ********************************************* 2025-08-29 17:39:41.474249 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:00.280) 0:00:19.667 ********* 2025-08-29 17:39:41.474261 | orchestrator | 2025-08-29 17:39:41.474273 | orchestrator | TASK [opensearch : Flush handlers] ********************************************* 2025-08-29 17:39:41.474284 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:00.062) 0:00:19.730 ********* 2025-08-29 17:39:41.474296 | orchestrator | 2025-08-29 17:39:41.474308 | orchestrator | TASK [opensearch : Flush handlers] ********************************************* 2025-08-29 17:39:41.474320 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:00.061) 0:00:19.792 ********* 2025-08-29 17:39:41.474332 | orchestrator | 2025-08-29 17:39:41.474344 | orchestrator | RUNNING HANDLER [opensearch : Disable shard allocation] ************************ 2025-08-29 17:39:41.474356 | orchestrator | Friday 29 August 2025 17:37:00 +0000 (0:00:00.236) 0:00:20.028 ********* 2025-08-29 17:39:41.474368 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:41.474381 | orchestrator | 2025-08-29 17:39:41.474412 | orchestrator | RUNNING HANDLER [opensearch : Perform a flush] ********************************* 2025-08-29 17:39:41.474425 | orchestrator | Friday 29 August 2025 17:37:01 +0000 (0:00:00.214) 0:00:20.242 ********* 2025-08-29 17:39:41.474437 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:41.474450 | orchestrator | 2025-08-29 17:39:41.474462 | orchestrator | RUNNING HANDLER [opensearch : Restart opensearch container] ******************** 2025-08-29 17:39:41.474473 | orchestrator | Friday 29 August 2025 17:37:01 +0000 (0:00:00.201) 0:00:20.444 ********* 2025-08-29 17:39:41.474484 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:41.474495 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:41.474506 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:41.474516 | orchestrator | 2025-08-29 17:39:41.474527 | orchestrator | RUNNING HANDLER [opensearch : Restart opensearch-dashboards container] ********* 2025-08-29 17:39:41.474538 | orchestrator | Friday 29 August 2025 17:38:08 +0000 (0:01:07.018) 0:01:27.463 ********* 2025-08-29 17:39:41.474549 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:41.474560 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:41.474570 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:41.474581 | orchestrator | 2025-08-29 17:39:41.474592 | orchestrator | TASK [opensearch : include_tasks] ********************************************** 2025-08-29 17:39:41.474602 | orchestrator | Friday 29 August 2025 17:39:28 +0000 (0:01:20.432) 0:02:47.895 ********* 2025-08-29 17:39:41.474613 | orchestrator | included: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:41.474624 | orchestrator | 2025-08-29 17:39:41.474635 | orchestrator | TASK [opensearch : Wait for OpenSearch to become ready] ************************ 2025-08-29 17:39:41.474646 | orchestrator | Friday 29 August 2025 17:39:29 +0000 (0:00:00.663) 0:02:48.559 ********* 2025-08-29 17:39:41.474657 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:41.474668 | orchestrator | 2025-08-29 17:39:41.474678 | orchestrator | TASK [opensearch : Check if a log retention policy exists] ********************* 2025-08-29 17:39:41.474689 | orchestrator | Friday 29 August 2025 17:39:31 +0000 (0:00:02.517) 0:02:51.077 ********* 2025-08-29 17:39:41.474700 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:41.474711 | orchestrator | 2025-08-29 17:39:41.474722 | orchestrator | TASK [opensearch : Create new log retention policy] **************************** 2025-08-29 17:39:41.474732 | orchestrator | Friday 29 August 2025 17:39:34 +0000 (0:00:02.241) 0:02:53.318 ********* 2025-08-29 17:39:41.474743 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:41.474754 | orchestrator | 2025-08-29 17:39:41.474765 | orchestrator | TASK [opensearch : Apply retention policy to existing indices] ***************** 2025-08-29 17:39:41.474776 | orchestrator | Friday 29 August 2025 17:39:37 +0000 (0:00:02.958) 0:02:56.277 ********* 2025-08-29 17:39:41.474787 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:41.474797 | orchestrator | 2025-08-29 17:39:41.474814 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:39:41.474827 | orchestrator | testbed-node-0 : ok=18  changed=11  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:39:41.474846 | orchestrator | testbed-node-1 : ok=14  changed=9  unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 17:39:41.474858 | orchestrator | testbed-node-2 : ok=14  changed=9  unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 17:39:41.474868 | orchestrator | 2025-08-29 17:39:41.474879 | orchestrator | 2025-08-29 17:39:41.474890 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:39:41.474906 | orchestrator | Friday 29 August 2025 17:39:39 +0000 (0:00:02.549) 0:02:58.827 ********* 2025-08-29 17:39:41.474917 | orchestrator | =============================================================================== 2025-08-29 17:39:41.474928 | orchestrator | opensearch : Restart opensearch-dashboards container ------------------- 80.43s 2025-08-29 17:39:41.474938 | orchestrator | opensearch : Restart opensearch container ------------------------------ 67.02s 2025-08-29 17:39:41.474949 | orchestrator | opensearch : Copying over opensearch service config file ---------------- 3.38s 2025-08-29 17:39:41.474960 | orchestrator | opensearch : Create new log retention policy ---------------------------- 2.96s 2025-08-29 17:39:41.474971 | orchestrator | service-cert-copy : opensearch | Copying over extra CA certificates ----- 2.73s 2025-08-29 17:39:41.474981 | orchestrator | opensearch : Apply retention policy to existing indices ----------------- 2.55s 2025-08-29 17:39:41.474992 | orchestrator | opensearch : Wait for OpenSearch to become ready ------------------------ 2.52s 2025-08-29 17:39:41.475003 | orchestrator | opensearch : Copying over config.json files for services ---------------- 2.41s 2025-08-29 17:39:41.475014 | orchestrator | opensearch : Check opensearch containers -------------------------------- 2.33s 2025-08-29 17:39:41.475024 | orchestrator | opensearch : Check if a log retention policy exists --------------------- 2.24s 2025-08-29 17:39:41.475035 | orchestrator | opensearch : Copying over opensearch-dashboards config file ------------- 1.73s 2025-08-29 17:39:41.475046 | orchestrator | opensearch : Ensuring config directories exist -------------------------- 1.65s 2025-08-29 17:39:41.475057 | orchestrator | service-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.39s 2025-08-29 17:39:41.475067 | orchestrator | service-cert-copy : opensearch | Copying over backend internal TLS key --- 1.03s 2025-08-29 17:39:41.475078 | orchestrator | opensearch : Setting sysctl values -------------------------------------- 0.77s 2025-08-29 17:39:41.475089 | orchestrator | opensearch : include_tasks ---------------------------------------------- 0.66s 2025-08-29 17:39:41.475100 | orchestrator | opensearch : include_tasks ---------------------------------------------- 0.50s 2025-08-29 17:39:41.475110 | orchestrator | opensearch : include_tasks ---------------------------------------------- 0.49s 2025-08-29 17:39:41.475121 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.42s 2025-08-29 17:39:41.475132 | orchestrator | opensearch : Flush handlers --------------------------------------------- 0.36s 2025-08-29 17:39:41.475143 | orchestrator | 2025-08-29 17:39:41 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:41.475154 | orchestrator | 2025-08-29 17:39:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:44.518239 | orchestrator | 2025-08-29 17:39:44 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:44.518930 | orchestrator | 2025-08-29 17:39:44 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:44.518955 | orchestrator | 2025-08-29 17:39:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:47.566985 | orchestrator | 2025-08-29 17:39:47 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:47.568996 | orchestrator | 2025-08-29 17:39:47 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:47.569044 | orchestrator | 2025-08-29 17:39:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:50.619191 | orchestrator | 2025-08-29 17:39:50 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:50.621236 | orchestrator | 2025-08-29 17:39:50 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:50.621564 | orchestrator | 2025-08-29 17:39:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:53.672284 | orchestrator | 2025-08-29 17:39:53 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state STARTED 2025-08-29 17:39:53.673808 | orchestrator | 2025-08-29 17:39:53 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:53.673845 | orchestrator | 2025-08-29 17:39:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:56.723973 | orchestrator | 2025-08-29 17:39:56 | INFO  | Task e4b7af23-426c-4cd2-9f25-339b63b65dfe is in state SUCCESS 2025-08-29 17:39:56.725177 | orchestrator | 2025-08-29 17:39:56.725232 | orchestrator | 2025-08-29 17:39:56.725241 | orchestrator | PLAY [Set kolla_action_mariadb] ************************************************ 2025-08-29 17:39:56.725247 | orchestrator | 2025-08-29 17:39:56.725253 | orchestrator | TASK [Inform the user about the following task] ******************************** 2025-08-29 17:39:56.725260 | orchestrator | Friday 29 August 2025 17:36:40 +0000 (0:00:00.102) 0:00:00.102 ********* 2025-08-29 17:39:56.725266 | orchestrator | ok: [localhost] => { 2025-08-29 17:39:56.725273 | orchestrator |  "msg": "The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine." 2025-08-29 17:39:56.725279 | orchestrator | } 2025-08-29 17:39:56.725285 | orchestrator | 2025-08-29 17:39:56.725291 | orchestrator | TASK [Check MariaDB service] *************************************************** 2025-08-29 17:39:56.725297 | orchestrator | Friday 29 August 2025 17:36:40 +0000 (0:00:00.048) 0:00:00.151 ********* 2025-08-29 17:39:56.725363 | orchestrator | fatal: [localhost]: FAILED! => {"changed": false, "elapsed": 2, "msg": "Timeout when waiting for search string MariaDB in 192.168.16.9:3306"} 2025-08-29 17:39:56.725373 | orchestrator | ...ignoring 2025-08-29 17:39:56.725678 | orchestrator | 2025-08-29 17:39:56.725700 | orchestrator | TASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ******** 2025-08-29 17:39:56.725712 | orchestrator | Friday 29 August 2025 17:36:43 +0000 (0:00:02.797) 0:00:02.949 ********* 2025-08-29 17:39:56.725723 | orchestrator | skipping: [localhost] 2025-08-29 17:39:56.725735 | orchestrator | 2025-08-29 17:39:56.725746 | orchestrator | TASK [Set kolla_action_mariadb = kolla_action_ng] ****************************** 2025-08-29 17:39:56.725757 | orchestrator | Friday 29 August 2025 17:36:43 +0000 (0:00:00.060) 0:00:03.009 ********* 2025-08-29 17:39:56.725768 | orchestrator | ok: [localhost] 2025-08-29 17:39:56.725779 | orchestrator | 2025-08-29 17:39:56.725790 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:39:56.725801 | orchestrator | 2025-08-29 17:39:56.725812 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:39:56.725824 | orchestrator | Friday 29 August 2025 17:36:43 +0000 (0:00:00.154) 0:00:03.164 ********* 2025-08-29 17:39:56.725835 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.725846 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.725857 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.725868 | orchestrator | 2025-08-29 17:39:56.725879 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:39:56.725891 | orchestrator | Friday 29 August 2025 17:36:44 +0000 (0:00:00.306) 0:00:03.470 ********* 2025-08-29 17:39:56.725902 | orchestrator | ok: [testbed-node-0] => (item=enable_mariadb_True) 2025-08-29 17:39:56.725914 | orchestrator | ok: [testbed-node-1] => (item=enable_mariadb_True) 2025-08-29 17:39:56.725925 | orchestrator | ok: [testbed-node-2] => (item=enable_mariadb_True) 2025-08-29 17:39:56.725935 | orchestrator | 2025-08-29 17:39:56.725946 | orchestrator | PLAY [Apply role mariadb] ****************************************************** 2025-08-29 17:39:56.725958 | orchestrator | 2025-08-29 17:39:56.725995 | orchestrator | TASK [mariadb : Group MariaDB hosts based on shards] *************************** 2025-08-29 17:39:56.726007 | orchestrator | Friday 29 August 2025 17:36:44 +0000 (0:00:00.712) 0:00:04.183 ********* 2025-08-29 17:39:56.726078 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 17:39:56.726093 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-1) 2025-08-29 17:39:56.726105 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-2) 2025-08-29 17:39:56.726116 | orchestrator | 2025-08-29 17:39:56.726127 | orchestrator | TASK [mariadb : include_tasks] ************************************************* 2025-08-29 17:39:56.726139 | orchestrator | Friday 29 August 2025 17:36:45 +0000 (0:00:00.400) 0:00:04.584 ********* 2025-08-29 17:39:56.726150 | orchestrator | included: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:56.726163 | orchestrator | 2025-08-29 17:39:56.726208 | orchestrator | TASK [mariadb : Ensuring config directories exist] ***************************** 2025-08-29 17:39:56.726220 | orchestrator | Friday 29 August 2025 17:36:46 +0000 (0:00:00.678) 0:00:05.262 ********* 2025-08-29 17:39:56.726256 | orchestrator | changed: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.726283 | orchestrator | changed: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.726310 | orchestrator | changed: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.726324 | orchestrator | 2025-08-29 17:39:56.726347 | orchestrator | TASK [mariadb : Ensuring database backup config directory exists] ************** 2025-08-29 17:39:56.726360 | orchestrator | Friday 29 August 2025 17:36:49 +0000 (0:00:03.524) 0:00:08.787 ********* 2025-08-29 17:39:56.726373 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.726386 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.726440 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.726454 | orchestrator | 2025-08-29 17:39:56.726467 | orchestrator | TASK [mariadb : Copying over my.cnf for mariabackup] *************************** 2025-08-29 17:39:56.726480 | orchestrator | Friday 29 August 2025 17:36:50 +0000 (0:00:00.771) 0:00:09.558 ********* 2025-08-29 17:39:56.726492 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.726504 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.726517 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.726529 | orchestrator | 2025-08-29 17:39:56.726541 | orchestrator | TASK [mariadb : Copying over config.json files for services] ******************* 2025-08-29 17:39:56.726561 | orchestrator | Friday 29 August 2025 17:36:51 +0000 (0:00:01.463) 0:00:11.022 ********* 2025-08-29 17:39:56.726576 | orchestrator | changed: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.726608 | orchestrator | changed: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.726629 | orchestrator | changed: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.726651 | orchestrator | 2025-08-29 17:39:56.726664 | orchestrator | TASK [mariadb : Copying over config.json files for mariabackup] **************** 2025-08-29 17:39:56.726676 | orchestrator | Friday 29 August 2025 17:36:55 +0000 (0:00:03.747) 0:00:14.770 ********* 2025-08-29 17:39:56.726688 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.726699 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.726710 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.726721 | orchestrator | 2025-08-29 17:39:56.726732 | orchestrator | TASK [mariadb : Copying over galera.cnf] *************************************** 2025-08-29 17:39:56.726743 | orchestrator | Friday 29 August 2025 17:36:56 +0000 (0:00:01.230) 0:00:16.000 ********* 2025-08-29 17:39:56.726753 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.726764 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:56.726776 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:56.726787 | orchestrator | 2025-08-29 17:39:56.726797 | orchestrator | TASK [mariadb : include_tasks] ************************************************* 2025-08-29 17:39:56.726808 | orchestrator | Friday 29 August 2025 17:37:01 +0000 (0:00:04.481) 0:00:20.482 ********* 2025-08-29 17:39:56.726819 | orchestrator | included: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:56.726830 | orchestrator | 2025-08-29 17:39:56.726841 | orchestrator | TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ******** 2025-08-29 17:39:56.726852 | orchestrator | Friday 29 August 2025 17:37:01 +0000 (0:00:00.581) 0:00:21.063 ********* 2025-08-29 17:39:56.726878 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.726899 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.726911 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.726924 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.726944 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.726963 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.726974 | orchestrator | 2025-08-29 17:39:56.727112 | orchestrator | TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] *** 2025-08-29 17:39:56.727136 | orchestrator | Friday 29 August 2025 17:37:05 +0000 (0:00:03.749) 0:00:24.813 ********* 2025-08-29 17:39:56.727149 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.727162 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.727183 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.727203 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.727216 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.727228 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.727238 | orchestrator | 2025-08-29 17:39:56.727249 | orchestrator | TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] ***** 2025-08-29 17:39:56.727261 | orchestrator | Friday 29 August 2025 17:37:09 +0000 (0:00:03.534) 0:00:28.347 ********* 2025-08-29 17:39:56.727324 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.727347 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.727365 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.727378 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.727390 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})  2025-08-29 17:39:56.727439 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.727451 | orchestrator | 2025-08-29 17:39:56.727463 | orchestrator | TASK [mariadb : Check mariadb containers] ************************************** 2025-08-29 17:39:56.727474 | orchestrator | Friday 29 August 2025 17:37:12 +0000 (0:00:03.164) 0:00:31.512 ********* 2025-08-29 17:39:56.727501 | orchestrator | changed: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.727516 | orchestrator | changed: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.727550 | orchestrator | changed: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) 2025-08-29 17:39:56.727564 | orchestrator | 2025-08-29 17:39:56.727577 | orchestrator | TASK [mariadb : Create MariaDB volume] ***************************************** 2025-08-29 17:39:56.727588 | orchestrator | Friday 29 August 2025 17:37:14 +0000 (0:00:02.591) 0:00:34.103 ********* 2025-08-29 17:39:56.727599 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.727610 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:56.727621 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:56.727632 | orchestrator | 2025-08-29 17:39:56.727643 | orchestrator | TASK [mariadb : Divide hosts by their MariaDB volume availability] ************* 2025-08-29 17:39:56.727654 | orchestrator | Friday 29 August 2025 17:37:15 +0000 (0:00:01.085) 0:00:35.188 ********* 2025-08-29 17:39:56.727665 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.727676 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.727687 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.727698 | orchestrator | 2025-08-29 17:39:56.727708 | orchestrator | TASK [mariadb : Establish whether the cluster has already existed] ************* 2025-08-29 17:39:56.727719 | orchestrator | Friday 29 August 2025 17:37:16 +0000 (0:00:00.322) 0:00:35.511 ********* 2025-08-29 17:39:56.727731 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.727742 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.727753 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.727764 | orchestrator | 2025-08-29 17:39:56.727775 | orchestrator | TASK [mariadb : Check MariaDB service port liveness] *************************** 2025-08-29 17:39:56.727787 | orchestrator | Friday 29 August 2025 17:37:16 +0000 (0:00:00.411) 0:00:35.922 ********* 2025-08-29 17:39:56.727799 | orchestrator | fatal: [testbed-node-1]: FAILED! => {"changed": false, "elapsed": 10, "msg": "Timeout when waiting for search string MariaDB in 192.168.16.11:3306"} 2025-08-29 17:39:56.727811 | orchestrator | ...ignoring 2025-08-29 17:39:56.727822 | orchestrator | fatal: [testbed-node-0]: FAILED! => {"changed": false, "elapsed": 10, "msg": "Timeout when waiting for search string MariaDB in 192.168.16.10:3306"} 2025-08-29 17:39:56.727833 | orchestrator | ...ignoring 2025-08-29 17:39:56.727844 | orchestrator | fatal: [testbed-node-2]: FAILED! => {"changed": false, "elapsed": 10, "msg": "Timeout when waiting for search string MariaDB in 192.168.16.12:3306"} 2025-08-29 17:39:56.727862 | orchestrator | ...ignoring 2025-08-29 17:39:56.727874 | orchestrator | 2025-08-29 17:39:56.727885 | orchestrator | TASK [mariadb : Divide hosts by their MariaDB service port liveness] *********** 2025-08-29 17:39:56.727896 | orchestrator | Friday 29 August 2025 17:37:27 +0000 (0:00:10.839) 0:00:46.761 ********* 2025-08-29 17:39:56.727906 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.727918 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.727928 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.727939 | orchestrator | 2025-08-29 17:39:56.727950 | orchestrator | TASK [mariadb : Fail on existing but stopped cluster] ************************** 2025-08-29 17:39:56.727962 | orchestrator | Friday 29 August 2025 17:37:28 +0000 (0:00:00.658) 0:00:47.420 ********* 2025-08-29 17:39:56.727973 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.727984 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.727995 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.728006 | orchestrator | 2025-08-29 17:39:56.728017 | orchestrator | TASK [mariadb : Check MariaDB service WSREP sync status] *********************** 2025-08-29 17:39:56.728028 | orchestrator | Friday 29 August 2025 17:37:28 +0000 (0:00:00.416) 0:00:47.837 ********* 2025-08-29 17:39:56.728039 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.728050 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.728061 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.728072 | orchestrator | 2025-08-29 17:39:56.728083 | orchestrator | TASK [mariadb : Extract MariaDB service WSREP sync status] ********************* 2025-08-29 17:39:56.728094 | orchestrator | Friday 29 August 2025 17:37:29 +0000 (0:00:00.401) 0:00:48.239 ********* 2025-08-29 17:39:56.728105 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.728116 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.728127 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.728138 | orchestrator | 2025-08-29 17:39:56.728149 | orchestrator | TASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] ******* 2025-08-29 17:39:56.728166 | orchestrator | Friday 29 August 2025 17:37:29 +0000 (0:00:00.414) 0:00:48.653 ********* 2025-08-29 17:39:56.728178 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.728189 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.728200 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.728211 | orchestrator | 2025-08-29 17:39:56.728222 | orchestrator | TASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] *** 2025-08-29 17:39:56.728233 | orchestrator | Friday 29 August 2025 17:37:30 +0000 (0:00:00.874) 0:00:49.527 ********* 2025-08-29 17:39:56.728244 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.728255 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.728266 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.728276 | orchestrator | 2025-08-29 17:39:56.728287 | orchestrator | TASK [mariadb : include_tasks] ************************************************* 2025-08-29 17:39:56.728298 | orchestrator | Friday 29 August 2025 17:37:30 +0000 (0:00:00.463) 0:00:49.991 ********* 2025-08-29 17:39:56.728309 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.728320 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.728336 | orchestrator | included: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0 2025-08-29 17:39:56.728347 | orchestrator | 2025-08-29 17:39:56.728358 | orchestrator | TASK [mariadb : Running MariaDB bootstrap container] *************************** 2025-08-29 17:39:56.728369 | orchestrator | Friday 29 August 2025 17:37:31 +0000 (0:00:00.394) 0:00:50.385 ********* 2025-08-29 17:39:56.728380 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.728391 | orchestrator | 2025-08-29 17:39:56.728626 | orchestrator | TASK [mariadb : Store bootstrap host name into facts] ************************** 2025-08-29 17:39:56.728637 | orchestrator | Friday 29 August 2025 17:37:41 +0000 (0:00:10.521) 0:01:00.907 ********* 2025-08-29 17:39:56.728648 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.728660 | orchestrator | 2025-08-29 17:39:56.728671 | orchestrator | TASK [mariadb : include_tasks] ************************************************* 2025-08-29 17:39:56.728680 | orchestrator | Friday 29 August 2025 17:37:41 +0000 (0:00:00.130) 0:01:01.038 ********* 2025-08-29 17:39:56.728700 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.728710 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.728794 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.728805 | orchestrator | 2025-08-29 17:39:56.728815 | orchestrator | RUNNING HANDLER [mariadb : Starting first MariaDB container] ******************* 2025-08-29 17:39:56.728825 | orchestrator | Friday 29 August 2025 17:37:42 +0000 (0:00:00.983) 0:01:02.021 ********* 2025-08-29 17:39:56.728834 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.728844 | orchestrator | 2025-08-29 17:39:56.728854 | orchestrator | RUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] ******* 2025-08-29 17:39:56.728864 | orchestrator | Friday 29 August 2025 17:37:50 +0000 (0:00:07.863) 0:01:09.885 ********* 2025-08-29 17:39:56.728874 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.728884 | orchestrator | 2025-08-29 17:39:56.728894 | orchestrator | RUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] ******* 2025-08-29 17:39:56.728903 | orchestrator | Friday 29 August 2025 17:37:52 +0000 (0:00:01.653) 0:01:11.539 ********* 2025-08-29 17:39:56.728913 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.728923 | orchestrator | 2025-08-29 17:39:56.728933 | orchestrator | RUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] *** 2025-08-29 17:39:56.728942 | orchestrator | Friday 29 August 2025 17:37:54 +0000 (0:00:02.473) 0:01:14.012 ********* 2025-08-29 17:39:56.728952 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.728962 | orchestrator | 2025-08-29 17:39:56.728972 | orchestrator | RUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ******** 2025-08-29 17:39:56.728982 | orchestrator | Friday 29 August 2025 17:37:54 +0000 (0:00:00.123) 0:01:14.136 ********* 2025-08-29 17:39:56.728991 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.729001 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.729011 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.729020 | orchestrator | 2025-08-29 17:39:56.729030 | orchestrator | RUNNING HANDLER [mariadb : Start MariaDB on new nodes] ************************* 2025-08-29 17:39:56.729040 | orchestrator | Friday 29 August 2025 17:37:55 +0000 (0:00:00.524) 0:01:14.660 ********* 2025-08-29 17:39:56.729049 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.729059 | orchestrator | [WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart 2025-08-29 17:39:56.729069 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:56.729079 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:56.729088 | orchestrator | 2025-08-29 17:39:56.729098 | orchestrator | PLAY [Restart mariadb services] ************************************************ 2025-08-29 17:39:56.729108 | orchestrator | skipping: no hosts matched 2025-08-29 17:39:56.729118 | orchestrator | 2025-08-29 17:39:56.729127 | orchestrator | PLAY [Start mariadb services] ************************************************** 2025-08-29 17:39:56.729137 | orchestrator | 2025-08-29 17:39:56.729147 | orchestrator | TASK [mariadb : Restart MariaDB container] ************************************* 2025-08-29 17:39:56.729157 | orchestrator | Friday 29 August 2025 17:37:55 +0000 (0:00:00.351) 0:01:15.012 ********* 2025-08-29 17:39:56.729166 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:39:56.729176 | orchestrator | 2025-08-29 17:39:56.729186 | orchestrator | TASK [mariadb : Wait for MariaDB service port liveness] ************************ 2025-08-29 17:39:56.729195 | orchestrator | Friday 29 August 2025 17:38:15 +0000 (0:00:19.413) 0:01:34.425 ********* 2025-08-29 17:39:56.729205 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.729215 | orchestrator | 2025-08-29 17:39:56.729225 | orchestrator | TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************ 2025-08-29 17:39:56.729234 | orchestrator | Friday 29 August 2025 17:38:35 +0000 (0:00:20.625) 0:01:55.050 ********* 2025-08-29 17:39:56.729244 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.729254 | orchestrator | 2025-08-29 17:39:56.729264 | orchestrator | PLAY [Start mariadb services] ************************************************** 2025-08-29 17:39:56.729274 | orchestrator | 2025-08-29 17:39:56.729284 | orchestrator | TASK [mariadb : Restart MariaDB container] ************************************* 2025-08-29 17:39:56.729293 | orchestrator | Friday 29 August 2025 17:38:38 +0000 (0:00:02.507) 0:01:57.557 ********* 2025-08-29 17:39:56.729312 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:39:56.729322 | orchestrator | 2025-08-29 17:39:56.729332 | orchestrator | TASK [mariadb : Wait for MariaDB service port liveness] ************************ 2025-08-29 17:39:56.729351 | orchestrator | Friday 29 August 2025 17:39:04 +0000 (0:00:25.768) 0:02:23.325 ********* 2025-08-29 17:39:56.729361 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.729371 | orchestrator | 2025-08-29 17:39:56.729381 | orchestrator | TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************ 2025-08-29 17:39:56.729391 | orchestrator | Friday 29 August 2025 17:39:19 +0000 (0:00:15.559) 0:02:38.884 ********* 2025-08-29 17:39:56.729425 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.729436 | orchestrator | 2025-08-29 17:39:56.729447 | orchestrator | PLAY [Restart bootstrap mariadb service] *************************************** 2025-08-29 17:39:56.729458 | orchestrator | 2025-08-29 17:39:56.729469 | orchestrator | TASK [mariadb : Restart MariaDB container] ************************************* 2025-08-29 17:39:56.729480 | orchestrator | Friday 29 August 2025 17:39:22 +0000 (0:00:02.722) 0:02:41.607 ********* 2025-08-29 17:39:56.729491 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.729503 | orchestrator | 2025-08-29 17:39:56.729514 | orchestrator | TASK [mariadb : Wait for MariaDB service port liveness] ************************ 2025-08-29 17:39:56.729525 | orchestrator | Friday 29 August 2025 17:39:34 +0000 (0:00:11.861) 0:02:53.468 ********* 2025-08-29 17:39:56.729542 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.729553 | orchestrator | 2025-08-29 17:39:56.729564 | orchestrator | TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************ 2025-08-29 17:39:56.729575 | orchestrator | Friday 29 August 2025 17:39:38 +0000 (0:00:04.618) 0:02:58.087 ********* 2025-08-29 17:39:56.729586 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.729597 | orchestrator | 2025-08-29 17:39:56.729608 | orchestrator | PLAY [Apply mariadb post-configuration] **************************************** 2025-08-29 17:39:56.729618 | orchestrator | 2025-08-29 17:39:56.729629 | orchestrator | TASK [Include mariadb post-deploy.yml] ***************************************** 2025-08-29 17:39:56.729640 | orchestrator | Friday 29 August 2025 17:39:41 +0000 (0:00:02.493) 0:03:00.580 ********* 2025-08-29 17:39:56.729652 | orchestrator | included: mariadb for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:39:56.729663 | orchestrator | 2025-08-29 17:39:56.729674 | orchestrator | TASK [mariadb : Creating shard root mysql user] ******************************** 2025-08-29 17:39:56.729684 | orchestrator | Friday 29 August 2025 17:39:41 +0000 (0:00:00.494) 0:03:01.075 ********* 2025-08-29 17:39:56.729694 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.729703 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.729713 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.729722 | orchestrator | 2025-08-29 17:39:56.729732 | orchestrator | TASK [mariadb : Creating mysql monitor user] *********************************** 2025-08-29 17:39:56.729742 | orchestrator | Friday 29 August 2025 17:39:44 +0000 (0:00:02.597) 0:03:03.673 ********* 2025-08-29 17:39:56.729752 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.729762 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.729771 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.729781 | orchestrator | 2025-08-29 17:39:56.729791 | orchestrator | TASK [mariadb : Creating database backup user and setting permissions] ********* 2025-08-29 17:39:56.729801 | orchestrator | Friday 29 August 2025 17:39:46 +0000 (0:00:02.091) 0:03:05.765 ********* 2025-08-29 17:39:56.729810 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.729820 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.729830 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.729839 | orchestrator | 2025-08-29 17:39:56.729849 | orchestrator | TASK [mariadb : Granting permissions on Mariabackup database to backup user] *** 2025-08-29 17:39:56.729859 | orchestrator | Friday 29 August 2025 17:39:48 +0000 (0:00:02.183) 0:03:07.948 ********* 2025-08-29 17:39:56.729869 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.729878 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.729888 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:39:56.729904 | orchestrator | 2025-08-29 17:39:56.729913 | orchestrator | TASK [mariadb : Wait for MariaDB service to be ready through VIP] ************** 2025-08-29 17:39:56.729923 | orchestrator | Friday 29 August 2025 17:39:51 +0000 (0:00:02.264) 0:03:10.213 ********* 2025-08-29 17:39:56.729933 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:39:56.729943 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:39:56.729952 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:39:56.729962 | orchestrator | 2025-08-29 17:39:56.729973 | orchestrator | TASK [Include mariadb post-upgrade.yml] **************************************** 2025-08-29 17:39:56.729983 | orchestrator | Friday 29 August 2025 17:39:54 +0000 (0:00:03.051) 0:03:13.265 ********* 2025-08-29 17:39:56.729993 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:39:56.730003 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:39:56.730012 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:39:56.730083 | orchestrator | 2025-08-29 17:39:56.730094 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:39:56.730104 | orchestrator | localhost : ok=3  changed=0 unreachable=0 failed=0 skipped=1  rescued=0 ignored=1  2025-08-29 17:39:56.730114 | orchestrator | testbed-node-0 : ok=34  changed=16  unreachable=0 failed=0 skipped=11  rescued=0 ignored=1  2025-08-29 17:39:56.730125 | orchestrator | testbed-node-1 : ok=20  changed=7  unreachable=0 failed=0 skipped=18  rescued=0 ignored=1  2025-08-29 17:39:56.730135 | orchestrator | testbed-node-2 : ok=20  changed=7  unreachable=0 failed=0 skipped=18  rescued=0 ignored=1  2025-08-29 17:39:56.730145 | orchestrator | 2025-08-29 17:39:56.730154 | orchestrator | 2025-08-29 17:39:56.730164 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:39:56.730173 | orchestrator | Friday 29 August 2025 17:39:54 +0000 (0:00:00.225) 0:03:13.490 ********* 2025-08-29 17:39:56.730183 | orchestrator | =============================================================================== 2025-08-29 17:39:56.730192 | orchestrator | mariadb : Restart MariaDB container ------------------------------------ 45.18s 2025-08-29 17:39:56.730202 | orchestrator | mariadb : Wait for MariaDB service port liveness ----------------------- 36.18s 2025-08-29 17:39:56.730219 | orchestrator | mariadb : Restart MariaDB container ------------------------------------ 11.86s 2025-08-29 17:39:56.730229 | orchestrator | mariadb : Check MariaDB service port liveness -------------------------- 10.84s 2025-08-29 17:39:56.730239 | orchestrator | mariadb : Running MariaDB bootstrap container -------------------------- 10.52s 2025-08-29 17:39:56.730249 | orchestrator | mariadb : Starting first MariaDB container ------------------------------ 7.86s 2025-08-29 17:39:56.730258 | orchestrator | mariadb : Wait for MariaDB service to sync WSREP ------------------------ 5.23s 2025-08-29 17:39:56.730268 | orchestrator | mariadb : Wait for MariaDB service port liveness ------------------------ 4.62s 2025-08-29 17:39:56.730277 | orchestrator | mariadb : Copying over galera.cnf --------------------------------------- 4.48s 2025-08-29 17:39:56.730287 | orchestrator | service-cert-copy : mariadb | Copying over extra CA certificates -------- 3.75s 2025-08-29 17:39:56.730297 | orchestrator | mariadb : Copying over config.json files for services ------------------- 3.75s 2025-08-29 17:39:56.730313 | orchestrator | service-cert-copy : mariadb | Copying over backend internal TLS certificate --- 3.53s 2025-08-29 17:39:56.730323 | orchestrator | mariadb : Ensuring config directories exist ----------------------------- 3.52s 2025-08-29 17:39:56.730332 | orchestrator | service-cert-copy : mariadb | Copying over backend internal TLS key ----- 3.16s 2025-08-29 17:39:56.730342 | orchestrator | mariadb : Wait for MariaDB service to be ready through VIP -------------- 3.05s 2025-08-29 17:39:56.730351 | orchestrator | Check MariaDB service --------------------------------------------------- 2.80s 2025-08-29 17:39:56.730361 | orchestrator | mariadb : Creating shard root mysql user -------------------------------- 2.60s 2025-08-29 17:39:56.730377 | orchestrator | mariadb : Check mariadb containers -------------------------------------- 2.59s 2025-08-29 17:39:56.730386 | orchestrator | mariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.49s 2025-08-29 17:39:56.730418 | orchestrator | mariadb : Wait for first MariaDB service to sync WSREP ------------------ 2.47s 2025-08-29 17:39:56.730429 | orchestrator | 2025-08-29 17:39:56 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:39:56.730439 | orchestrator | 2025-08-29 17:39:56 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:56.730449 | orchestrator | 2025-08-29 17:39:56 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:39:56.730459 | orchestrator | 2025-08-29 17:39:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:39:59.780896 | orchestrator | 2025-08-29 17:39:59 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:39:59.780976 | orchestrator | 2025-08-29 17:39:59 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:39:59.781675 | orchestrator | 2025-08-29 17:39:59 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:39:59.781699 | orchestrator | 2025-08-29 17:39:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:02.828817 | orchestrator | 2025-08-29 17:40:02 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:02.828880 | orchestrator | 2025-08-29 17:40:02 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:02.830134 | orchestrator | 2025-08-29 17:40:02 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:02.830166 | orchestrator | 2025-08-29 17:40:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:05.873270 | orchestrator | 2025-08-29 17:40:05 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:05.873943 | orchestrator | 2025-08-29 17:40:05 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:05.875719 | orchestrator | 2025-08-29 17:40:05 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:05.875758 | orchestrator | 2025-08-29 17:40:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:08.912657 | orchestrator | 2025-08-29 17:40:08 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:08.913438 | orchestrator | 2025-08-29 17:40:08 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:08.915534 | orchestrator | 2025-08-29 17:40:08 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:08.915872 | orchestrator | 2025-08-29 17:40:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:11.947902 | orchestrator | 2025-08-29 17:40:11 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:11.947996 | orchestrator | 2025-08-29 17:40:11 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:11.949218 | orchestrator | 2025-08-29 17:40:11 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:11.949250 | orchestrator | 2025-08-29 17:40:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:14.980640 | orchestrator | 2025-08-29 17:40:14 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:14.980828 | orchestrator | 2025-08-29 17:40:14 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:14.981526 | orchestrator | 2025-08-29 17:40:14 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:14.981585 | orchestrator | 2025-08-29 17:40:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:18.019304 | orchestrator | 2025-08-29 17:40:18 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:18.019434 | orchestrator | 2025-08-29 17:40:18 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:18.021160 | orchestrator | 2025-08-29 17:40:18 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:18.021197 | orchestrator | 2025-08-29 17:40:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:21.055741 | orchestrator | 2025-08-29 17:40:21 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:21.057497 | orchestrator | 2025-08-29 17:40:21 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:21.057526 | orchestrator | 2025-08-29 17:40:21 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:21.057537 | orchestrator | 2025-08-29 17:40:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:24.094779 | orchestrator | 2025-08-29 17:40:24 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:24.097259 | orchestrator | 2025-08-29 17:40:24 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:24.099147 | orchestrator | 2025-08-29 17:40:24 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:24.099192 | orchestrator | 2025-08-29 17:40:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:27.130932 | orchestrator | 2025-08-29 17:40:27 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:27.132641 | orchestrator | 2025-08-29 17:40:27 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:27.133855 | orchestrator | 2025-08-29 17:40:27 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:27.134090 | orchestrator | 2025-08-29 17:40:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:30.168778 | orchestrator | 2025-08-29 17:40:30 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:30.169521 | orchestrator | 2025-08-29 17:40:30 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:30.170823 | orchestrator | 2025-08-29 17:40:30 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:30.170852 | orchestrator | 2025-08-29 17:40:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:33.216529 | orchestrator | 2025-08-29 17:40:33 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:33.218704 | orchestrator | 2025-08-29 17:40:33 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:33.221037 | orchestrator | 2025-08-29 17:40:33 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:33.221064 | orchestrator | 2025-08-29 17:40:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:36.272520 | orchestrator | 2025-08-29 17:40:36 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:36.275208 | orchestrator | 2025-08-29 17:40:36 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:36.276957 | orchestrator | 2025-08-29 17:40:36 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:36.277020 | orchestrator | 2025-08-29 17:40:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:39.320608 | orchestrator | 2025-08-29 17:40:39 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:39.321954 | orchestrator | 2025-08-29 17:40:39 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:39.323169 | orchestrator | 2025-08-29 17:40:39 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:39.323477 | orchestrator | 2025-08-29 17:40:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:42.361257 | orchestrator | 2025-08-29 17:40:42 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:42.363177 | orchestrator | 2025-08-29 17:40:42 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:42.365793 | orchestrator | 2025-08-29 17:40:42 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:42.366053 | orchestrator | 2025-08-29 17:40:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:45.405688 | orchestrator | 2025-08-29 17:40:45 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:45.407235 | orchestrator | 2025-08-29 17:40:45 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:45.409604 | orchestrator | 2025-08-29 17:40:45 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:45.409822 | orchestrator | 2025-08-29 17:40:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:48.446231 | orchestrator | 2025-08-29 17:40:48 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:48.447124 | orchestrator | 2025-08-29 17:40:48 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:48.448813 | orchestrator | 2025-08-29 17:40:48 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:48.449078 | orchestrator | 2025-08-29 17:40:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:51.493543 | orchestrator | 2025-08-29 17:40:51 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:51.496383 | orchestrator | 2025-08-29 17:40:51 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:51.498173 | orchestrator | 2025-08-29 17:40:51 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:51.498201 | orchestrator | 2025-08-29 17:40:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:54.539861 | orchestrator | 2025-08-29 17:40:54 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:54.541870 | orchestrator | 2025-08-29 17:40:54 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:54.543696 | orchestrator | 2025-08-29 17:40:54 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:54.543719 | orchestrator | 2025-08-29 17:40:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:40:57.585146 | orchestrator | 2025-08-29 17:40:57 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:40:57.588390 | orchestrator | 2025-08-29 17:40:57 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:40:57.589804 | orchestrator | 2025-08-29 17:40:57 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:40:57.589829 | orchestrator | 2025-08-29 17:40:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:00.645974 | orchestrator | 2025-08-29 17:41:00 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:00.647154 | orchestrator | 2025-08-29 17:41:00 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:41:00.649760 | orchestrator | 2025-08-29 17:41:00 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:00.649839 | orchestrator | 2025-08-29 17:41:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:03.689907 | orchestrator | 2025-08-29 17:41:03 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:03.691957 | orchestrator | 2025-08-29 17:41:03 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:41:03.694240 | orchestrator | 2025-08-29 17:41:03 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:03.694269 | orchestrator | 2025-08-29 17:41:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:06.735994 | orchestrator | 2025-08-29 17:41:06 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:06.737957 | orchestrator | 2025-08-29 17:41:06 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:41:06.739943 | orchestrator | 2025-08-29 17:41:06 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:06.739979 | orchestrator | 2025-08-29 17:41:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:09.784637 | orchestrator | 2025-08-29 17:41:09 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:09.785125 | orchestrator | 2025-08-29 17:41:09 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:41:09.785690 | orchestrator | 2025-08-29 17:41:09 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:09.785767 | orchestrator | 2025-08-29 17:41:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:12.840856 | orchestrator | 2025-08-29 17:41:12 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:12.842195 | orchestrator | 2025-08-29 17:41:12 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state STARTED 2025-08-29 17:41:12.844247 | orchestrator | 2025-08-29 17:41:12 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:12.844544 | orchestrator | 2025-08-29 17:41:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:15.887862 | orchestrator | 2025-08-29 17:41:15 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:15.890843 | orchestrator | 2025-08-29 17:41:15 | INFO  | Task d12f167b-302d-458d-bade-26edcabbb36c is in state SUCCESS 2025-08-29 17:41:15.893087 | orchestrator | 2025-08-29 17:41:15.893151 | orchestrator | 2025-08-29 17:41:15.893869 | orchestrator | PLAY [Create ceph pools] ******************************************************* 2025-08-29 17:41:15.893888 | orchestrator | 2025-08-29 17:41:15.893900 | orchestrator | TASK [ceph-facts : Include facts.yml] ****************************************** 2025-08-29 17:41:15.893911 | orchestrator | Friday 29 August 2025 17:39:04 +0000 (0:00:00.600) 0:00:00.600 ********* 2025-08-29 17:41:15.893923 | orchestrator | included: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:41:15.893934 | orchestrator | 2025-08-29 17:41:15.893945 | orchestrator | TASK [ceph-facts : Check if it is atomic host] ********************************* 2025-08-29 17:41:15.893957 | orchestrator | Friday 29 August 2025 17:39:05 +0000 (0:00:00.616) 0:00:01.217 ********* 2025-08-29 17:41:15.893968 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894003 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894133 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894153 | orchestrator | 2025-08-29 17:41:15.894165 | orchestrator | TASK [ceph-facts : Set_fact is_atomic] ***************************************** 2025-08-29 17:41:15.894176 | orchestrator | Friday 29 August 2025 17:39:05 +0000 (0:00:00.631) 0:00:01.849 ********* 2025-08-29 17:41:15.894186 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894197 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894208 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894218 | orchestrator | 2025-08-29 17:41:15.894229 | orchestrator | TASK [ceph-facts : Check if podman binary is present] ************************** 2025-08-29 17:41:15.894240 | orchestrator | Friday 29 August 2025 17:39:06 +0000 (0:00:00.278) 0:00:02.128 ********* 2025-08-29 17:41:15.894250 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894261 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894272 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894282 | orchestrator | 2025-08-29 17:41:15.894293 | orchestrator | TASK [ceph-facts : Set_fact container_binary] ********************************** 2025-08-29 17:41:15.894304 | orchestrator | Friday 29 August 2025 17:39:06 +0000 (0:00:00.771) 0:00:02.899 ********* 2025-08-29 17:41:15.894314 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894325 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894336 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894346 | orchestrator | 2025-08-29 17:41:15.894357 | orchestrator | TASK [ceph-facts : Set_fact ceph_cmd] ****************************************** 2025-08-29 17:41:15.894368 | orchestrator | Friday 29 August 2025 17:39:07 +0000 (0:00:00.328) 0:00:03.227 ********* 2025-08-29 17:41:15.894379 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894389 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894400 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894430 | orchestrator | 2025-08-29 17:41:15.894441 | orchestrator | TASK [ceph-facts : Set_fact discovered_interpreter_python] ********************* 2025-08-29 17:41:15.894454 | orchestrator | Friday 29 August 2025 17:39:07 +0000 (0:00:00.297) 0:00:03.525 ********* 2025-08-29 17:41:15.894465 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894478 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894490 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894502 | orchestrator | 2025-08-29 17:41:15.894514 | orchestrator | TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] *** 2025-08-29 17:41:15.894526 | orchestrator | Friday 29 August 2025 17:39:07 +0000 (0:00:00.301) 0:00:03.826 ********* 2025-08-29 17:41:15.894539 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.894551 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.894563 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.894575 | orchestrator | 2025-08-29 17:41:15.894587 | orchestrator | TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ****************** 2025-08-29 17:41:15.894600 | orchestrator | Friday 29 August 2025 17:39:08 +0000 (0:00:00.492) 0:00:04.319 ********* 2025-08-29 17:41:15.894612 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894624 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894636 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894648 | orchestrator | 2025-08-29 17:41:15.894660 | orchestrator | TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************ 2025-08-29 17:41:15.894672 | orchestrator | Friday 29 August 2025 17:39:08 +0000 (0:00:00.289) 0:00:04.609 ********* 2025-08-29 17:41:15.894683 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0) 2025-08-29 17:41:15.894695 | orchestrator | ok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:41:15.894708 | orchestrator | ok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:41:15.894720 | orchestrator | 2025-08-29 17:41:15.894732 | orchestrator | TASK [ceph-facts : Set_fact container_exec_cmd] ******************************** 2025-08-29 17:41:15.894744 | orchestrator | Friday 29 August 2025 17:39:09 +0000 (0:00:00.614) 0:00:05.224 ********* 2025-08-29 17:41:15.894756 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.894780 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.894792 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.894805 | orchestrator | 2025-08-29 17:41:15.894818 | orchestrator | TASK [ceph-facts : Find a running mon container] ******************************* 2025-08-29 17:41:15.894830 | orchestrator | Friday 29 August 2025 17:39:09 +0000 (0:00:00.456) 0:00:05.680 ********* 2025-08-29 17:41:15.894841 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0) 2025-08-29 17:41:15.894851 | orchestrator | ok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:41:15.894862 | orchestrator | ok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:41:15.894873 | orchestrator | 2025-08-29 17:41:15.894894 | orchestrator | TASK [ceph-facts : Check for a ceph mon socket] ******************************** 2025-08-29 17:41:15.894906 | orchestrator | Friday 29 August 2025 17:39:11 +0000 (0:00:02.118) 0:00:07.799 ********* 2025-08-29 17:41:15.894917 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-0)  2025-08-29 17:41:15.894927 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-1)  2025-08-29 17:41:15.894939 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-2)  2025-08-29 17:41:15.894950 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.894961 | orchestrator | 2025-08-29 17:41:15.894972 | orchestrator | TASK [ceph-facts : Check if the ceph mon socket is in-use] ********************* 2025-08-29 17:41:15.895037 | orchestrator | Friday 29 August 2025 17:39:12 +0000 (0:00:00.411) 0:00:08.211 ********* 2025-08-29 17:41:15.895053 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.895066 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.895077 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.895088 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.895099 | orchestrator | 2025-08-29 17:41:15.895111 | orchestrator | TASK [ceph-facts : Set_fact running_mon - non_container] *********************** 2025-08-29 17:41:15.895122 | orchestrator | Friday 29 August 2025 17:39:12 +0000 (0:00:00.755) 0:00:08.966 ********* 2025-08-29 17:41:15.895160 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.895174 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.895186 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.895197 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.895216 | orchestrator | 2025-08-29 17:41:15.895227 | orchestrator | TASK [ceph-facts : Set_fact running_mon - container] *************************** 2025-08-29 17:41:15.895238 | orchestrator | Friday 29 August 2025 17:39:13 +0000 (0:00:00.163) 0:00:09.130 ********* 2025-08-29 17:41:15.895251 | orchestrator | ok: [testbed-node-3] => (item={'changed': False, 'stdout': '0e5cc467e26a', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2025-08-29 17:39:10.292873', 'end': '2025-08-29 17:39:10.326492', 'delta': '0:00:00.033619', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['0e5cc467e26a'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) 2025-08-29 17:41:15.895271 | orchestrator | ok: [testbed-node-3] => (item={'changed': False, 'stdout': '9d107394c3ff', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2025-08-29 17:39:11.057245', 'end': '2025-08-29 17:39:11.107181', 'delta': '0:00:00.049936', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['9d107394c3ff'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) 2025-08-29 17:41:15.895317 | orchestrator | ok: [testbed-node-3] => (item={'changed': False, 'stdout': 'bc3d2c50e18a', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2025-08-29 17:39:11.613123', 'end': '2025-08-29 17:39:11.653720', 'delta': '0:00:00.040597', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['bc3d2c50e18a'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) 2025-08-29 17:41:15.895331 | orchestrator | 2025-08-29 17:41:15.895342 | orchestrator | TASK [ceph-facts : Set_fact _container_exec_cmd] ******************************* 2025-08-29 17:41:15.895353 | orchestrator | Friday 29 August 2025 17:39:13 +0000 (0:00:00.364) 0:00:09.495 ********* 2025-08-29 17:41:15.895364 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.895375 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.895386 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.895396 | orchestrator | 2025-08-29 17:41:15.895435 | orchestrator | TASK [ceph-facts : Get current fsid if cluster is already running] ************* 2025-08-29 17:41:15.895447 | orchestrator | Friday 29 August 2025 17:39:13 +0000 (0:00:00.409) 0:00:09.904 ********* 2025-08-29 17:41:15.895458 | orchestrator | ok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] 2025-08-29 17:41:15.895468 | orchestrator | 2025-08-29 17:41:15.895479 | orchestrator | TASK [ceph-facts : Set_fact current_fsid rc 1] ********************************* 2025-08-29 17:41:15.895490 | orchestrator | Friday 29 August 2025 17:39:15 +0000 (0:00:01.765) 0:00:11.670 ********* 2025-08-29 17:41:15.895501 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.895512 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.895523 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.895533 | orchestrator | 2025-08-29 17:41:15.895544 | orchestrator | TASK [ceph-facts : Get current fsid] ******************************************* 2025-08-29 17:41:15.895555 | orchestrator | Friday 29 August 2025 17:39:15 +0000 (0:00:00.279) 0:00:11.949 ********* 2025-08-29 17:41:15.895575 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.895586 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.895597 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.895608 | orchestrator | 2025-08-29 17:41:15.895622 | orchestrator | TASK [ceph-facts : Set_fact fsid] ********************************************** 2025-08-29 17:41:15.895642 | orchestrator | Friday 29 August 2025 17:39:16 +0000 (0:00:00.405) 0:00:12.355 ********* 2025-08-29 17:41:15.895661 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.895679 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.895697 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.895715 | orchestrator | 2025-08-29 17:41:15.895733 | orchestrator | TASK [ceph-facts : Set_fact fsid from current_fsid] **************************** 2025-08-29 17:41:15.895752 | orchestrator | Friday 29 August 2025 17:39:16 +0000 (0:00:00.452) 0:00:12.807 ********* 2025-08-29 17:41:15.895771 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.895789 | orchestrator | 2025-08-29 17:41:15.895810 | orchestrator | TASK [ceph-facts : Generate cluster fsid] ************************************** 2025-08-29 17:41:15.895830 | orchestrator | Friday 29 August 2025 17:39:16 +0000 (0:00:00.129) 0:00:12.937 ********* 2025-08-29 17:41:15.895944 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.895956 | orchestrator | 2025-08-29 17:41:15.895967 | orchestrator | TASK [ceph-facts : Set_fact fsid] ********************************************** 2025-08-29 17:41:15.895977 | orchestrator | Friday 29 August 2025 17:39:17 +0000 (0:00:00.224) 0:00:13.161 ********* 2025-08-29 17:41:15.895988 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.895999 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.896010 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.896022 | orchestrator | 2025-08-29 17:41:15.896042 | orchestrator | TASK [ceph-facts : Resolve device link(s)] ************************************* 2025-08-29 17:41:15.896060 | orchestrator | Friday 29 August 2025 17:39:17 +0000 (0:00:00.274) 0:00:13.436 ********* 2025-08-29 17:41:15.896080 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.896101 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.896121 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.896132 | orchestrator | 2025-08-29 17:41:15.896143 | orchestrator | TASK [ceph-facts : Set_fact build devices from resolved symlinks] ************** 2025-08-29 17:41:15.896154 | orchestrator | Friday 29 August 2025 17:39:17 +0000 (0:00:00.311) 0:00:13.747 ********* 2025-08-29 17:41:15.896170 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.896188 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.896207 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.896225 | orchestrator | 2025-08-29 17:41:15.896243 | orchestrator | TASK [ceph-facts : Resolve dedicated_device link(s)] *************************** 2025-08-29 17:41:15.896357 | orchestrator | Friday 29 August 2025 17:39:18 +0000 (0:00:00.462) 0:00:14.209 ********* 2025-08-29 17:41:15.896377 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.896388 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.896399 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.896457 | orchestrator | 2025-08-29 17:41:15.896470 | orchestrator | TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] **** 2025-08-29 17:41:15.896490 | orchestrator | Friday 29 August 2025 17:39:18 +0000 (0:00:00.375) 0:00:14.584 ********* 2025-08-29 17:41:15.896501 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.896512 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.896523 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.896534 | orchestrator | 2025-08-29 17:41:15.896545 | orchestrator | TASK [ceph-facts : Resolve bluestore_wal_device link(s)] *********************** 2025-08-29 17:41:15.896556 | orchestrator | Friday 29 August 2025 17:39:18 +0000 (0:00:00.307) 0:00:14.891 ********* 2025-08-29 17:41:15.896566 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.896577 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.896588 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.896599 | orchestrator | 2025-08-29 17:41:15.896610 | orchestrator | TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] *** 2025-08-29 17:41:15.896669 | orchestrator | Friday 29 August 2025 17:39:19 +0000 (0:00:00.306) 0:00:15.198 ********* 2025-08-29 17:41:15.896683 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.896694 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.896704 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.896715 | orchestrator | 2025-08-29 17:41:15.896726 | orchestrator | TASK [ceph-facts : Collect existed devices] ************************************ 2025-08-29 17:41:15.896737 | orchestrator | Friday 29 August 2025 17:39:19 +0000 (0:00:00.463) 0:00:15.661 ********* 2025-08-29 17:41:15.896750 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b', 'dm-uuid-LVM-8SW2ubuTvGOnhFOTau03JFgatBW9cwlrT4nSVIJUb9pyMNhhp0p0QmNBRj3JLwK3'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896763 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a', 'dm-uuid-LVM-5q8dGM2db5NrMc0chc8fHWOTnADjoH0OBy9rcTAnAINXxTj11VL1o8QghnCwxTrh'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896775 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896787 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896799 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896810 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896826 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896872 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896886 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896897 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2', 'dm-uuid-LVM-GUlkxvsBUEoBNxRHTurm2I6vdb71iOQT09A53JwNvatREyMn391iO8eZzncdLhOV'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896909 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896920 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2', 'dm-uuid-LVM-gVhXWy9gnL13qoE5gEc3ufITkM48FBGMjmEmplIXepwpRgq3AOhOj1huYwO1lOSE'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896966 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.896988 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.896999 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Ln6jHt-qsqU-X1le-h5ao-5dTL-pKYQ-cOGmh1', 'scsi-0QEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322', 'scsi-SQEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897012 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897023 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uPkwe4-XhPc-lzhV-DN2T-ebZn-SA0c-JkEUrj', 'scsi-0QEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f', 'scsi-SQEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897035 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5', 'scsi-SQEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897047 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897093 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-17-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897107 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897119 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897131 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.897143 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897154 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897166 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897202 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897221 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe', 'dm-uuid-LVM-tOwLLnFJLaczcEpQ6NBpurAfznGF1LI9EZRJAMwx4l3TllcGTbf8jWoatcKCkzDd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897233 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-0A5zqG-61nB-VreW-lmbz-tCSc-hwpf-PGsoeL', 'scsi-0QEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3', 'scsi-SQEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897245 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288', 'dm-uuid-LVM-y0l4STA9mSjwec3ylN7nLUCBF2I0AWgcWG3yGA1QpITOXDaU9fQLHWqzTqgQaUq4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897257 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-IdAuJW-KdwX-YTF8-XPAR-1cwa-KggL-hFVogq', 'scsi-0QEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6', 'scsi-SQEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897272 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897288 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474', 'scsi-SQEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897308 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897321 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897332 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897344 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.897355 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897366 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897376 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897392 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897402 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}})  2025-08-29 17:41:15.897443 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897456 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-CYMrCT-OsQ3-cdVr-EcBN-TKOH-xE0q-pkdaCm', 'scsi-0QEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6', 'scsi-SQEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897467 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XNIeJ-OvKH-SWIj-T9QN-B9GX-Jw9F-bWcj1x', 'scsi-0QEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6', 'scsi-SQEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897486 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290', 'scsi-SQEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897501 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-18-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}})  2025-08-29 17:41:15.897512 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.897522 | orchestrator | 2025-08-29 17:41:15.897532 | orchestrator | TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] *** 2025-08-29 17:41:15.897542 | orchestrator | Friday 29 August 2025 17:39:20 +0000 (0:00:00.514) 0:00:16.176 ********* 2025-08-29 17:41:15.897552 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b', 'dm-uuid-LVM-8SW2ubuTvGOnhFOTau03JFgatBW9cwlrT4nSVIJUb9pyMNhhp0p0QmNBRj3JLwK3'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897563 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a', 'dm-uuid-LVM-5q8dGM2db5NrMc0chc8fHWOTnADjoH0OBy9rcTAnAINXxTj11VL1o8QghnCwxTrh'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897573 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897588 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897622 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897640 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897651 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897661 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897671 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897686 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897708 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16', 'scsi-SQEMU_QEMU_HARDDISK_15412d66-e28a-4d80-903c-2982b517c5a1-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897721 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b-osd--block--25b2a0ab--1f04--5c39--a4c5--8c8c13772d9b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-Ln6jHt-qsqU-X1le-h5ao-5dTL-pKYQ-cOGmh1', 'scsi-0QEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322', 'scsi-SQEMU_QEMU_HARDDISK_90c73877-8ee8-469e-bb0b-db8fbc3e9322'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897732 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2', 'dm-uuid-LVM-GUlkxvsBUEoBNxRHTurm2I6vdb71iOQT09A53JwNvatREyMn391iO8eZzncdLhOV'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897751 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--67ff47d9--d75a--55af--b095--c4dbbf8f796a-osd--block--67ff47d9--d75a--55af--b095--c4dbbf8f796a'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-uPkwe4-XhPc-lzhV-DN2T-ebZn-SA0c-JkEUrj', 'scsi-0QEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f', 'scsi-SQEMU_QEMU_HARDDISK_8ff97e3e-a6a2-4dcf-a003-6cc5f2bd734f'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897768 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2', 'dm-uuid-LVM-gVhXWy9gnL13qoE5gEc3ufITkM48FBGMjmEmplIXepwpRgq3AOhOj1huYwO1lOSE'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897779 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5', 'scsi-SQEMU_QEMU_HARDDISK_8cbeb260-c625-4217-b6ca-cc546e4ae7e5'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897789 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897799 | orchestrator | skipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-17-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897815 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.897825 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897842 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897858 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897868 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897879 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897889 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe', 'dm-uuid-LVM-tOwLLnFJLaczcEpQ6NBpurAfznGF1LI9EZRJAMwx4l3TllcGTbf8jWoatcKCkzDd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897904 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897918 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288', 'dm-uuid-LVM-y0l4STA9mSjwec3ylN7nLUCBF2I0AWgcWG3yGA1QpITOXDaU9fQLHWqzTqgQaUq4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897935 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897945 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897957 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16', 'scsi-SQEMU_QEMU_HARDDISK_8099cd87-683e-4d8a-b9af-8a40d3773320-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897978 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.897995 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--b89384ec--5219--5f2a--8735--84f78c8179d2-osd--block--b89384ec--5219--5f2a--8735--84f78c8179d2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-0A5zqG-61nB-VreW-lmbz-tCSc-hwpf-PGsoeL', 'scsi-0QEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3', 'scsi-SQEMU_QEMU_HARDDISK_8606d01f-f80c-4f5c-bc07-f0f70fbd69b3'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898006 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898042 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--860f9296--4089--56f8--8238--0f24b03dbca2-osd--block--860f9296--4089--56f8--8238--0f24b03dbca2'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-IdAuJW-KdwX-YTF8-XPAR-1cwa-KggL-hFVogq', 'scsi-0QEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6', 'scsi-SQEMU_QEMU_HARDDISK_8e58fc08-3f1a-44a4-9385-6e01a94c76b6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898060 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898071 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474', 'scsi-SQEMU_QEMU_HARDDISK_8a1a83e3-79d3-46d9-abf0-648ca1b5a474'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898091 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898102 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898112 | orchestrator | skipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-14-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898127 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.898138 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898148 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898170 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16', 'scsi-SQEMU_QEMU_HARDDISK_4f525629-0253-4144-8364-3c913388a34f-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898182 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe-osd--block--e5f8acb9--1955--50e0--bba9--ecbc3a5da5fe'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-CYMrCT-OsQ3-cdVr-EcBN-TKOH-xE0q-pkdaCm', 'scsi-0QEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6', 'scsi-SQEMU_QEMU_HARDDISK_588899d3-5e4c-43b2-b3cb-36cecbb176f6'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898197 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--21b183ef--757d--561c--bead--7bb3aee28288-osd--block--21b183ef--757d--561c--bead--7bb3aee28288'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-2XNIeJ-OvKH-SWIj-T9QN-B9GX-Jw9F-bWcj1x', 'scsi-0QEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6', 'scsi-SQEMU_QEMU_HARDDISK_5c881ad7-dcaa-4170-926f-a62d483103a6'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898211 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290', 'scsi-SQEMU_QEMU_HARDDISK_c3b7c3a0-9d38-47db-a77f-489f49c79290'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898228 | orchestrator | skipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2025-08-29-16-48-18-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'})  2025-08-29 17:41:15.898238 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.898248 | orchestrator | 2025-08-29 17:41:15.898258 | orchestrator | TASK [ceph-facts : Check if the ceph conf exists] ****************************** 2025-08-29 17:41:15.898268 | orchestrator | Friday 29 August 2025 17:39:20 +0000 (0:00:00.563) 0:00:16.740 ********* 2025-08-29 17:41:15.898278 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.898288 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.898298 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.898307 | orchestrator | 2025-08-29 17:41:15.898317 | orchestrator | TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] *************** 2025-08-29 17:41:15.898326 | orchestrator | Friday 29 August 2025 17:39:21 +0000 (0:00:00.732) 0:00:17.473 ********* 2025-08-29 17:41:15.898341 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.898351 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.898361 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.898370 | orchestrator | 2025-08-29 17:41:15.898380 | orchestrator | TASK [ceph-facts : Read osd pool default crush rule] *************************** 2025-08-29 17:41:15.898390 | orchestrator | Friday 29 August 2025 17:39:21 +0000 (0:00:00.455) 0:00:17.928 ********* 2025-08-29 17:41:15.898399 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.898425 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.898435 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.898445 | orchestrator | 2025-08-29 17:41:15.898455 | orchestrator | TASK [ceph-facts : Set osd_pool_default_crush_rule fact] *********************** 2025-08-29 17:41:15.898464 | orchestrator | Friday 29 August 2025 17:39:22 +0000 (0:00:00.684) 0:00:18.613 ********* 2025-08-29 17:41:15.898474 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.898484 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.898493 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.898503 | orchestrator | 2025-08-29 17:41:15.898512 | orchestrator | TASK [ceph-facts : Read osd pool default crush rule] *************************** 2025-08-29 17:41:15.898522 | orchestrator | Friday 29 August 2025 17:39:22 +0000 (0:00:00.285) 0:00:18.898 ********* 2025-08-29 17:41:15.898531 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.898541 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.898550 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.898560 | orchestrator | 2025-08-29 17:41:15.898570 | orchestrator | TASK [ceph-facts : Set osd_pool_default_crush_rule fact] *********************** 2025-08-29 17:41:15.898579 | orchestrator | Friday 29 August 2025 17:39:23 +0000 (0:00:00.422) 0:00:19.320 ********* 2025-08-29 17:41:15.898589 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.898598 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.898608 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.898617 | orchestrator | 2025-08-29 17:41:15.898627 | orchestrator | TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] ************************* 2025-08-29 17:41:15.898636 | orchestrator | Friday 29 August 2025 17:39:23 +0000 (0:00:00.487) 0:00:19.808 ********* 2025-08-29 17:41:15.898646 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-0) 2025-08-29 17:41:15.898656 | orchestrator | ok: [testbed-node-4] => (item=testbed-node-0) 2025-08-29 17:41:15.898665 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-1) 2025-08-29 17:41:15.898674 | orchestrator | ok: [testbed-node-5] => (item=testbed-node-0) 2025-08-29 17:41:15.898684 | orchestrator | ok: [testbed-node-4] => (item=testbed-node-1) 2025-08-29 17:41:15.898693 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-2) 2025-08-29 17:41:15.898703 | orchestrator | ok: [testbed-node-5] => (item=testbed-node-1) 2025-08-29 17:41:15.898712 | orchestrator | ok: [testbed-node-4] => (item=testbed-node-2) 2025-08-29 17:41:15.898721 | orchestrator | ok: [testbed-node-5] => (item=testbed-node-2) 2025-08-29 17:41:15.898731 | orchestrator | 2025-08-29 17:41:15.898741 | orchestrator | TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] ************************* 2025-08-29 17:41:15.898750 | orchestrator | Friday 29 August 2025 17:39:24 +0000 (0:00:00.936) 0:00:20.745 ********* 2025-08-29 17:41:15.898760 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-0)  2025-08-29 17:41:15.898769 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-1)  2025-08-29 17:41:15.898779 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-2)  2025-08-29 17:41:15.898788 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.898798 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-0)  2025-08-29 17:41:15.898807 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-1)  2025-08-29 17:41:15.898817 | orchestrator | skipping: [testbed-node-4] => (item=testbed-node-2)  2025-08-29 17:41:15.898826 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.898835 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-0)  2025-08-29 17:41:15.898850 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-1)  2025-08-29 17:41:15.898860 | orchestrator | skipping: [testbed-node-5] => (item=testbed-node-2)  2025-08-29 17:41:15.898876 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.898892 | orchestrator | 2025-08-29 17:41:15.898908 | orchestrator | TASK [ceph-facts : Import_tasks set_radosgw_address.yml] *********************** 2025-08-29 17:41:15.898931 | orchestrator | Friday 29 August 2025 17:39:25 +0000 (0:00:00.339) 0:00:21.084 ********* 2025-08-29 17:41:15.898950 | orchestrator | included: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:41:15.898966 | orchestrator | 2025-08-29 17:41:15.898983 | orchestrator | TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface from node "{{ ceph_dashboard_call_item }}"] *** 2025-08-29 17:41:15.898999 | orchestrator | Friday 29 August 2025 17:39:25 +0000 (0:00:00.784) 0:00:21.868 ********* 2025-08-29 17:41:15.899016 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.899032 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.899049 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.899067 | orchestrator | 2025-08-29 17:41:15.899093 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] **** 2025-08-29 17:41:15.899104 | orchestrator | Friday 29 August 2025 17:39:26 +0000 (0:00:00.337) 0:00:22.206 ********* 2025-08-29 17:41:15.899114 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.899123 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.899133 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.899142 | orchestrator | 2025-08-29 17:41:15.899152 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] **** 2025-08-29 17:41:15.899162 | orchestrator | Friday 29 August 2025 17:39:26 +0000 (0:00:00.295) 0:00:22.502 ********* 2025-08-29 17:41:15.899172 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.899181 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.899191 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:41:15.899200 | orchestrator | 2025-08-29 17:41:15.899210 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] *************** 2025-08-29 17:41:15.899219 | orchestrator | Friday 29 August 2025 17:39:26 +0000 (0:00:00.309) 0:00:22.811 ********* 2025-08-29 17:41:15.899229 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.899238 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.899248 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.899257 | orchestrator | 2025-08-29 17:41:15.899267 | orchestrator | TASK [ceph-facts : Set_fact _interface] **************************************** 2025-08-29 17:41:15.899277 | orchestrator | Friday 29 August 2025 17:39:27 +0000 (0:00:00.608) 0:00:23.420 ********* 2025-08-29 17:41:15.899286 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:41:15.899296 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:41:15.899306 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:41:15.899315 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.899325 | orchestrator | 2025-08-29 17:41:15.899334 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ****** 2025-08-29 17:41:15.899344 | orchestrator | Friday 29 August 2025 17:39:27 +0000 (0:00:00.362) 0:00:23.783 ********* 2025-08-29 17:41:15.899353 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:41:15.899363 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:41:15.899373 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:41:15.899382 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.899392 | orchestrator | 2025-08-29 17:41:15.899401 | orchestrator | TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ****** 2025-08-29 17:41:15.899464 | orchestrator | Friday 29 August 2025 17:39:28 +0000 (0:00:00.359) 0:00:24.142 ********* 2025-08-29 17:41:15.899484 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-3)  2025-08-29 17:41:15.899501 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-4)  2025-08-29 17:41:15.899529 | orchestrator | skipping: [testbed-node-3] => (item=testbed-node-5)  2025-08-29 17:41:15.899546 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.899560 | orchestrator | 2025-08-29 17:41:15.899573 | orchestrator | TASK [ceph-facts : Reset rgw_instances (workaround)] *************************** 2025-08-29 17:41:15.899586 | orchestrator | Friday 29 August 2025 17:39:28 +0000 (0:00:00.356) 0:00:24.499 ********* 2025-08-29 17:41:15.899599 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:41:15.899611 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:41:15.899623 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:41:15.899635 | orchestrator | 2025-08-29 17:41:15.899648 | orchestrator | TASK [ceph-facts : Set_fact rgw_instances] ************************************* 2025-08-29 17:41:15.899662 | orchestrator | Friday 29 August 2025 17:39:28 +0000 (0:00:00.310) 0:00:24.810 ********* 2025-08-29 17:41:15.899674 | orchestrator | ok: [testbed-node-4] => (item=0) 2025-08-29 17:41:15.899687 | orchestrator | ok: [testbed-node-3] => (item=0) 2025-08-29 17:41:15.899701 | orchestrator | ok: [testbed-node-5] => (item=0) 2025-08-29 17:41:15.899715 | orchestrator | 2025-08-29 17:41:15.899729 | orchestrator | TASK [ceph-facts : Set_fact ceph_run_cmd] ************************************** 2025-08-29 17:41:15.899742 | orchestrator | Friday 29 August 2025 17:39:29 +0000 (0:00:00.618) 0:00:25.428 ********* 2025-08-29 17:41:15.899756 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0) 2025-08-29 17:41:15.899770 | orchestrator | ok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:41:15.899783 | orchestrator | ok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:41:15.899797 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-3) 2025-08-29 17:41:15.899809 | orchestrator | ok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4) 2025-08-29 17:41:15.899823 | orchestrator | ok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5) 2025-08-29 17:41:15.899835 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager) 2025-08-29 17:41:15.899850 | orchestrator | 2025-08-29 17:41:15.899864 | orchestrator | TASK [ceph-facts : Set_fact ceph_admin_command] ******************************** 2025-08-29 17:41:15.899877 | orchestrator | Friday 29 August 2025 17:39:30 +0000 (0:00:00.983) 0:00:26.412 ********* 2025-08-29 17:41:15.899891 | orchestrator | ok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0) 2025-08-29 17:41:15.899915 | orchestrator | ok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1) 2025-08-29 17:41:15.899930 | orchestrator | ok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2) 2025-08-29 17:41:15.899943 | orchestrator | ok: [testbed-node-3] => (item=testbed-node-3) 2025-08-29 17:41:15.899957 | orchestrator | ok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4) 2025-08-29 17:41:15.899971 | orchestrator | ok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5) 2025-08-29 17:41:15.899984 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager) 2025-08-29 17:41:15.899998 | orchestrator | 2025-08-29 17:41:15.900021 | orchestrator | TASK [Include tasks from the ceph-osd role] ************************************ 2025-08-29 17:41:15.900036 | orchestrator | Friday 29 August 2025 17:39:32 +0000 (0:00:01.909) 0:00:28.321 ********* 2025-08-29 17:41:15.900051 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:41:15.900064 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:41:15.900079 | orchestrator | included: /ansible/tasks/openstack_config.yml for testbed-node-5 2025-08-29 17:41:15.900093 | orchestrator | 2025-08-29 17:41:15.900106 | orchestrator | TASK [create openstack pool(s)] ************************************************ 2025-08-29 17:41:15.900117 | orchestrator | Friday 29 August 2025 17:39:32 +0000 (0:00:00.371) 0:00:28.693 ********* 2025-08-29 17:41:15.900126 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1}) 2025-08-29 17:41:15.900144 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1}) 2025-08-29 17:41:15.900152 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1}) 2025-08-29 17:41:15.900161 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1}) 2025-08-29 17:41:15.900169 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1}) 2025-08-29 17:41:15.900176 | orchestrator | 2025-08-29 17:41:15.900184 | orchestrator | TASK [generate keys] *********************************************************** 2025-08-29 17:41:15.900192 | orchestrator | Friday 29 August 2025 17:40:19 +0000 (0:00:46.820) 0:01:15.513 ********* 2025-08-29 17:41:15.900200 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900208 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900216 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900223 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900231 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900239 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900246 | orchestrator | changed: [testbed-node-5 -> {{ groups[mon_group_name][0] }}] 2025-08-29 17:41:15.900254 | orchestrator | 2025-08-29 17:41:15.900262 | orchestrator | TASK [get keys from monitors] ************************************************** 2025-08-29 17:41:15.900270 | orchestrator | Friday 29 August 2025 17:40:44 +0000 (0:00:24.512) 0:01:40.026 ********* 2025-08-29 17:41:15.900278 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900285 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900293 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900301 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900308 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900316 | orchestrator | ok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900324 | orchestrator | ok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}] 2025-08-29 17:41:15.900332 | orchestrator | 2025-08-29 17:41:15.900340 | orchestrator | TASK [copy ceph key(s) if needed] ********************************************** 2025-08-29 17:41:15.900347 | orchestrator | Friday 29 August 2025 17:40:56 +0000 (0:00:12.127) 0:01:52.153 ********* 2025-08-29 17:41:15.900360 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900368 | orchestrator | changed: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:41:15.900376 | orchestrator | changed: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:41:15.900387 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900395 | orchestrator | changed: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:41:15.900403 | orchestrator | changed: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:41:15.900444 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900459 | orchestrator | changed: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:41:15.900472 | orchestrator | changed: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:41:15.900485 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900496 | orchestrator | changed: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:41:15.900510 | orchestrator | changed: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:41:15.900524 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900537 | orchestrator | changed: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:41:15.900551 | orchestrator | changed: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:41:15.900562 | orchestrator | changed: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None) 2025-08-29 17:41:15.900570 | orchestrator | changed: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None) 2025-08-29 17:41:15.900578 | orchestrator | changed: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None) 2025-08-29 17:41:15.900585 | orchestrator | changed: [testbed-node-5 -> {{ item.1 }}] 2025-08-29 17:41:15.900593 | orchestrator | 2025-08-29 17:41:15.900601 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:41:15.900609 | orchestrator | testbed-node-3 : ok=25  changed=0 unreachable=0 failed=0 skipped=28  rescued=0 ignored=0 2025-08-29 17:41:15.900618 | orchestrator | testbed-node-4 : ok=18  changed=0 unreachable=0 failed=0 skipped=21  rescued=0 ignored=0 2025-08-29 17:41:15.900626 | orchestrator | testbed-node-5 : ok=23  changed=3  unreachable=0 failed=0 skipped=20  rescued=0 ignored=0 2025-08-29 17:41:15.900634 | orchestrator | 2025-08-29 17:41:15.900642 | orchestrator | 2025-08-29 17:41:15.900650 | orchestrator | 2025-08-29 17:41:15.900658 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:41:15.900665 | orchestrator | Friday 29 August 2025 17:41:13 +0000 (0:00:17.504) 0:02:09.658 ********* 2025-08-29 17:41:15.900673 | orchestrator | =============================================================================== 2025-08-29 17:41:15.900681 | orchestrator | create openstack pool(s) ----------------------------------------------- 46.82s 2025-08-29 17:41:15.900688 | orchestrator | generate keys ---------------------------------------------------------- 24.51s 2025-08-29 17:41:15.900696 | orchestrator | copy ceph key(s) if needed --------------------------------------------- 17.50s 2025-08-29 17:41:15.900704 | orchestrator | get keys from monitors ------------------------------------------------- 12.13s 2025-08-29 17:41:15.900712 | orchestrator | ceph-facts : Find a running mon container ------------------------------- 2.12s 2025-08-29 17:41:15.900720 | orchestrator | ceph-facts : Set_fact ceph_admin_command -------------------------------- 1.91s 2025-08-29 17:41:15.900728 | orchestrator | ceph-facts : Get current fsid if cluster is already running ------------- 1.77s 2025-08-29 17:41:15.900736 | orchestrator | ceph-facts : Set_fact ceph_run_cmd -------------------------------------- 0.98s 2025-08-29 17:41:15.900744 | orchestrator | ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 0.94s 2025-08-29 17:41:15.900752 | orchestrator | ceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.78s 2025-08-29 17:41:15.900760 | orchestrator | ceph-facts : Check if podman binary is present -------------------------- 0.77s 2025-08-29 17:41:15.900775 | orchestrator | ceph-facts : Check if the ceph mon socket is in-use --------------------- 0.76s 2025-08-29 17:41:15.900783 | orchestrator | ceph-facts : Check if the ceph conf exists ------------------------------ 0.73s 2025-08-29 17:41:15.900791 | orchestrator | ceph-facts : Read osd pool default crush rule --------------------------- 0.68s 2025-08-29 17:41:15.900798 | orchestrator | ceph-facts : Check if it is atomic host --------------------------------- 0.63s 2025-08-29 17:41:15.900806 | orchestrator | ceph-facts : Set_fact rgw_instances ------------------------------------- 0.62s 2025-08-29 17:41:15.900814 | orchestrator | ceph-facts : Include facts.yml ------------------------------------------ 0.62s 2025-08-29 17:41:15.900821 | orchestrator | ceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.61s 2025-08-29 17:41:15.900829 | orchestrator | ceph-facts : Set_fact _radosgw_address to radosgw_address --------------- 0.61s 2025-08-29 17:41:15.900837 | orchestrator | ceph-facts : Set_fact devices generate device list when osd_auto_discovery --- 0.56s 2025-08-29 17:41:15.900845 | orchestrator | 2025-08-29 17:41:15 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:15.900857 | orchestrator | 2025-08-29 17:41:15 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:15.900865 | orchestrator | 2025-08-29 17:41:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:18.934834 | orchestrator | 2025-08-29 17:41:18 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:18.935698 | orchestrator | 2025-08-29 17:41:18 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:18.937394 | orchestrator | 2025-08-29 17:41:18 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:18.937471 | orchestrator | 2025-08-29 17:41:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:21.984225 | orchestrator | 2025-08-29 17:41:21 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:21.986375 | orchestrator | 2025-08-29 17:41:21 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:21.990206 | orchestrator | 2025-08-29 17:41:21 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:21.990247 | orchestrator | 2025-08-29 17:41:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:25.038919 | orchestrator | 2025-08-29 17:41:25 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:25.043532 | orchestrator | 2025-08-29 17:41:25 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:25.045147 | orchestrator | 2025-08-29 17:41:25 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:25.045540 | orchestrator | 2025-08-29 17:41:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:28.096836 | orchestrator | 2025-08-29 17:41:28 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:28.099022 | orchestrator | 2025-08-29 17:41:28 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:28.100947 | orchestrator | 2025-08-29 17:41:28 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:28.100957 | orchestrator | 2025-08-29 17:41:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:31.155164 | orchestrator | 2025-08-29 17:41:31 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:31.157560 | orchestrator | 2025-08-29 17:41:31 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:31.159365 | orchestrator | 2025-08-29 17:41:31 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:31.159633 | orchestrator | 2025-08-29 17:41:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:34.206303 | orchestrator | 2025-08-29 17:41:34 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:34.207231 | orchestrator | 2025-08-29 17:41:34 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state STARTED 2025-08-29 17:41:34.208768 | orchestrator | 2025-08-29 17:41:34 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:34.208808 | orchestrator | 2025-08-29 17:41:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:37.265272 | orchestrator | 2025-08-29 17:41:37 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:37.267568 | orchestrator | 2025-08-29 17:41:37 | INFO  | Task 5b910b2d-09b3-4896-8194-a8a3048dc9db is in state SUCCESS 2025-08-29 17:41:37.269826 | orchestrator | 2025-08-29 17:41:37.269911 | orchestrator | 2025-08-29 17:41:37.269928 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:41:37.269940 | orchestrator | 2025-08-29 17:41:37.269952 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:41:37.269963 | orchestrator | Friday 29 August 2025 17:39:58 +0000 (0:00:00.192) 0:00:00.192 ********* 2025-08-29 17:41:37.269974 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.269986 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.269996 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.270007 | orchestrator | 2025-08-29 17:41:37.270076 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:41:37.270089 | orchestrator | Friday 29 August 2025 17:39:58 +0000 (0:00:00.235) 0:00:00.427 ********* 2025-08-29 17:41:37.270100 | orchestrator | ok: [testbed-node-0] => (item=enable_horizon_True) 2025-08-29 17:41:37.270112 | orchestrator | ok: [testbed-node-1] => (item=enable_horizon_True) 2025-08-29 17:41:37.270123 | orchestrator | ok: [testbed-node-2] => (item=enable_horizon_True) 2025-08-29 17:41:37.270135 | orchestrator | 2025-08-29 17:41:37.270146 | orchestrator | PLAY [Apply role horizon] ****************************************************** 2025-08-29 17:41:37.270157 | orchestrator | 2025-08-29 17:41:37.270168 | orchestrator | TASK [horizon : include_tasks] ************************************************* 2025-08-29 17:41:37.270196 | orchestrator | Friday 29 August 2025 17:39:58 +0000 (0:00:00.323) 0:00:00.751 ********* 2025-08-29 17:41:37.270207 | orchestrator | included: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:41:37.270219 | orchestrator | 2025-08-29 17:41:37.270230 | orchestrator | TASK [horizon : Ensuring config directories exist] ***************************** 2025-08-29 17:41:37.270241 | orchestrator | Friday 29 August 2025 17:39:59 +0000 (0:00:00.435) 0:00:01.186 ********* 2025-08-29 17:41:37.270366 | orchestrator | changed: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.270490 | orchestrator | changed: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.270593 | orchestrator | changed: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.270619 | orchestrator | 2025-08-29 17:41:37.270631 | orchestrator | TASK [horizon : Set empty custom policy] *************************************** 2025-08-29 17:41:37.270642 | orchestrator | Friday 29 August 2025 17:40:00 +0000 (0:00:00.894) 0:00:02.081 ********* 2025-08-29 17:41:37.270653 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.270664 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.270674 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.270685 | orchestrator | 2025-08-29 17:41:37.270696 | orchestrator | TASK [horizon : include_tasks] ************************************************* 2025-08-29 17:41:37.270706 | orchestrator | Friday 29 August 2025 17:40:00 +0000 (0:00:00.378) 0:00:02.459 ********* 2025-08-29 17:41:37.270725 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False})  2025-08-29 17:41:37.270737 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'})  2025-08-29 17:41:37.270748 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False})  2025-08-29 17:41:37.270758 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False})  2025-08-29 17:41:37.270769 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False})  2025-08-29 17:41:37.270780 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False})  2025-08-29 17:41:37.270791 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False})  2025-08-29 17:41:37.270802 | orchestrator | skipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False})  2025-08-29 17:41:37.270813 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False})  2025-08-29 17:41:37.270823 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'})  2025-08-29 17:41:37.270834 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False})  2025-08-29 17:41:37.270851 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False})  2025-08-29 17:41:37.270862 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False})  2025-08-29 17:41:37.270873 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False})  2025-08-29 17:41:37.270884 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False})  2025-08-29 17:41:37.270894 | orchestrator | skipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False})  2025-08-29 17:41:37.270905 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False})  2025-08-29 17:41:37.270916 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'})  2025-08-29 17:41:37.270933 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False})  2025-08-29 17:41:37.270944 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False})  2025-08-29 17:41:37.270955 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False})  2025-08-29 17:41:37.270965 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False})  2025-08-29 17:41:37.270976 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False})  2025-08-29 17:41:37.270987 | orchestrator | skipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False})  2025-08-29 17:41:37.270999 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'}) 2025-08-29 17:41:37.271012 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'}) 2025-08-29 17:41:37.271023 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True}) 2025-08-29 17:41:37.271034 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True}) 2025-08-29 17:41:37.271044 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True}) 2025-08-29 17:41:37.271055 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True}) 2025-08-29 17:41:37.271066 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True}) 2025-08-29 17:41:37.271077 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True}) 2025-08-29 17:41:37.271088 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True}) 2025-08-29 17:41:37.271100 | orchestrator | included: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True}) 2025-08-29 17:41:37.271111 | orchestrator | 2025-08-29 17:41:37.271122 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.271133 | orchestrator | Friday 29 August 2025 17:40:01 +0000 (0:00:00.626) 0:00:03.086 ********* 2025-08-29 17:41:37.271144 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.271154 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.271165 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.271176 | orchestrator | 2025-08-29 17:41:37.271187 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.271198 | orchestrator | Friday 29 August 2025 17:40:01 +0000 (0:00:00.260) 0:00:03.346 ********* 2025-08-29 17:41:37.271215 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271229 | orchestrator | 2025-08-29 17:41:37.271241 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.271253 | orchestrator | Friday 29 August 2025 17:40:01 +0000 (0:00:00.114) 0:00:03.461 ********* 2025-08-29 17:41:37.271265 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271277 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.271289 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.271300 | orchestrator | 2025-08-29 17:41:37.271312 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.271324 | orchestrator | Friday 29 August 2025 17:40:01 +0000 (0:00:00.372) 0:00:03.834 ********* 2025-08-29 17:41:37.271342 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.271355 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.271367 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.271379 | orchestrator | 2025-08-29 17:41:37.271391 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.271403 | orchestrator | Friday 29 August 2025 17:40:02 +0000 (0:00:00.259) 0:00:04.094 ********* 2025-08-29 17:41:37.271437 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271449 | orchestrator | 2025-08-29 17:41:37.271461 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.271478 | orchestrator | Friday 29 August 2025 17:40:02 +0000 (0:00:00.117) 0:00:04.211 ********* 2025-08-29 17:41:37.271491 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271503 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.271515 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.271528 | orchestrator | 2025-08-29 17:41:37.271540 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.271553 | orchestrator | Friday 29 August 2025 17:40:02 +0000 (0:00:00.244) 0:00:04.455 ********* 2025-08-29 17:41:37.271565 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.271576 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.271587 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.271598 | orchestrator | 2025-08-29 17:41:37.271609 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.271620 | orchestrator | Friday 29 August 2025 17:40:02 +0000 (0:00:00.255) 0:00:04.711 ********* 2025-08-29 17:41:37.271631 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271641 | orchestrator | 2025-08-29 17:41:37.271652 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.271663 | orchestrator | Friday 29 August 2025 17:40:03 +0000 (0:00:00.238) 0:00:04.950 ********* 2025-08-29 17:41:37.271674 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271684 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.271695 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.271706 | orchestrator | 2025-08-29 17:41:37.271716 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.271727 | orchestrator | Friday 29 August 2025 17:40:03 +0000 (0:00:00.303) 0:00:05.254 ********* 2025-08-29 17:41:37.271738 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.271748 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.271759 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.271770 | orchestrator | 2025-08-29 17:41:37.271780 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.271791 | orchestrator | Friday 29 August 2025 17:40:03 +0000 (0:00:00.292) 0:00:05.546 ********* 2025-08-29 17:41:37.271802 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271813 | orchestrator | 2025-08-29 17:41:37.271824 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.271834 | orchestrator | Friday 29 August 2025 17:40:03 +0000 (0:00:00.120) 0:00:05.667 ********* 2025-08-29 17:41:37.271845 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271856 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.271866 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.271877 | orchestrator | 2025-08-29 17:41:37.271888 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.271898 | orchestrator | Friday 29 August 2025 17:40:04 +0000 (0:00:00.291) 0:00:05.958 ********* 2025-08-29 17:41:37.271909 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.271920 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.271930 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.271941 | orchestrator | 2025-08-29 17:41:37.271952 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.271963 | orchestrator | Friday 29 August 2025 17:40:04 +0000 (0:00:00.486) 0:00:06.445 ********* 2025-08-29 17:41:37.271973 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.271990 | orchestrator | 2025-08-29 17:41:37.272001 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.272012 | orchestrator | Friday 29 August 2025 17:40:04 +0000 (0:00:00.123) 0:00:06.568 ********* 2025-08-29 17:41:37.272023 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272034 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.272044 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.272055 | orchestrator | 2025-08-29 17:41:37.272066 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.272077 | orchestrator | Friday 29 August 2025 17:40:04 +0000 (0:00:00.284) 0:00:06.852 ********* 2025-08-29 17:41:37.272088 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.272098 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.272109 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.272120 | orchestrator | 2025-08-29 17:41:37.272131 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.272142 | orchestrator | Friday 29 August 2025 17:40:05 +0000 (0:00:00.288) 0:00:07.141 ********* 2025-08-29 17:41:37.272152 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272163 | orchestrator | 2025-08-29 17:41:37.272174 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.272185 | orchestrator | Friday 29 August 2025 17:40:05 +0000 (0:00:00.113) 0:00:07.254 ********* 2025-08-29 17:41:37.272196 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272207 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.272217 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.272228 | orchestrator | 2025-08-29 17:41:37.272239 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.272256 | orchestrator | Friday 29 August 2025 17:40:05 +0000 (0:00:00.514) 0:00:07.768 ********* 2025-08-29 17:41:37.272267 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.272278 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.272288 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.272299 | orchestrator | 2025-08-29 17:41:37.272310 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.272321 | orchestrator | Friday 29 August 2025 17:40:06 +0000 (0:00:00.309) 0:00:08.078 ********* 2025-08-29 17:41:37.272332 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272343 | orchestrator | 2025-08-29 17:41:37.272353 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.272364 | orchestrator | Friday 29 August 2025 17:40:06 +0000 (0:00:00.144) 0:00:08.223 ********* 2025-08-29 17:41:37.272375 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272386 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.272396 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.272407 | orchestrator | 2025-08-29 17:41:37.272476 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.272504 | orchestrator | Friday 29 August 2025 17:40:06 +0000 (0:00:00.277) 0:00:08.500 ********* 2025-08-29 17:41:37.272522 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.272538 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.272555 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.272573 | orchestrator | 2025-08-29 17:41:37.272597 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.272615 | orchestrator | Friday 29 August 2025 17:40:06 +0000 (0:00:00.315) 0:00:08.816 ********* 2025-08-29 17:41:37.272632 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272650 | orchestrator | 2025-08-29 17:41:37.272670 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.272687 | orchestrator | Friday 29 August 2025 17:40:07 +0000 (0:00:00.125) 0:00:08.941 ********* 2025-08-29 17:41:37.272705 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272717 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.272727 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.272748 | orchestrator | 2025-08-29 17:41:37.272759 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.272770 | orchestrator | Friday 29 August 2025 17:40:07 +0000 (0:00:00.399) 0:00:09.341 ********* 2025-08-29 17:41:37.272781 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.272791 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.272802 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.272812 | orchestrator | 2025-08-29 17:41:37.272823 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.272834 | orchestrator | Friday 29 August 2025 17:40:07 +0000 (0:00:00.282) 0:00:09.623 ********* 2025-08-29 17:41:37.272844 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272855 | orchestrator | 2025-08-29 17:41:37.272866 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.272877 | orchestrator | Friday 29 August 2025 17:40:07 +0000 (0:00:00.109) 0:00:09.733 ********* 2025-08-29 17:41:37.272887 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.272898 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.272909 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.272919 | orchestrator | 2025-08-29 17:41:37.272930 | orchestrator | TASK [horizon : Update policy file name] *************************************** 2025-08-29 17:41:37.272941 | orchestrator | Friday 29 August 2025 17:40:08 +0000 (0:00:00.290) 0:00:10.024 ********* 2025-08-29 17:41:37.272951 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:41:37.272962 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:41:37.272973 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:41:37.272983 | orchestrator | 2025-08-29 17:41:37.272994 | orchestrator | TASK [horizon : Check if policies shall be overwritten] ************************ 2025-08-29 17:41:37.273005 | orchestrator | Friday 29 August 2025 17:40:08 +0000 (0:00:00.435) 0:00:10.459 ********* 2025-08-29 17:41:37.273020 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.273039 | orchestrator | 2025-08-29 17:41:37.273056 | orchestrator | TASK [horizon : Update custom policy file name] ******************************** 2025-08-29 17:41:37.273074 | orchestrator | Friday 29 August 2025 17:40:08 +0000 (0:00:00.115) 0:00:10.574 ********* 2025-08-29 17:41:37.273090 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.273108 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.273126 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.273145 | orchestrator | 2025-08-29 17:41:37.273163 | orchestrator | TASK [horizon : Copying over config.json files for services] ******************* 2025-08-29 17:41:37.273183 | orchestrator | Friday 29 August 2025 17:40:08 +0000 (0:00:00.267) 0:00:10.841 ********* 2025-08-29 17:41:37.273194 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:41:37.273205 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:41:37.273215 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:41:37.273230 | orchestrator | 2025-08-29 17:41:37.273249 | orchestrator | TASK [horizon : Copying over horizon.conf] ************************************* 2025-08-29 17:41:37.273266 | orchestrator | Friday 29 August 2025 17:40:10 +0000 (0:00:01.620) 0:00:12.462 ********* 2025-08-29 17:41:37.273283 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2) 2025-08-29 17:41:37.273300 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2) 2025-08-29 17:41:37.273318 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2) 2025-08-29 17:41:37.273337 | orchestrator | 2025-08-29 17:41:37.273356 | orchestrator | TASK [horizon : Copying over kolla-settings.py] ******************************** 2025-08-29 17:41:37.273373 | orchestrator | Friday 29 August 2025 17:40:12 +0000 (0:00:01.618) 0:00:14.081 ********* 2025-08-29 17:41:37.273392 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2) 2025-08-29 17:41:37.273404 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2) 2025-08-29 17:41:37.273438 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2) 2025-08-29 17:41:37.273484 | orchestrator | 2025-08-29 17:41:37.273505 | orchestrator | TASK [horizon : Copying over custom-settings.py] ******************************* 2025-08-29 17:41:37.273517 | orchestrator | Friday 29 August 2025 17:40:14 +0000 (0:00:02.510) 0:00:16.591 ********* 2025-08-29 17:41:37.273528 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2) 2025-08-29 17:41:37.273539 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2) 2025-08-29 17:41:37.273550 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2) 2025-08-29 17:41:37.273561 | orchestrator | 2025-08-29 17:41:37.273572 | orchestrator | TASK [horizon : Copying over existing policy file] ***************************** 2025-08-29 17:41:37.273582 | orchestrator | Friday 29 August 2025 17:40:16 +0000 (0:00:01.604) 0:00:18.195 ********* 2025-08-29 17:41:37.273593 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.273604 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.273615 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.273628 | orchestrator | 2025-08-29 17:41:37.273646 | orchestrator | TASK [horizon : Copying over custom themes] ************************************ 2025-08-29 17:41:37.273664 | orchestrator | Friday 29 August 2025 17:40:16 +0000 (0:00:00.252) 0:00:18.448 ********* 2025-08-29 17:41:37.273682 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.273706 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.273723 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.273740 | orchestrator | 2025-08-29 17:41:37.273756 | orchestrator | TASK [horizon : include_tasks] ************************************************* 2025-08-29 17:41:37.273773 | orchestrator | Friday 29 August 2025 17:40:16 +0000 (0:00:00.246) 0:00:18.695 ********* 2025-08-29 17:41:37.273791 | orchestrator | included: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:41:37.273808 | orchestrator | 2025-08-29 17:41:37.273824 | orchestrator | TASK [service-cert-copy : horizon | Copying over extra CA certificates] ******** 2025-08-29 17:41:37.273840 | orchestrator | Friday 29 August 2025 17:40:17 +0000 (0:00:00.672) 0:00:19.367 ********* 2025-08-29 17:41:37.273861 | orchestrator | changed: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.273929 | orchestrator | changed: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.273952 | orchestrator | changed: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.274008 | orchestrator | 2025-08-29 17:41:37.274077 | orchestrator | TASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] *** 2025-08-29 17:41:37.274098 | orchestrator | Friday 29 August 2025 17:40:19 +0000 (0:00:01.624) 0:00:20.992 ********* 2025-08-29 17:41:37.274143 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:41:37.274166 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.274198 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:41:37.274246 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.274274 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:41:37.274295 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.274314 | orchestrator | 2025-08-29 17:41:37.274333 | orchestrator | TASK [service-cert-copy : horizon | Copying over backend internal TLS key] ***** 2025-08-29 17:41:37.274352 | orchestrator | Friday 29 August 2025 17:40:19 +0000 (0:00:00.558) 0:00:21.551 ********* 2025-08-29 17:41:37.274383 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:41:37.274437 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.274465 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:41:37.274494 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.274529 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})  2025-08-29 17:41:37.274550 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.274568 | orchestrator | 2025-08-29 17:41:37.274585 | orchestrator | TASK [horizon : Deploy horizon container] ************************************** 2025-08-29 17:41:37.274602 | orchestrator | Friday 29 August 2025 17:40:20 +0000 (0:00:01.042) 0:00:22.594 ********* 2025-08-29 17:41:37.274623 | orchestrator | changed: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.274675 | orchestrator | changed: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.274697 | orchestrator | changed: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/horizon:25.1.1.20250711', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) 2025-08-29 17:41:37.274731 | orchestrator | 2025-08-29 17:41:37.274750 | orchestrator | TASK [horizon : include_tasks] ************************************************* 2025-08-29 17:41:37.274768 | orchestrator | Friday 29 August 2025 17:40:21 +0000 (0:00:01.188) 0:00:23.783 ********* 2025-08-29 17:41:37.274786 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:41:37.274804 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:41:37.274823 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:41:37.274841 | orchestrator | 2025-08-29 17:41:37.274859 | orchestrator | TASK [horizon : include_tasks] ************************************************* 2025-08-29 17:41:37.274886 | orchestrator | Friday 29 August 2025 17:40:22 +0000 (0:00:00.294) 0:00:24.077 ********* 2025-08-29 17:41:37.274905 | orchestrator | included: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:41:37.274923 | orchestrator | 2025-08-29 17:41:37.274942 | orchestrator | TASK [horizon : Creating Horizon database] ************************************* 2025-08-29 17:41:37.274961 | orchestrator | Friday 29 August 2025 17:40:22 +0000 (0:00:00.578) 0:00:24.656 ********* 2025-08-29 17:41:37.274979 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:41:37.274997 | orchestrator | 2025-08-29 17:41:37.275015 | orchestrator | TASK [horizon : Creating Horizon database user and setting permissions] ******** 2025-08-29 17:41:37.275032 | orchestrator | Friday 29 August 2025 17:40:25 +0000 (0:00:02.227) 0:00:26.883 ********* 2025-08-29 17:41:37.275048 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:41:37.275067 | orchestrator | 2025-08-29 17:41:37.275084 | orchestrator | TASK [horizon : Running Horizon bootstrap container] *************************** 2025-08-29 17:41:37.275101 | orchestrator | Friday 29 August 2025 17:40:27 +0000 (0:00:02.156) 0:00:29.039 ********* 2025-08-29 17:41:37.275117 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:41:37.275133 | orchestrator | 2025-08-29 17:41:37.275150 | orchestrator | TASK [horizon : Flush handlers] ************************************************ 2025-08-29 17:41:37.275165 | orchestrator | Friday 29 August 2025 17:40:42 +0000 (0:00:15.582) 0:00:44.622 ********* 2025-08-29 17:41:37.275181 | orchestrator | 2025-08-29 17:41:37.275229 | orchestrator | TASK [horizon : Flush handlers] ************************************************ 2025-08-29 17:41:37.275247 | orchestrator | Friday 29 August 2025 17:40:42 +0000 (0:00:00.058) 0:00:44.681 ********* 2025-08-29 17:41:37.275264 | orchestrator | 2025-08-29 17:41:37.275281 | orchestrator | TASK [horizon : Flush handlers] ************************************************ 2025-08-29 17:41:37.275298 | orchestrator | Friday 29 August 2025 17:40:42 +0000 (0:00:00.058) 0:00:44.739 ********* 2025-08-29 17:41:37.275315 | orchestrator | 2025-08-29 17:41:37.275333 | orchestrator | RUNNING HANDLER [horizon : Restart horizon container] ************************** 2025-08-29 17:41:37.275350 | orchestrator | Friday 29 August 2025 17:40:42 +0000 (0:00:00.059) 0:00:44.798 ********* 2025-08-29 17:41:37.275367 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:41:37.275383 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:41:37.275401 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:41:37.275523 | orchestrator | 2025-08-29 17:41:37.275544 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:41:37.275563 | orchestrator | testbed-node-0 : ok=37  changed=11  unreachable=0 failed=0 skipped=25  rescued=0 ignored=0 2025-08-29 17:41:37.275580 | orchestrator | testbed-node-1 : ok=34  changed=8  unreachable=0 failed=0 skipped=15  rescued=0 ignored=0 2025-08-29 17:41:37.275596 | orchestrator | testbed-node-2 : ok=34  changed=8  unreachable=0 failed=0 skipped=15  rescued=0 ignored=0 2025-08-29 17:41:37.275612 | orchestrator | 2025-08-29 17:41:37.275629 | orchestrator | 2025-08-29 17:41:37.275646 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:41:37.275664 | orchestrator | Friday 29 August 2025 17:41:35 +0000 (0:00:52.832) 0:01:37.630 ********* 2025-08-29 17:41:37.275682 | orchestrator | =============================================================================== 2025-08-29 17:41:37.275698 | orchestrator | horizon : Restart horizon container ------------------------------------ 52.83s 2025-08-29 17:41:37.275717 | orchestrator | horizon : Running Horizon bootstrap container -------------------------- 15.58s 2025-08-29 17:41:37.275734 | orchestrator | horizon : Copying over kolla-settings.py -------------------------------- 2.51s 2025-08-29 17:41:37.275750 | orchestrator | horizon : Creating Horizon database ------------------------------------- 2.23s 2025-08-29 17:41:37.275767 | orchestrator | horizon : Creating Horizon database user and setting permissions -------- 2.16s 2025-08-29 17:41:37.275784 | orchestrator | service-cert-copy : horizon | Copying over extra CA certificates -------- 1.62s 2025-08-29 17:41:37.275800 | orchestrator | horizon : Copying over config.json files for services ------------------- 1.62s 2025-08-29 17:41:37.275817 | orchestrator | horizon : Copying over horizon.conf ------------------------------------- 1.62s 2025-08-29 17:41:37.275831 | orchestrator | horizon : Copying over custom-settings.py ------------------------------- 1.60s 2025-08-29 17:41:37.275847 | orchestrator | horizon : Deploy horizon container -------------------------------------- 1.19s 2025-08-29 17:41:37.275863 | orchestrator | service-cert-copy : horizon | Copying over backend internal TLS key ----- 1.04s 2025-08-29 17:41:37.275877 | orchestrator | horizon : Ensuring config directories exist ----------------------------- 0.89s 2025-08-29 17:41:37.275893 | orchestrator | horizon : include_tasks ------------------------------------------------- 0.67s 2025-08-29 17:41:37.275910 | orchestrator | horizon : include_tasks ------------------------------------------------- 0.63s 2025-08-29 17:41:37.275927 | orchestrator | horizon : include_tasks ------------------------------------------------- 0.58s 2025-08-29 17:41:37.275944 | orchestrator | service-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.56s 2025-08-29 17:41:37.275961 | orchestrator | horizon : Update custom policy file name -------------------------------- 0.51s 2025-08-29 17:41:37.275978 | orchestrator | horizon : Update policy file name --------------------------------------- 0.49s 2025-08-29 17:41:37.275995 | orchestrator | horizon : include_tasks ------------------------------------------------- 0.44s 2025-08-29 17:41:37.276013 | orchestrator | horizon : Update policy file name --------------------------------------- 0.44s 2025-08-29 17:41:37.276029 | orchestrator | 2025-08-29 17:41:37 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:37.276059 | orchestrator | 2025-08-29 17:41:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:40.323814 | orchestrator | 2025-08-29 17:41:40 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:40.326378 | orchestrator | 2025-08-29 17:41:40 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:40.326504 | orchestrator | 2025-08-29 17:41:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:43.366921 | orchestrator | 2025-08-29 17:41:43 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:43.367064 | orchestrator | 2025-08-29 17:41:43 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state STARTED 2025-08-29 17:41:43.367082 | orchestrator | 2025-08-29 17:41:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:46.416054 | orchestrator | 2025-08-29 17:41:46 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:46.417671 | orchestrator | 2025-08-29 17:41:46 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:41:46.420739 | orchestrator | 2025-08-29 17:41:46 | INFO  | Task 4b8e9bf9-f296-478a-aca8-2afef083ab7b is in state SUCCESS 2025-08-29 17:41:46.420821 | orchestrator | 2025-08-29 17:41:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:49.473672 | orchestrator | 2025-08-29 17:41:49 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:49.478389 | orchestrator | 2025-08-29 17:41:49 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:41:49.479664 | orchestrator | 2025-08-29 17:41:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:52.524379 | orchestrator | 2025-08-29 17:41:52 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:52.524556 | orchestrator | 2025-08-29 17:41:52 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:41:52.524574 | orchestrator | 2025-08-29 17:41:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:55.574084 | orchestrator | 2025-08-29 17:41:55 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:55.576484 | orchestrator | 2025-08-29 17:41:55 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:41:55.576525 | orchestrator | 2025-08-29 17:41:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:41:58.625302 | orchestrator | 2025-08-29 17:41:58 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:41:58.627585 | orchestrator | 2025-08-29 17:41:58 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:41:58.627986 | orchestrator | 2025-08-29 17:41:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:01.669198 | orchestrator | 2025-08-29 17:42:01 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:01.671609 | orchestrator | 2025-08-29 17:42:01 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:01.671650 | orchestrator | 2025-08-29 17:42:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:04.717117 | orchestrator | 2025-08-29 17:42:04 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:04.719499 | orchestrator | 2025-08-29 17:42:04 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:04.719532 | orchestrator | 2025-08-29 17:42:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:07.770830 | orchestrator | 2025-08-29 17:42:07 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:07.771542 | orchestrator | 2025-08-29 17:42:07 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:07.771833 | orchestrator | 2025-08-29 17:42:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:10.806518 | orchestrator | 2025-08-29 17:42:10 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:10.807917 | orchestrator | 2025-08-29 17:42:10 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:10.808049 | orchestrator | 2025-08-29 17:42:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:13.839122 | orchestrator | 2025-08-29 17:42:13 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:13.840327 | orchestrator | 2025-08-29 17:42:13 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:13.840378 | orchestrator | 2025-08-29 17:42:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:16.886960 | orchestrator | 2025-08-29 17:42:16 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:16.887778 | orchestrator | 2025-08-29 17:42:16 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:16.887923 | orchestrator | 2025-08-29 17:42:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:19.933767 | orchestrator | 2025-08-29 17:42:19 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:19.935312 | orchestrator | 2025-08-29 17:42:19 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:19.935355 | orchestrator | 2025-08-29 17:42:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:22.975980 | orchestrator | 2025-08-29 17:42:22 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:22.976996 | orchestrator | 2025-08-29 17:42:22 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:22.977057 | orchestrator | 2025-08-29 17:42:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:26.021502 | orchestrator | 2025-08-29 17:42:26 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:26.022227 | orchestrator | 2025-08-29 17:42:26 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:26.022264 | orchestrator | 2025-08-29 17:42:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:29.062964 | orchestrator | 2025-08-29 17:42:29 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:29.065240 | orchestrator | 2025-08-29 17:42:29 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:29.065685 | orchestrator | 2025-08-29 17:42:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:32.114787 | orchestrator | 2025-08-29 17:42:32 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state STARTED 2025-08-29 17:42:32.115337 | orchestrator | 2025-08-29 17:42:32 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:32.115620 | orchestrator | 2025-08-29 17:42:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:35.156491 | orchestrator | 2025-08-29 17:42:35 | INFO  | Task d66407d9-4347-40e1-b4a7-fe23a391f4b3 is in state SUCCESS 2025-08-29 17:42:35.160794 | orchestrator | 2025-08-29 17:42:35.160884 | orchestrator | 2025-08-29 17:42:35.160900 | orchestrator | PLAY [Copy ceph keys to the configuration repository] ************************** 2025-08-29 17:42:35.160912 | orchestrator | 2025-08-29 17:42:35.160924 | orchestrator | TASK [Fetch all ceph keys] ***************************************************** 2025-08-29 17:42:35.160936 | orchestrator | Friday 29 August 2025 17:41:17 +0000 (0:00:00.149) 0:00:00.149 ********* 2025-08-29 17:42:35.160948 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring) 2025-08-29 17:42:35.160981 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.160993 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161027 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring) 2025-08-29 17:42:35.161038 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161050 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring) 2025-08-29 17:42:35.161060 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring) 2025-08-29 17:42:35.161071 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring) 2025-08-29 17:42:35.161082 | orchestrator | ok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring) 2025-08-29 17:42:35.161093 | orchestrator | 2025-08-29 17:42:35.161104 | orchestrator | TASK [Create share directory] ************************************************** 2025-08-29 17:42:35.161115 | orchestrator | Friday 29 August 2025 17:41:21 +0000 (0:00:04.301) 0:00:04.450 ********* 2025-08-29 17:42:35.161126 | orchestrator | changed: [testbed-manager -> localhost] 2025-08-29 17:42:35.161137 | orchestrator | 2025-08-29 17:42:35.161147 | orchestrator | TASK [Write ceph keys to the share directory] ********************************** 2025-08-29 17:42:35.161158 | orchestrator | Friday 29 August 2025 17:41:22 +0000 (0:00:01.006) 0:00:05.457 ********* 2025-08-29 17:42:35.161169 | orchestrator | changed: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring) 2025-08-29 17:42:35.161180 | orchestrator | changed: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161190 | orchestrator | ok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161201 | orchestrator | changed: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring) 2025-08-29 17:42:35.161212 | orchestrator | ok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161222 | orchestrator | changed: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring) 2025-08-29 17:42:35.161233 | orchestrator | changed: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring) 2025-08-29 17:42:35.161243 | orchestrator | changed: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring) 2025-08-29 17:42:35.161254 | orchestrator | changed: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring) 2025-08-29 17:42:35.161265 | orchestrator | 2025-08-29 17:42:35.161275 | orchestrator | TASK [Write ceph keys to the configuration directory] ************************** 2025-08-29 17:42:35.161286 | orchestrator | Friday 29 August 2025 17:41:36 +0000 (0:00:13.568) 0:00:19.026 ********* 2025-08-29 17:42:35.161311 | orchestrator | changed: [testbed-manager] => (item=ceph.client.admin.keyring) 2025-08-29 17:42:35.161325 | orchestrator | changed: [testbed-manager] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161337 | orchestrator | changed: [testbed-manager] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161357 | orchestrator | changed: [testbed-manager] => (item=ceph.client.cinder-backup.keyring) 2025-08-29 17:42:35.161371 | orchestrator | changed: [testbed-manager] => (item=ceph.client.cinder.keyring) 2025-08-29 17:42:35.161383 | orchestrator | changed: [testbed-manager] => (item=ceph.client.nova.keyring) 2025-08-29 17:42:35.161395 | orchestrator | changed: [testbed-manager] => (item=ceph.client.glance.keyring) 2025-08-29 17:42:35.161407 | orchestrator | changed: [testbed-manager] => (item=ceph.client.gnocchi.keyring) 2025-08-29 17:42:35.161420 | orchestrator | changed: [testbed-manager] => (item=ceph.client.manila.keyring) 2025-08-29 17:42:35.161467 | orchestrator | 2025-08-29 17:42:35.161480 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:42:35.161492 | orchestrator | testbed-manager : ok=4  changed=3  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:42:35.161506 | orchestrator | 2025-08-29 17:42:35.161518 | orchestrator | 2025-08-29 17:42:35.161529 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:42:35.161549 | orchestrator | Friday 29 August 2025 17:41:43 +0000 (0:00:07.200) 0:00:26.226 ********* 2025-08-29 17:42:35.161561 | orchestrator | =============================================================================== 2025-08-29 17:42:35.161574 | orchestrator | Write ceph keys to the share directory --------------------------------- 13.57s 2025-08-29 17:42:35.161587 | orchestrator | Write ceph keys to the configuration directory -------------------------- 7.20s 2025-08-29 17:42:35.161599 | orchestrator | Fetch all ceph keys ----------------------------------------------------- 4.30s 2025-08-29 17:42:35.161610 | orchestrator | Create share directory -------------------------------------------------- 1.01s 2025-08-29 17:42:35.161622 | orchestrator | 2025-08-29 17:42:35.161634 | orchestrator | 2025-08-29 17:42:35.161647 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:42:35.161659 | orchestrator | 2025-08-29 17:42:35.161720 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:42:35.161733 | orchestrator | Friday 29 August 2025 17:39:58 +0000 (0:00:00.229) 0:00:00.229 ********* 2025-08-29 17:42:35.161745 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:42:35.161756 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:42:35.161767 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:42:35.161778 | orchestrator | 2025-08-29 17:42:35.161789 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:42:35.161799 | orchestrator | Friday 29 August 2025 17:39:58 +0000 (0:00:00.252) 0:00:00.481 ********* 2025-08-29 17:42:35.161810 | orchestrator | ok: [testbed-node-0] => (item=enable_keystone_True) 2025-08-29 17:42:35.161837 | orchestrator | ok: [testbed-node-1] => (item=enable_keystone_True) 2025-08-29 17:42:35.161849 | orchestrator | ok: [testbed-node-2] => (item=enable_keystone_True) 2025-08-29 17:42:35.161860 | orchestrator | 2025-08-29 17:42:35.161871 | orchestrator | PLAY [Apply role keystone] ***************************************************** 2025-08-29 17:42:35.161881 | orchestrator | 2025-08-29 17:42:35.161892 | orchestrator | TASK [keystone : include_tasks] ************************************************ 2025-08-29 17:42:35.161903 | orchestrator | Friday 29 August 2025 17:39:59 +0000 (0:00:00.350) 0:00:00.831 ********* 2025-08-29 17:42:35.161914 | orchestrator | included: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:42:35.161925 | orchestrator | 2025-08-29 17:42:35.161936 | orchestrator | TASK [keystone : Ensuring config directories exist] **************************** 2025-08-29 17:42:35.161947 | orchestrator | Friday 29 August 2025 17:39:59 +0000 (0:00:00.474) 0:00:01.306 ********* 2025-08-29 17:42:35.161977 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.162011 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.162166 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.162183 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162196 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162208 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162220 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162268 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162282 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162294 | orchestrator | 2025-08-29 17:42:35.162305 | orchestrator | TASK [keystone : Check if policies shall be overwritten] *********************** 2025-08-29 17:42:35.162316 | orchestrator | Friday 29 August 2025 17:40:01 +0000 (0:00:01.605) 0:00:02.912 ********* 2025-08-29 17:42:35.162334 | orchestrator | ok: [testbed-node-0 -> localhost] => (item=/opt/configuration/environments/kolla/files/overlays/keystone/policy.yaml) 2025-08-29 17:42:35.162346 | orchestrator | 2025-08-29 17:42:35.162357 | orchestrator | TASK [keystone : Set keystone policy file] ************************************* 2025-08-29 17:42:35.162368 | orchestrator | Friday 29 August 2025 17:40:01 +0000 (0:00:00.737) 0:00:03.650 ********* 2025-08-29 17:42:35.162378 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:42:35.162389 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:42:35.162400 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:42:35.162411 | orchestrator | 2025-08-29 17:42:35.162482 | orchestrator | TASK [keystone : Check if Keystone domain-specific config is supplied] ********* 2025-08-29 17:42:35.162496 | orchestrator | Friday 29 August 2025 17:40:02 +0000 (0:00:00.382) 0:00:04.032 ********* 2025-08-29 17:42:35.162507 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:42:35.162517 | orchestrator | 2025-08-29 17:42:35.162528 | orchestrator | TASK [keystone : include_tasks] ************************************************ 2025-08-29 17:42:35.162539 | orchestrator | Friday 29 August 2025 17:40:02 +0000 (0:00:00.618) 0:00:04.650 ********* 2025-08-29 17:42:35.162550 | orchestrator | included: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:42:35.162561 | orchestrator | 2025-08-29 17:42:35.162571 | orchestrator | TASK [service-cert-copy : keystone | Copying over extra CA certificates] ******* 2025-08-29 17:42:35.162582 | orchestrator | Friday 29 August 2025 17:40:03 +0000 (0:00:00.510) 0:00:05.160 ********* 2025-08-29 17:42:35.162594 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.162620 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.162634 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.162656 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162669 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162680 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162698 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162714 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162726 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.162737 | orchestrator | 2025-08-29 17:42:35.162748 | orchestrator | TASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] *** 2025-08-29 17:42:35.162759 | orchestrator | Friday 29 August 2025 17:40:06 +0000 (0:00:03.415) 0:00:08.576 ********* 2025-08-29 17:42:35.162779 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:42:35.162792 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.162808 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:42:35.162818 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.162833 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:42:35.162844 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.162861 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:42:35.162871 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.162882 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:42:35.162898 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.162909 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:42:35.162919 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.162929 | orchestrator | 2025-08-29 17:42:35.162938 | orchestrator | TASK [service-cert-copy : keystone | Copying over backend internal TLS key] **** 2025-08-29 17:42:35.162952 | orchestrator | Friday 29 August 2025 17:40:07 +0000 (0:00:00.491) 0:00:09.068 ********* 2025-08-29 17:42:35.162963 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:42:35.162996 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163008 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:42:35.163023 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.163034 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:42:35.163044 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163059 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:42:35.163069 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.163086 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})  2025-08-29 17:42:35.163098 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163113 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})  2025-08-29 17:42:35.163123 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.163133 | orchestrator | 2025-08-29 17:42:35.163143 | orchestrator | TASK [keystone : Copying over config.json files for services] ****************** 2025-08-29 17:42:35.163153 | orchestrator | Friday 29 August 2025 17:40:08 +0000 (0:00:00.733) 0:00:09.801 ********* 2025-08-29 17:42:35.163168 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163179 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163197 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163213 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163223 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163234 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163248 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163259 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163276 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163294 | orchestrator | 2025-08-29 17:42:35.163304 | orchestrator | TASK [keystone : Copying over keystone.conf] *********************************** 2025-08-29 17:42:35.163314 | orchestrator | Friday 29 August 2025 17:40:11 +0000 (0:00:03.304) 0:00:13.106 ********* 2025-08-29 17:42:35.163325 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163336 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163351 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163362 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163378 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163394 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163404 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163415 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163444 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163455 | orchestrator | 2025-08-29 17:42:35.163465 | orchestrator | TASK [keystone : Copying keystone-startup script for keystone] ***************** 2025-08-29 17:42:35.163475 | orchestrator | Friday 29 August 2025 17:40:16 +0000 (0:00:04.853) 0:00:17.959 ********* 2025-08-29 17:42:35.163485 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.163494 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:42:35.163504 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:42:35.163514 | orchestrator | 2025-08-29 17:42:35.163524 | orchestrator | TASK [keystone : Create Keystone domain-specific config directory] ************* 2025-08-29 17:42:35.163543 | orchestrator | Friday 29 August 2025 17:40:17 +0000 (0:00:01.329) 0:00:19.289 ********* 2025-08-29 17:42:35.163553 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.163562 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.163572 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.163582 | orchestrator | 2025-08-29 17:42:35.163591 | orchestrator | TASK [keystone : Get file list in custom domains folder] *********************** 2025-08-29 17:42:35.163606 | orchestrator | Friday 29 August 2025 17:40:18 +0000 (0:00:00.497) 0:00:19.786 ********* 2025-08-29 17:42:35.163617 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.163626 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.163636 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.163649 | orchestrator | 2025-08-29 17:42:35.163666 | orchestrator | TASK [keystone : Copying Keystone Domain specific settings] ******************** 2025-08-29 17:42:35.163676 | orchestrator | Friday 29 August 2025 17:40:18 +0000 (0:00:00.254) 0:00:20.041 ********* 2025-08-29 17:42:35.163685 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.163695 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.163704 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.163714 | orchestrator | 2025-08-29 17:42:35.163724 | orchestrator | TASK [keystone : Copying over existing policy file] **************************** 2025-08-29 17:42:35.163733 | orchestrator | Friday 29 August 2025 17:40:18 +0000 (0:00:00.389) 0:00:20.430 ********* 2025-08-29 17:42:35.163744 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163755 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163769 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163787 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163804 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.163815 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})  2025-08-29 17:42:35.163826 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163836 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163851 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.163867 | orchestrator | 2025-08-29 17:42:35.163877 | orchestrator | TASK [keystone : include_tasks] ************************************************ 2025-08-29 17:42:35.163887 | orchestrator | Friday 29 August 2025 17:40:20 +0000 (0:00:02.154) 0:00:22.585 ********* 2025-08-29 17:42:35.163897 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.163907 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.163916 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.163926 | orchestrator | 2025-08-29 17:42:35.163936 | orchestrator | TASK [keystone : Copying over wsgi-keystone.conf] ****************************** 2025-08-29 17:42:35.163945 | orchestrator | Friday 29 August 2025 17:40:21 +0000 (0:00:00.304) 0:00:22.889 ********* 2025-08-29 17:42:35.163955 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2) 2025-08-29 17:42:35.163964 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2) 2025-08-29 17:42:35.163974 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2) 2025-08-29 17:42:35.163984 | orchestrator | 2025-08-29 17:42:35.163998 | orchestrator | TASK [keystone : Checking whether keystone-paste.ini file exists] ************** 2025-08-29 17:42:35.164008 | orchestrator | Friday 29 August 2025 17:40:22 +0000 (0:00:01.555) 0:00:24.445 ********* 2025-08-29 17:42:35.164018 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:42:35.164027 | orchestrator | 2025-08-29 17:42:35.164037 | orchestrator | TASK [keystone : Copying over keystone-paste.ini] ****************************** 2025-08-29 17:42:35.164047 | orchestrator | Friday 29 August 2025 17:40:23 +0000 (0:00:01.087) 0:00:25.533 ********* 2025-08-29 17:42:35.164056 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.164066 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.164075 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.164085 | orchestrator | 2025-08-29 17:42:35.164094 | orchestrator | TASK [keystone : Generate the required cron jobs for the node] ***************** 2025-08-29 17:42:35.164104 | orchestrator | Friday 29 August 2025 17:40:24 +0000 (0:00:00.469) 0:00:26.002 ********* 2025-08-29 17:42:35.164114 | orchestrator | ok: [testbed-node-1 -> localhost] 2025-08-29 17:42:35.164124 | orchestrator | ok: [testbed-node-2 -> localhost] 2025-08-29 17:42:35.164133 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:42:35.164143 | orchestrator | 2025-08-29 17:42:35.164152 | orchestrator | TASK [keystone : Set fact with the generated cron jobs for building the crontab later] *** 2025-08-29 17:42:35.164162 | orchestrator | Friday 29 August 2025 17:40:25 +0000 (0:00:00.851) 0:00:26.854 ********* 2025-08-29 17:42:35.164171 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:42:35.164181 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:42:35.164190 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:42:35.164200 | orchestrator | 2025-08-29 17:42:35.164210 | orchestrator | TASK [keystone : Copying files for keystone-fernet] **************************** 2025-08-29 17:42:35.164219 | orchestrator | Friday 29 August 2025 17:40:25 +0000 (0:00:00.243) 0:00:27.098 ********* 2025-08-29 17:42:35.164229 | orchestrator | changed: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'}) 2025-08-29 17:42:35.164239 | orchestrator | changed: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'}) 2025-08-29 17:42:35.164248 | orchestrator | changed: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'}) 2025-08-29 17:42:35.164258 | orchestrator | changed: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'}) 2025-08-29 17:42:35.164268 | orchestrator | changed: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'}) 2025-08-29 17:42:35.164278 | orchestrator | changed: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'}) 2025-08-29 17:42:35.164293 | orchestrator | changed: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'}) 2025-08-29 17:42:35.164303 | orchestrator | changed: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'}) 2025-08-29 17:42:35.164313 | orchestrator | changed: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'}) 2025-08-29 17:42:35.164323 | orchestrator | changed: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'}) 2025-08-29 17:42:35.164332 | orchestrator | changed: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'}) 2025-08-29 17:42:35.164342 | orchestrator | changed: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'}) 2025-08-29 17:42:35.164352 | orchestrator | changed: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'}) 2025-08-29 17:42:35.164362 | orchestrator | changed: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'}) 2025-08-29 17:42:35.164372 | orchestrator | changed: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'}) 2025-08-29 17:42:35.164386 | orchestrator | changed: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) 2025-08-29 17:42:35.164396 | orchestrator | changed: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) 2025-08-29 17:42:35.164406 | orchestrator | changed: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) 2025-08-29 17:42:35.164415 | orchestrator | changed: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) 2025-08-29 17:42:35.164472 | orchestrator | changed: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) 2025-08-29 17:42:35.164484 | orchestrator | changed: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) 2025-08-29 17:42:35.164494 | orchestrator | 2025-08-29 17:42:35.164504 | orchestrator | TASK [keystone : Copying files for keystone-ssh] ******************************* 2025-08-29 17:42:35.164513 | orchestrator | Friday 29 August 2025 17:40:34 +0000 (0:00:08.775) 0:00:35.873 ********* 2025-08-29 17:42:35.164523 | orchestrator | changed: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) 2025-08-29 17:42:35.164533 | orchestrator | changed: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) 2025-08-29 17:42:35.164542 | orchestrator | changed: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) 2025-08-29 17:42:35.164552 | orchestrator | changed: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) 2025-08-29 17:42:35.164561 | orchestrator | changed: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) 2025-08-29 17:42:35.164571 | orchestrator | changed: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) 2025-08-29 17:42:35.164580 | orchestrator | 2025-08-29 17:42:35.164590 | orchestrator | TASK [keystone : Check keystone containers] ************************************ 2025-08-29 17:42:35.164606 | orchestrator | Friday 29 August 2025 17:40:36 +0000 (0:00:02.743) 0:00:38.617 ********* 2025-08-29 17:42:35.164617 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.164635 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.164651 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) 2025-08-29 17:42:35.164663 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.164705 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.164717 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) 2025-08-29 17:42:35.164733 | orchestrator | changed: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.164743 | orchestrator | changed: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.164757 | orchestrator | changed: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) 2025-08-29 17:42:35.164767 | orchestrator | 2025-08-29 17:42:35.164777 | orchestrator | TASK [keystone : include_tasks] ************************************************ 2025-08-29 17:42:35.164787 | orchestrator | Friday 29 August 2025 17:40:39 +0000 (0:00:02.375) 0:00:40.992 ********* 2025-08-29 17:42:35.164797 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.164806 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.164814 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.164821 | orchestrator | 2025-08-29 17:42:35.164829 | orchestrator | TASK [keystone : Creating keystone database] *********************************** 2025-08-29 17:42:35.164837 | orchestrator | Friday 29 August 2025 17:40:39 +0000 (0:00:00.262) 0:00:41.255 ********* 2025-08-29 17:42:35.164845 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.164853 | orchestrator | 2025-08-29 17:42:35.164860 | orchestrator | TASK [keystone : Creating Keystone database user and setting permissions] ****** 2025-08-29 17:42:35.164868 | orchestrator | Friday 29 August 2025 17:40:41 +0000 (0:00:02.318) 0:00:43.573 ********* 2025-08-29 17:42:35.164876 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.164884 | orchestrator | 2025-08-29 17:42:35.164892 | orchestrator | TASK [keystone : Checking for any running keystone_fernet containers] ********** 2025-08-29 17:42:35.164899 | orchestrator | Friday 29 August 2025 17:40:44 +0000 (0:00:02.205) 0:00:45.779 ********* 2025-08-29 17:42:35.164907 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:42:35.164915 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:42:35.164923 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:42:35.164930 | orchestrator | 2025-08-29 17:42:35.164938 | orchestrator | TASK [keystone : Group nodes where keystone_fernet is running] ***************** 2025-08-29 17:42:35.164946 | orchestrator | Friday 29 August 2025 17:40:45 +0000 (0:00:01.372) 0:00:47.152 ********* 2025-08-29 17:42:35.164954 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:42:35.164961 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:42:35.164977 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:42:35.164985 | orchestrator | 2025-08-29 17:42:35.164996 | orchestrator | TASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] *** 2025-08-29 17:42:35.165005 | orchestrator | Friday 29 August 2025 17:40:45 +0000 (0:00:00.275) 0:00:47.427 ********* 2025-08-29 17:42:35.165013 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.165020 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.165029 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.165036 | orchestrator | 2025-08-29 17:42:35.165044 | orchestrator | TASK [keystone : Running Keystone bootstrap container] ************************* 2025-08-29 17:42:35.165052 | orchestrator | Friday 29 August 2025 17:40:46 +0000 (0:00:00.295) 0:00:47.722 ********* 2025-08-29 17:42:35.165060 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.165068 | orchestrator | 2025-08-29 17:42:35.165075 | orchestrator | TASK [keystone : Running Keystone fernet bootstrap container] ****************** 2025-08-29 17:42:35.165083 | orchestrator | Friday 29 August 2025 17:40:59 +0000 (0:00:13.514) 0:01:01.237 ********* 2025-08-29 17:42:35.165091 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.165099 | orchestrator | 2025-08-29 17:42:35.165112 | orchestrator | TASK [keystone : Flush handlers] *********************************************** 2025-08-29 17:42:35.165125 | orchestrator | Friday 29 August 2025 17:41:09 +0000 (0:00:10.389) 0:01:11.627 ********* 2025-08-29 17:42:35.165138 | orchestrator | 2025-08-29 17:42:35.165156 | orchestrator | TASK [keystone : Flush handlers] *********************************************** 2025-08-29 17:42:35.165177 | orchestrator | Friday 29 August 2025 17:41:09 +0000 (0:00:00.059) 0:01:11.686 ********* 2025-08-29 17:42:35.165189 | orchestrator | 2025-08-29 17:42:35.165200 | orchestrator | TASK [keystone : Flush handlers] *********************************************** 2025-08-29 17:42:35.165212 | orchestrator | Friday 29 August 2025 17:41:10 +0000 (0:00:00.169) 0:01:11.855 ********* 2025-08-29 17:42:35.165225 | orchestrator | 2025-08-29 17:42:35.165237 | orchestrator | RUNNING HANDLER [keystone : Restart keystone-ssh container] ******************** 2025-08-29 17:42:35.165249 | orchestrator | Friday 29 August 2025 17:41:10 +0000 (0:00:00.060) 0:01:11.916 ********* 2025-08-29 17:42:35.165261 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.165272 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:42:35.165283 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:42:35.165296 | orchestrator | 2025-08-29 17:42:35.165307 | orchestrator | RUNNING HANDLER [keystone : Restart keystone-fernet container] ***************** 2025-08-29 17:42:35.165319 | orchestrator | Friday 29 August 2025 17:41:26 +0000 (0:00:16.628) 0:01:28.545 ********* 2025-08-29 17:42:35.165330 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:42:35.165342 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:42:35.165354 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.165366 | orchestrator | 2025-08-29 17:42:35.165377 | orchestrator | RUNNING HANDLER [keystone : Restart keystone container] ************************ 2025-08-29 17:42:35.165389 | orchestrator | Friday 29 August 2025 17:41:34 +0000 (0:00:07.427) 0:01:35.973 ********* 2025-08-29 17:42:35.165402 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.165415 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:42:35.165450 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:42:35.165464 | orchestrator | 2025-08-29 17:42:35.165475 | orchestrator | TASK [keystone : include_tasks] ************************************************ 2025-08-29 17:42:35.165488 | orchestrator | Friday 29 August 2025 17:41:45 +0000 (0:00:11.472) 0:01:47.445 ********* 2025-08-29 17:42:35.165500 | orchestrator | included: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:42:35.165513 | orchestrator | 2025-08-29 17:42:35.165525 | orchestrator | TASK [keystone : Waiting for Keystone SSH port to be UP] *********************** 2025-08-29 17:42:35.165540 | orchestrator | Friday 29 August 2025 17:41:46 +0000 (0:00:00.820) 0:01:48.266 ********* 2025-08-29 17:42:35.165553 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:42:35.165566 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:42:35.165579 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:42:35.165590 | orchestrator | 2025-08-29 17:42:35.165607 | orchestrator | TASK [keystone : Run key distribution] ***************************************** 2025-08-29 17:42:35.165615 | orchestrator | Friday 29 August 2025 17:41:47 +0000 (0:00:00.764) 0:01:49.031 ********* 2025-08-29 17:42:35.165623 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:42:35.165631 | orchestrator | 2025-08-29 17:42:35.165639 | orchestrator | TASK [keystone : Creating admin project, user, role, service, and endpoint] **** 2025-08-29 17:42:35.165652 | orchestrator | Friday 29 August 2025 17:41:49 +0000 (0:00:01.816) 0:01:50.848 ********* 2025-08-29 17:42:35.165660 | orchestrator | changed: [testbed-node-0] => (item=RegionOne) 2025-08-29 17:42:35.165669 | orchestrator | 2025-08-29 17:42:35.165677 | orchestrator | TASK [service-ks-register : keystone | Creating services] ********************** 2025-08-29 17:42:35.165685 | orchestrator | Friday 29 August 2025 17:42:00 +0000 (0:00:11.159) 0:02:02.008 ********* 2025-08-29 17:42:35.165693 | orchestrator | changed: [testbed-node-0] => (item=keystone (identity)) 2025-08-29 17:42:35.165701 | orchestrator | 2025-08-29 17:42:35.165709 | orchestrator | TASK [service-ks-register : keystone | Creating endpoints] ********************* 2025-08-29 17:42:35.165716 | orchestrator | Friday 29 August 2025 17:42:22 +0000 (0:00:21.998) 0:02:24.007 ********* 2025-08-29 17:42:35.165724 | orchestrator | ok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal) 2025-08-29 17:42:35.165733 | orchestrator | ok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public) 2025-08-29 17:42:35.165741 | orchestrator | 2025-08-29 17:42:35.165749 | orchestrator | TASK [service-ks-register : keystone | Creating projects] ********************** 2025-08-29 17:42:35.165757 | orchestrator | Friday 29 August 2025 17:42:29 +0000 (0:00:06.904) 0:02:30.911 ********* 2025-08-29 17:42:35.165765 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.165772 | orchestrator | 2025-08-29 17:42:35.165780 | orchestrator | TASK [service-ks-register : keystone | Creating users] ************************* 2025-08-29 17:42:35.165788 | orchestrator | Friday 29 August 2025 17:42:29 +0000 (0:00:00.129) 0:02:31.041 ********* 2025-08-29 17:42:35.165796 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.165803 | orchestrator | 2025-08-29 17:42:35.165811 | orchestrator | TASK [service-ks-register : keystone | Creating roles] ************************* 2025-08-29 17:42:35.165819 | orchestrator | Friday 29 August 2025 17:42:29 +0000 (0:00:00.286) 0:02:31.328 ********* 2025-08-29 17:42:35.165827 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.165835 | orchestrator | 2025-08-29 17:42:35.165851 | orchestrator | TASK [service-ks-register : keystone | Granting user roles] ******************** 2025-08-29 17:42:35.165859 | orchestrator | Friday 29 August 2025 17:42:29 +0000 (0:00:00.127) 0:02:31.455 ********* 2025-08-29 17:42:35.165867 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.165874 | orchestrator | 2025-08-29 17:42:35.165882 | orchestrator | TASK [keystone : Creating default user role] *********************************** 2025-08-29 17:42:35.165890 | orchestrator | Friday 29 August 2025 17:42:30 +0000 (0:00:00.325) 0:02:31.781 ********* 2025-08-29 17:42:35.165898 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:42:35.165906 | orchestrator | 2025-08-29 17:42:35.165914 | orchestrator | TASK [keystone : include_tasks] ************************************************ 2025-08-29 17:42:35.165921 | orchestrator | Friday 29 August 2025 17:42:33 +0000 (0:00:03.402) 0:02:35.183 ********* 2025-08-29 17:42:35.165929 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:42:35.165937 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:42:35.165945 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:42:35.165952 | orchestrator | 2025-08-29 17:42:35.165960 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:42:35.165968 | orchestrator | testbed-node-0 : ok=36  changed=20  unreachable=0 failed=0 skipped=14  rescued=0 ignored=0 2025-08-29 17:42:35.165977 | orchestrator | testbed-node-1 : ok=24  changed=13  unreachable=0 failed=0 skipped=10  rescued=0 ignored=0 2025-08-29 17:42:35.165985 | orchestrator | testbed-node-2 : ok=24  changed=13  unreachable=0 failed=0 skipped=10  rescued=0 ignored=0 2025-08-29 17:42:35.165998 | orchestrator | 2025-08-29 17:42:35.166006 | orchestrator | 2025-08-29 17:42:35.166045 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:42:35.166055 | orchestrator | Friday 29 August 2025 17:42:33 +0000 (0:00:00.422) 0:02:35.606 ********* 2025-08-29 17:42:35.166063 | orchestrator | =============================================================================== 2025-08-29 17:42:35.166071 | orchestrator | service-ks-register : keystone | Creating services --------------------- 22.00s 2025-08-29 17:42:35.166078 | orchestrator | keystone : Restart keystone-ssh container ------------------------------ 16.63s 2025-08-29 17:42:35.166086 | orchestrator | keystone : Running Keystone bootstrap container ------------------------ 13.51s 2025-08-29 17:42:35.166094 | orchestrator | keystone : Restart keystone container ---------------------------------- 11.47s 2025-08-29 17:42:35.166103 | orchestrator | keystone : Creating admin project, user, role, service, and endpoint --- 11.16s 2025-08-29 17:42:35.166110 | orchestrator | keystone : Running Keystone fernet bootstrap container ----------------- 10.39s 2025-08-29 17:42:35.166118 | orchestrator | keystone : Copying files for keystone-fernet ---------------------------- 8.78s 2025-08-29 17:42:35.166126 | orchestrator | keystone : Restart keystone-fernet container ---------------------------- 7.43s 2025-08-29 17:42:35.166134 | orchestrator | service-ks-register : keystone | Creating endpoints --------------------- 6.90s 2025-08-29 17:42:35.166142 | orchestrator | keystone : Copying over keystone.conf ----------------------------------- 4.85s 2025-08-29 17:42:35.166150 | orchestrator | service-cert-copy : keystone | Copying over extra CA certificates ------- 3.42s 2025-08-29 17:42:35.166158 | orchestrator | keystone : Creating default user role ----------------------------------- 3.40s 2025-08-29 17:42:35.166166 | orchestrator | keystone : Copying over config.json files for services ------------------ 3.30s 2025-08-29 17:42:35.166174 | orchestrator | keystone : Copying files for keystone-ssh ------------------------------- 2.74s 2025-08-29 17:42:35.166181 | orchestrator | keystone : Check keystone containers ------------------------------------ 2.38s 2025-08-29 17:42:35.166189 | orchestrator | keystone : Creating keystone database ----------------------------------- 2.32s 2025-08-29 17:42:35.166201 | orchestrator | keystone : Creating Keystone database user and setting permissions ------ 2.21s 2025-08-29 17:42:35.166209 | orchestrator | keystone : Copying over existing policy file ---------------------------- 2.15s 2025-08-29 17:42:35.166217 | orchestrator | keystone : Run key distribution ----------------------------------------- 1.82s 2025-08-29 17:42:35.166225 | orchestrator | keystone : Ensuring config directories exist ---------------------------- 1.61s 2025-08-29 17:42:35.166233 | orchestrator | 2025-08-29 17:42:35 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:35.166241 | orchestrator | 2025-08-29 17:42:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:38.191184 | orchestrator | 2025-08-29 17:42:38 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:38.191376 | orchestrator | 2025-08-29 17:42:38 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:38.254934 | orchestrator | 2025-08-29 17:42:38 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:38.255034 | orchestrator | 2025-08-29 17:42:38 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:38.255046 | orchestrator | 2025-08-29 17:42:38 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:38.255057 | orchestrator | 2025-08-29 17:42:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:41.217542 | orchestrator | 2025-08-29 17:42:41 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:41.217628 | orchestrator | 2025-08-29 17:42:41 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:41.218057 | orchestrator | 2025-08-29 17:42:41 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:41.218555 | orchestrator | 2025-08-29 17:42:41 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:41.219735 | orchestrator | 2025-08-29 17:42:41 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:41.219755 | orchestrator | 2025-08-29 17:42:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:44.256673 | orchestrator | 2025-08-29 17:42:44 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:44.257890 | orchestrator | 2025-08-29 17:42:44 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:44.259305 | orchestrator | 2025-08-29 17:42:44 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:44.260707 | orchestrator | 2025-08-29 17:42:44 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state STARTED 2025-08-29 17:42:44.262002 | orchestrator | 2025-08-29 17:42:44 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:44.262233 | orchestrator | 2025-08-29 17:42:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:47.316377 | orchestrator | 2025-08-29 17:42:47 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:47.317894 | orchestrator | 2025-08-29 17:42:47 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:47.319988 | orchestrator | 2025-08-29 17:42:47 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:47.323168 | orchestrator | 2025-08-29 17:42:47 | INFO  | Task 66ffc838-e54f-49ac-9223-aadd32376fcf is in state SUCCESS 2025-08-29 17:42:47.324719 | orchestrator | 2025-08-29 17:42:47 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:42:47.326108 | orchestrator | 2025-08-29 17:42:47 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:47.326147 | orchestrator | 2025-08-29 17:42:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:50.367883 | orchestrator | 2025-08-29 17:42:50 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:50.369638 | orchestrator | 2025-08-29 17:42:50 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:50.371626 | orchestrator | 2025-08-29 17:42:50 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:50.374081 | orchestrator | 2025-08-29 17:42:50 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:42:50.375823 | orchestrator | 2025-08-29 17:42:50 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:50.375860 | orchestrator | 2025-08-29 17:42:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:53.432633 | orchestrator | 2025-08-29 17:42:53 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:53.432710 | orchestrator | 2025-08-29 17:42:53 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:53.433568 | orchestrator | 2025-08-29 17:42:53 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:53.433658 | orchestrator | 2025-08-29 17:42:53 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:42:53.434882 | orchestrator | 2025-08-29 17:42:53 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:53.434921 | orchestrator | 2025-08-29 17:42:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:56.491149 | orchestrator | 2025-08-29 17:42:56 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:56.494922 | orchestrator | 2025-08-29 17:42:56 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:56.496377 | orchestrator | 2025-08-29 17:42:56 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:56.498071 | orchestrator | 2025-08-29 17:42:56 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:42:56.499767 | orchestrator | 2025-08-29 17:42:56 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:56.499814 | orchestrator | 2025-08-29 17:42:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:42:59.542994 | orchestrator | 2025-08-29 17:42:59 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:42:59.544409 | orchestrator | 2025-08-29 17:42:59 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:42:59.545214 | orchestrator | 2025-08-29 17:42:59 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:42:59.546258 | orchestrator | 2025-08-29 17:42:59 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:42:59.547279 | orchestrator | 2025-08-29 17:42:59 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:42:59.547317 | orchestrator | 2025-08-29 17:42:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:02.599078 | orchestrator | 2025-08-29 17:43:02 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:02.601267 | orchestrator | 2025-08-29 17:43:02 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:02.603571 | orchestrator | 2025-08-29 17:43:02 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:02.604617 | orchestrator | 2025-08-29 17:43:02 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:02.606305 | orchestrator | 2025-08-29 17:43:02 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:43:02.606358 | orchestrator | 2025-08-29 17:43:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:05.644146 | orchestrator | 2025-08-29 17:43:05 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:05.644536 | orchestrator | 2025-08-29 17:43:05 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:05.645617 | orchestrator | 2025-08-29 17:43:05 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:05.646341 | orchestrator | 2025-08-29 17:43:05 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:05.647530 | orchestrator | 2025-08-29 17:43:05 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:43:05.647614 | orchestrator | 2025-08-29 17:43:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:08.684740 | orchestrator | 2025-08-29 17:43:08 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:08.686490 | orchestrator | 2025-08-29 17:43:08 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:08.686775 | orchestrator | 2025-08-29 17:43:08 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:08.688244 | orchestrator | 2025-08-29 17:43:08 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:08.690501 | orchestrator | 2025-08-29 17:43:08 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:43:08.690558 | orchestrator | 2025-08-29 17:43:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:11.746701 | orchestrator | 2025-08-29 17:43:11 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:11.747199 | orchestrator | 2025-08-29 17:43:11 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:11.748367 | orchestrator | 2025-08-29 17:43:11 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:11.749382 | orchestrator | 2025-08-29 17:43:11 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:11.750473 | orchestrator | 2025-08-29 17:43:11 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:43:11.750526 | orchestrator | 2025-08-29 17:43:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:14.793528 | orchestrator | 2025-08-29 17:43:14 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:14.793904 | orchestrator | 2025-08-29 17:43:14 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:14.794611 | orchestrator | 2025-08-29 17:43:14 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:14.795678 | orchestrator | 2025-08-29 17:43:14 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:14.796516 | orchestrator | 2025-08-29 17:43:14 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state STARTED 2025-08-29 17:43:14.796539 | orchestrator | 2025-08-29 17:43:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:17.844349 | orchestrator | 2025-08-29 17:43:17 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:17.844771 | orchestrator | 2025-08-29 17:43:17 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:17.845671 | orchestrator | 2025-08-29 17:43:17 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:17.847171 | orchestrator | 2025-08-29 17:43:17 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:17.847766 | orchestrator | 2025-08-29 17:43:17 | INFO  | Task 450211fc-e7c3-4e5f-83b7-eeaeb96ded7e is in state SUCCESS 2025-08-29 17:43:17.847997 | orchestrator | 2025-08-29 17:43:17.848015 | orchestrator | 2025-08-29 17:43:17.848029 | orchestrator | PLAY [Apply role cephclient] *************************************************** 2025-08-29 17:43:17.848042 | orchestrator | 2025-08-29 17:43:17.848055 | orchestrator | TASK [osism.services.cephclient : Include container tasks] ********************* 2025-08-29 17:43:17.848068 | orchestrator | Friday 29 August 2025 17:41:48 +0000 (0:00:00.229) 0:00:00.229 ********* 2025-08-29 17:43:17.848080 | orchestrator | included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager 2025-08-29 17:43:17.848092 | orchestrator | 2025-08-29 17:43:17.848104 | orchestrator | TASK [osism.services.cephclient : Create required directories] ***************** 2025-08-29 17:43:17.848115 | orchestrator | Friday 29 August 2025 17:41:48 +0000 (0:00:00.231) 0:00:00.461 ********* 2025-08-29 17:43:17.848126 | orchestrator | changed: [testbed-manager] => (item=/opt/cephclient/configuration) 2025-08-29 17:43:17.848137 | orchestrator | changed: [testbed-manager] => (item=/opt/cephclient/data) 2025-08-29 17:43:17.848149 | orchestrator | ok: [testbed-manager] => (item=/opt/cephclient) 2025-08-29 17:43:17.848160 | orchestrator | 2025-08-29 17:43:17.848171 | orchestrator | TASK [osism.services.cephclient : Copy configuration files] ******************** 2025-08-29 17:43:17.848210 | orchestrator | Friday 29 August 2025 17:41:49 +0000 (0:00:01.188) 0:00:01.649 ********* 2025-08-29 17:43:17.848222 | orchestrator | changed: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'}) 2025-08-29 17:43:17.848233 | orchestrator | 2025-08-29 17:43:17.848245 | orchestrator | TASK [osism.services.cephclient : Copy keyring file] *************************** 2025-08-29 17:43:17.848257 | orchestrator | Friday 29 August 2025 17:41:50 +0000 (0:00:01.165) 0:00:02.814 ********* 2025-08-29 17:43:17.848268 | orchestrator | changed: [testbed-manager] 2025-08-29 17:43:17.848279 | orchestrator | 2025-08-29 17:43:17.848289 | orchestrator | TASK [osism.services.cephclient : Copy docker-compose.yml file] **************** 2025-08-29 17:43:17.848300 | orchestrator | Friday 29 August 2025 17:41:51 +0000 (0:00:00.988) 0:00:03.803 ********* 2025-08-29 17:43:17.848311 | orchestrator | changed: [testbed-manager] 2025-08-29 17:43:17.848322 | orchestrator | 2025-08-29 17:43:17.848333 | orchestrator | TASK [osism.services.cephclient : Manage cephclient service] ******************* 2025-08-29 17:43:17.848344 | orchestrator | Friday 29 August 2025 17:41:52 +0000 (0:00:00.926) 0:00:04.729 ********* 2025-08-29 17:43:17.848354 | orchestrator | FAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left). 2025-08-29 17:43:17.848365 | orchestrator | ok: [testbed-manager] 2025-08-29 17:43:17.848376 | orchestrator | 2025-08-29 17:43:17.848387 | orchestrator | TASK [osism.services.cephclient : Copy wrapper scripts] ************************ 2025-08-29 17:43:17.848398 | orchestrator | Friday 29 August 2025 17:42:34 +0000 (0:00:41.564) 0:00:46.294 ********* 2025-08-29 17:43:17.848409 | orchestrator | changed: [testbed-manager] => (item=ceph) 2025-08-29 17:43:17.848420 | orchestrator | changed: [testbed-manager] => (item=ceph-authtool) 2025-08-29 17:43:17.848477 | orchestrator | changed: [testbed-manager] => (item=rados) 2025-08-29 17:43:17.848498 | orchestrator | changed: [testbed-manager] => (item=radosgw-admin) 2025-08-29 17:43:17.848515 | orchestrator | changed: [testbed-manager] => (item=rbd) 2025-08-29 17:43:17.848535 | orchestrator | 2025-08-29 17:43:17.848553 | orchestrator | TASK [osism.services.cephclient : Remove old wrapper scripts] ****************** 2025-08-29 17:43:17.848571 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:05.184) 0:00:51.479 ********* 2025-08-29 17:43:17.848589 | orchestrator | ok: [testbed-manager] => (item=crushtool) 2025-08-29 17:43:17.848600 | orchestrator | 2025-08-29 17:43:17.848611 | orchestrator | TASK [osism.services.cephclient : Include package tasks] *********************** 2025-08-29 17:43:17.848622 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.431) 0:00:51.911 ********* 2025-08-29 17:43:17.848633 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:43:17.848643 | orchestrator | 2025-08-29 17:43:17.848654 | orchestrator | TASK [osism.services.cephclient : Include rook task] *************************** 2025-08-29 17:43:17.848665 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.119) 0:00:52.031 ********* 2025-08-29 17:43:17.848676 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:43:17.848686 | orchestrator | 2025-08-29 17:43:17.848697 | orchestrator | RUNNING HANDLER [osism.services.cephclient : Restart cephclient service] ******* 2025-08-29 17:43:17.848708 | orchestrator | Friday 29 August 2025 17:42:40 +0000 (0:00:00.233) 0:00:52.265 ********* 2025-08-29 17:43:17.848718 | orchestrator | changed: [testbed-manager] 2025-08-29 17:43:17.848729 | orchestrator | 2025-08-29 17:43:17.848740 | orchestrator | RUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] *** 2025-08-29 17:43:17.848751 | orchestrator | Friday 29 August 2025 17:42:41 +0000 (0:00:01.465) 0:00:53.730 ********* 2025-08-29 17:43:17.848761 | orchestrator | changed: [testbed-manager] 2025-08-29 17:43:17.848772 | orchestrator | 2025-08-29 17:43:17.848783 | orchestrator | RUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ****** 2025-08-29 17:43:17.848794 | orchestrator | Friday 29 August 2025 17:42:42 +0000 (0:00:00.636) 0:00:54.367 ********* 2025-08-29 17:43:17.848804 | orchestrator | changed: [testbed-manager] 2025-08-29 17:43:17.848815 | orchestrator | 2025-08-29 17:43:17.848825 | orchestrator | RUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] ***** 2025-08-29 17:43:17.848847 | orchestrator | Friday 29 August 2025 17:42:42 +0000 (0:00:00.567) 0:00:54.934 ********* 2025-08-29 17:43:17.848858 | orchestrator | ok: [testbed-manager] => (item=ceph) 2025-08-29 17:43:17.848869 | orchestrator | ok: [testbed-manager] => (item=rados) 2025-08-29 17:43:17.848880 | orchestrator | ok: [testbed-manager] => (item=radosgw-admin) 2025-08-29 17:43:17.848890 | orchestrator | ok: [testbed-manager] => (item=rbd) 2025-08-29 17:43:17.848901 | orchestrator | 2025-08-29 17:43:17.848912 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:43:17.848923 | orchestrator | testbed-manager : ok=12  changed=8  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 17:43:17.848935 | orchestrator | 2025-08-29 17:43:17.849045 | orchestrator | 2025-08-29 17:43:17.849075 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:43:17.849086 | orchestrator | Friday 29 August 2025 17:42:44 +0000 (0:00:01.331) 0:00:56.266 ********* 2025-08-29 17:43:17.849097 | orchestrator | =============================================================================== 2025-08-29 17:43:17.849108 | orchestrator | osism.services.cephclient : Manage cephclient service ------------------ 41.56s 2025-08-29 17:43:17.849119 | orchestrator | osism.services.cephclient : Copy wrapper scripts ------------------------ 5.18s 2025-08-29 17:43:17.849129 | orchestrator | osism.services.cephclient : Restart cephclient service ------------------ 1.47s 2025-08-29 17:43:17.849140 | orchestrator | osism.services.cephclient : Copy bash completion scripts ---------------- 1.33s 2025-08-29 17:43:17.849151 | orchestrator | osism.services.cephclient : Create required directories ----------------- 1.19s 2025-08-29 17:43:17.849161 | orchestrator | osism.services.cephclient : Copy configuration files -------------------- 1.17s 2025-08-29 17:43:17.849172 | orchestrator | osism.services.cephclient : Copy keyring file --------------------------- 0.99s 2025-08-29 17:43:17.849183 | orchestrator | osism.services.cephclient : Copy docker-compose.yml file ---------------- 0.93s 2025-08-29 17:43:17.849193 | orchestrator | osism.services.cephclient : Ensure that all containers are up ----------- 0.64s 2025-08-29 17:43:17.849204 | orchestrator | osism.services.cephclient : Wait for an healthy service ----------------- 0.57s 2025-08-29 17:43:17.849215 | orchestrator | osism.services.cephclient : Remove old wrapper scripts ------------------ 0.43s 2025-08-29 17:43:17.849225 | orchestrator | osism.services.cephclient : Include rook task --------------------------- 0.23s 2025-08-29 17:43:17.849236 | orchestrator | osism.services.cephclient : Include container tasks --------------------- 0.23s 2025-08-29 17:43:17.849246 | orchestrator | osism.services.cephclient : Include package tasks ----------------------- 0.12s 2025-08-29 17:43:17.849258 | orchestrator | 2025-08-29 17:43:17.849277 | orchestrator | 2025-08-29 17:43:17.849295 | orchestrator | PLAY [Download ironic ipa images] ********************************************** 2025-08-29 17:43:17.849312 | orchestrator | 2025-08-29 17:43:17.849330 | orchestrator | TASK [Ensure the destination directory exists] ********************************* 2025-08-29 17:43:17.849347 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.120) 0:00:00.120 ********* 2025-08-29 17:43:17.849365 | orchestrator | changed: [localhost] 2025-08-29 17:43:17.849382 | orchestrator | 2025-08-29 17:43:17.849400 | orchestrator | TASK [Download ironic-agent initramfs] ***************************************** 2025-08-29 17:43:17.849420 | orchestrator | Friday 29 August 2025 17:42:40 +0000 (0:00:01.043) 0:00:01.164 ********* 2025-08-29 17:43:17.849523 | orchestrator | changed: [localhost] 2025-08-29 17:43:17.849546 | orchestrator | 2025-08-29 17:43:17.849566 | orchestrator | TASK [Download ironic-agent kernel] ******************************************** 2025-08-29 17:43:17.849585 | orchestrator | Friday 29 August 2025 17:43:10 +0000 (0:00:29.985) 0:00:31.150 ********* 2025-08-29 17:43:17.849598 | orchestrator | changed: [localhost] 2025-08-29 17:43:17.849608 | orchestrator | 2025-08-29 17:43:17.849628 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:43:17.849639 | orchestrator | 2025-08-29 17:43:17.849651 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:43:17.849673 | orchestrator | Friday 29 August 2025 17:43:14 +0000 (0:00:04.039) 0:00:35.189 ********* 2025-08-29 17:43:17.849684 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:43:17.849695 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:43:17.849706 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:43:17.849717 | orchestrator | 2025-08-29 17:43:17.849728 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:43:17.849739 | orchestrator | Friday 29 August 2025 17:43:14 +0000 (0:00:00.292) 0:00:35.482 ********* 2025-08-29 17:43:17.849750 | orchestrator | [WARNING]: Could not match supplied host pattern, ignoring: enable_ironic_True 2025-08-29 17:43:17.849761 | orchestrator | ok: [testbed-node-0] => (item=enable_ironic_False) 2025-08-29 17:43:17.849772 | orchestrator | ok: [testbed-node-1] => (item=enable_ironic_False) 2025-08-29 17:43:17.849783 | orchestrator | ok: [testbed-node-2] => (item=enable_ironic_False) 2025-08-29 17:43:17.849794 | orchestrator | 2025-08-29 17:43:17.849805 | orchestrator | PLAY [Apply role ironic] ******************************************************* 2025-08-29 17:43:17.849816 | orchestrator | skipping: no hosts matched 2025-08-29 17:43:17.849827 | orchestrator | 2025-08-29 17:43:17.849837 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:43:17.849848 | orchestrator | localhost : ok=3  changed=3  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:43:17.849860 | orchestrator | testbed-node-0 : ok=2  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:43:17.849872 | orchestrator | testbed-node-1 : ok=2  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:43:17.849883 | orchestrator | testbed-node-2 : ok=2  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:43:17.849894 | orchestrator | 2025-08-29 17:43:17.849905 | orchestrator | 2025-08-29 17:43:17.849914 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:43:17.849924 | orchestrator | Friday 29 August 2025 17:43:14 +0000 (0:00:00.464) 0:00:35.946 ********* 2025-08-29 17:43:17.849934 | orchestrator | =============================================================================== 2025-08-29 17:43:17.849943 | orchestrator | Download ironic-agent initramfs ---------------------------------------- 29.99s 2025-08-29 17:43:17.849953 | orchestrator | Download ironic-agent kernel -------------------------------------------- 4.04s 2025-08-29 17:43:17.849963 | orchestrator | Ensure the destination directory exists --------------------------------- 1.04s 2025-08-29 17:43:17.849973 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.46s 2025-08-29 17:43:17.849991 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.29s 2025-08-29 17:43:17.850001 | orchestrator | 2025-08-29 17:43:17 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:17.850011 | orchestrator | 2025-08-29 17:43:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:20.889615 | orchestrator | 2025-08-29 17:43:20 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:20.891040 | orchestrator | 2025-08-29 17:43:20 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:20.892676 | orchestrator | 2025-08-29 17:43:20 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:20.893578 | orchestrator | 2025-08-29 17:43:20 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:20.894352 | orchestrator | 2025-08-29 17:43:20 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:20.894619 | orchestrator | 2025-08-29 17:43:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:23.916762 | orchestrator | 2025-08-29 17:43:23 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:23.916874 | orchestrator | 2025-08-29 17:43:23 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:23.917475 | orchestrator | 2025-08-29 17:43:23 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:23.918715 | orchestrator | 2025-08-29 17:43:23 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:23.919225 | orchestrator | 2025-08-29 17:43:23 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:23.919246 | orchestrator | 2025-08-29 17:43:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:26.962741 | orchestrator | 2025-08-29 17:43:26 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:26.964829 | orchestrator | 2025-08-29 17:43:26 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:26.968069 | orchestrator | 2025-08-29 17:43:26 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:26.970896 | orchestrator | 2025-08-29 17:43:26 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:26.973148 | orchestrator | 2025-08-29 17:43:26 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:26.973597 | orchestrator | 2025-08-29 17:43:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:30.010620 | orchestrator | 2025-08-29 17:43:30 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:30.011002 | orchestrator | 2025-08-29 17:43:30 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:30.011627 | orchestrator | 2025-08-29 17:43:30 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:30.012369 | orchestrator | 2025-08-29 17:43:30 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:30.013238 | orchestrator | 2025-08-29 17:43:30 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:30.013262 | orchestrator | 2025-08-29 17:43:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:33.042269 | orchestrator | 2025-08-29 17:43:33 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:33.042508 | orchestrator | 2025-08-29 17:43:33 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:33.043207 | orchestrator | 2025-08-29 17:43:33 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:33.044862 | orchestrator | 2025-08-29 17:43:33 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:33.045604 | orchestrator | 2025-08-29 17:43:33 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:33.045751 | orchestrator | 2025-08-29 17:43:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:36.074243 | orchestrator | 2025-08-29 17:43:36 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:36.075754 | orchestrator | 2025-08-29 17:43:36 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:36.077511 | orchestrator | 2025-08-29 17:43:36 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:36.078247 | orchestrator | 2025-08-29 17:43:36 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:36.079394 | orchestrator | 2025-08-29 17:43:36 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:36.079493 | orchestrator | 2025-08-29 17:43:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:39.103595 | orchestrator | 2025-08-29 17:43:39 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:39.103698 | orchestrator | 2025-08-29 17:43:39 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:39.104129 | orchestrator | 2025-08-29 17:43:39 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:39.104587 | orchestrator | 2025-08-29 17:43:39 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:39.105409 | orchestrator | 2025-08-29 17:43:39 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:39.105516 | orchestrator | 2025-08-29 17:43:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:42.131351 | orchestrator | 2025-08-29 17:43:42 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:42.131576 | orchestrator | 2025-08-29 17:43:42 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:42.132125 | orchestrator | 2025-08-29 17:43:42 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:42.132797 | orchestrator | 2025-08-29 17:43:42 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:42.133536 | orchestrator | 2025-08-29 17:43:42 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:42.133557 | orchestrator | 2025-08-29 17:43:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:45.163906 | orchestrator | 2025-08-29 17:43:45 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:45.166663 | orchestrator | 2025-08-29 17:43:45 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:45.168112 | orchestrator | 2025-08-29 17:43:45 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:45.171116 | orchestrator | 2025-08-29 17:43:45 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:45.175130 | orchestrator | 2025-08-29 17:43:45 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:45.175725 | orchestrator | 2025-08-29 17:43:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:48.211145 | orchestrator | 2025-08-29 17:43:48 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:48.211677 | orchestrator | 2025-08-29 17:43:48 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:48.212400 | orchestrator | 2025-08-29 17:43:48 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:48.213240 | orchestrator | 2025-08-29 17:43:48 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:48.213984 | orchestrator | 2025-08-29 17:43:48 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:48.214083 | orchestrator | 2025-08-29 17:43:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:51.240576 | orchestrator | 2025-08-29 17:43:51 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:51.242606 | orchestrator | 2025-08-29 17:43:51 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:51.243297 | orchestrator | 2025-08-29 17:43:51 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:51.244093 | orchestrator | 2025-08-29 17:43:51 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:51.244742 | orchestrator | 2025-08-29 17:43:51 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:51.244866 | orchestrator | 2025-08-29 17:43:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:54.281504 | orchestrator | 2025-08-29 17:43:54 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:54.281936 | orchestrator | 2025-08-29 17:43:54 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:54.282612 | orchestrator | 2025-08-29 17:43:54 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:54.283262 | orchestrator | 2025-08-29 17:43:54 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:54.283828 | orchestrator | 2025-08-29 17:43:54 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:54.283854 | orchestrator | 2025-08-29 17:43:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:43:57.326961 | orchestrator | 2025-08-29 17:43:57 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:43:57.327222 | orchestrator | 2025-08-29 17:43:57 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:43:57.327858 | orchestrator | 2025-08-29 17:43:57 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:43:57.328355 | orchestrator | 2025-08-29 17:43:57 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:43:57.329083 | orchestrator | 2025-08-29 17:43:57 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:43:57.329108 | orchestrator | 2025-08-29 17:43:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:00.351424 | orchestrator | 2025-08-29 17:44:00 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:00.351531 | orchestrator | 2025-08-29 17:44:00 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:00.352005 | orchestrator | 2025-08-29 17:44:00 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:00.352428 | orchestrator | 2025-08-29 17:44:00 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:44:00.353052 | orchestrator | 2025-08-29 17:44:00 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:00.353107 | orchestrator | 2025-08-29 17:44:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:03.387024 | orchestrator | 2025-08-29 17:44:03 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:03.387372 | orchestrator | 2025-08-29 17:44:03 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:03.388162 | orchestrator | 2025-08-29 17:44:03 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:03.388629 | orchestrator | 2025-08-29 17:44:03 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:44:03.389250 | orchestrator | 2025-08-29 17:44:03 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:03.389291 | orchestrator | 2025-08-29 17:44:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:06.413629 | orchestrator | 2025-08-29 17:44:06 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:06.416110 | orchestrator | 2025-08-29 17:44:06 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:06.418056 | orchestrator | 2025-08-29 17:44:06 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:06.419808 | orchestrator | 2025-08-29 17:44:06 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:44:06.421716 | orchestrator | 2025-08-29 17:44:06 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:06.421745 | orchestrator | 2025-08-29 17:44:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:09.451302 | orchestrator | 2025-08-29 17:44:09 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:09.451404 | orchestrator | 2025-08-29 17:44:09 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:09.452025 | orchestrator | 2025-08-29 17:44:09 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:09.452617 | orchestrator | 2025-08-29 17:44:09 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:44:09.453283 | orchestrator | 2025-08-29 17:44:09 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:09.453313 | orchestrator | 2025-08-29 17:44:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:12.486905 | orchestrator | 2025-08-29 17:44:12 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:12.487287 | orchestrator | 2025-08-29 17:44:12 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:12.488071 | orchestrator | 2025-08-29 17:44:12 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:12.488923 | orchestrator | 2025-08-29 17:44:12 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:44:12.489750 | orchestrator | 2025-08-29 17:44:12 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:12.489797 | orchestrator | 2025-08-29 17:44:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:15.519021 | orchestrator | 2025-08-29 17:44:15 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:15.519636 | orchestrator | 2025-08-29 17:44:15 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:15.520617 | orchestrator | 2025-08-29 17:44:15 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:15.521655 | orchestrator | 2025-08-29 17:44:15 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state STARTED 2025-08-29 17:44:15.522726 | orchestrator | 2025-08-29 17:44:15 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:15.522750 | orchestrator | 2025-08-29 17:44:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:18.552581 | orchestrator | 2025-08-29 17:44:18 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:18.552845 | orchestrator | 2025-08-29 17:44:18 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:18.553506 | orchestrator | 2025-08-29 17:44:18 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:18.554114 | orchestrator | 2025-08-29 17:44:18 | INFO  | Task 4d843cda-a2d0-4e49-8576-d9858e9994cc is in state SUCCESS 2025-08-29 17:44:18.554910 | orchestrator | 2025-08-29 17:44:18 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:18.554943 | orchestrator | 2025-08-29 17:44:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:21.601652 | orchestrator | 2025-08-29 17:44:21 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:21.603979 | orchestrator | 2025-08-29 17:44:21 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:21.604138 | orchestrator | 2025-08-29 17:44:21 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:21.604166 | orchestrator | 2025-08-29 17:44:21 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:21.604187 | orchestrator | 2025-08-29 17:44:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:24.639171 | orchestrator | 2025-08-29 17:44:24 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:24.639543 | orchestrator | 2025-08-29 17:44:24 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:24.640918 | orchestrator | 2025-08-29 17:44:24 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:24.643552 | orchestrator | 2025-08-29 17:44:24 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:24.643609 | orchestrator | 2025-08-29 17:44:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:27.689143 | orchestrator | 2025-08-29 17:44:27 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:27.689632 | orchestrator | 2025-08-29 17:44:27 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:27.690437 | orchestrator | 2025-08-29 17:44:27 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:27.691439 | orchestrator | 2025-08-29 17:44:27 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:27.691500 | orchestrator | 2025-08-29 17:44:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:30.727059 | orchestrator | 2025-08-29 17:44:30 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:30.727572 | orchestrator | 2025-08-29 17:44:30 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:30.728219 | orchestrator | 2025-08-29 17:44:30 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:30.729044 | orchestrator | 2025-08-29 17:44:30 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:30.729068 | orchestrator | 2025-08-29 17:44:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:33.758138 | orchestrator | 2025-08-29 17:44:33 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:33.758439 | orchestrator | 2025-08-29 17:44:33 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:33.759112 | orchestrator | 2025-08-29 17:44:33 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:33.759817 | orchestrator | 2025-08-29 17:44:33 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:33.759843 | orchestrator | 2025-08-29 17:44:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:36.792080 | orchestrator | 2025-08-29 17:44:36 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:36.792526 | orchestrator | 2025-08-29 17:44:36 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:36.793272 | orchestrator | 2025-08-29 17:44:36 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:36.793839 | orchestrator | 2025-08-29 17:44:36 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:36.793873 | orchestrator | 2025-08-29 17:44:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:39.841988 | orchestrator | 2025-08-29 17:44:39 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:39.842237 | orchestrator | 2025-08-29 17:44:39 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:39.842876 | orchestrator | 2025-08-29 17:44:39 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:39.843819 | orchestrator | 2025-08-29 17:44:39 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:39.843945 | orchestrator | 2025-08-29 17:44:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:42.877367 | orchestrator | 2025-08-29 17:44:42 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:42.879617 | orchestrator | 2025-08-29 17:44:42 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:42.880291 | orchestrator | 2025-08-29 17:44:42 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:42.881961 | orchestrator | 2025-08-29 17:44:42 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state STARTED 2025-08-29 17:44:42.881993 | orchestrator | 2025-08-29 17:44:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:45.902547 | orchestrator | 2025-08-29 17:44:45 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:45.902965 | orchestrator | 2025-08-29 17:44:45 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:45.904900 | orchestrator | 2025-08-29 17:44:45 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:45.906615 | orchestrator | 2025-08-29 17:44:45 | INFO  | Task 2fd641ba-8b8a-40aa-b2ec-5d8abf47a423 is in state SUCCESS 2025-08-29 17:44:45.908249 | orchestrator | 2025-08-29 17:44:45.908290 | orchestrator | 2025-08-29 17:44:45.908303 | orchestrator | PLAY [Bootstraph ceph dashboard] *********************************************** 2025-08-29 17:44:45.908315 | orchestrator | 2025-08-29 17:44:45.908326 | orchestrator | TASK [Disable the ceph dashboard] ********************************************** 2025-08-29 17:44:45.908337 | orchestrator | Friday 29 August 2025 17:42:48 +0000 (0:00:00.312) 0:00:00.313 ********* 2025-08-29 17:44:45.908348 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908360 | orchestrator | 2025-08-29 17:44:45.908374 | orchestrator | TASK [Set mgr/dashboard/ssl to false] ****************************************** 2025-08-29 17:44:45.908393 | orchestrator | Friday 29 August 2025 17:42:50 +0000 (0:00:01.972) 0:00:02.285 ********* 2025-08-29 17:44:45.908412 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908430 | orchestrator | 2025-08-29 17:44:45.908473 | orchestrator | TASK [Set mgr/dashboard/server_port to 7000] *********************************** 2025-08-29 17:44:45.908496 | orchestrator | Friday 29 August 2025 17:42:51 +0000 (0:00:01.054) 0:00:03.340 ********* 2025-08-29 17:44:45.908515 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908534 | orchestrator | 2025-08-29 17:44:45.908554 | orchestrator | TASK [Set mgr/dashboard/server_addr to 0.0.0.0] ******************************** 2025-08-29 17:44:45.908573 | orchestrator | Friday 29 August 2025 17:42:53 +0000 (0:00:01.130) 0:00:04.471 ********* 2025-08-29 17:44:45.908592 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908610 | orchestrator | 2025-08-29 17:44:45.908630 | orchestrator | TASK [Set mgr/dashboard/standby_behaviour to error] **************************** 2025-08-29 17:44:45.908649 | orchestrator | Friday 29 August 2025 17:42:54 +0000 (0:00:01.237) 0:00:05.708 ********* 2025-08-29 17:44:45.908668 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908715 | orchestrator | 2025-08-29 17:44:45.908734 | orchestrator | TASK [Set mgr/dashboard/standby_error_status_code to 404] ********************** 2025-08-29 17:44:45.908752 | orchestrator | Friday 29 August 2025 17:42:55 +0000 (0:00:01.166) 0:00:06.875 ********* 2025-08-29 17:44:45.908770 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908788 | orchestrator | 2025-08-29 17:44:45.908807 | orchestrator | TASK [Enable the ceph dashboard] *********************************************** 2025-08-29 17:44:45.908826 | orchestrator | Friday 29 August 2025 17:42:56 +0000 (0:00:01.013) 0:00:07.888 ********* 2025-08-29 17:44:45.908845 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908865 | orchestrator | 2025-08-29 17:44:45.908884 | orchestrator | TASK [Write ceph_dashboard_password to temporary file] ************************* 2025-08-29 17:44:45.908903 | orchestrator | Friday 29 August 2025 17:42:58 +0000 (0:00:02.094) 0:00:09.983 ********* 2025-08-29 17:44:45.908922 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.908942 | orchestrator | 2025-08-29 17:44:45.908961 | orchestrator | TASK [Create admin user] ******************************************************* 2025-08-29 17:44:45.908980 | orchestrator | Friday 29 August 2025 17:42:59 +0000 (0:00:01.138) 0:00:11.122 ********* 2025-08-29 17:44:45.908999 | orchestrator | changed: [testbed-manager] 2025-08-29 17:44:45.909019 | orchestrator | 2025-08-29 17:44:45.909038 | orchestrator | TASK [Remove temporary file for ceph_dashboard_password] *********************** 2025-08-29 17:44:45.909057 | orchestrator | Friday 29 August 2025 17:43:52 +0000 (0:00:52.609) 0:01:03.732 ********* 2025-08-29 17:44:45.909076 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:44:45.909096 | orchestrator | 2025-08-29 17:44:45.909115 | orchestrator | PLAY [Restart ceph manager services] ******************************************* 2025-08-29 17:44:45.909133 | orchestrator | 2025-08-29 17:44:45.909152 | orchestrator | TASK [Restart ceph manager service] ******************************************** 2025-08-29 17:44:45.909173 | orchestrator | Friday 29 August 2025 17:43:52 +0000 (0:00:00.127) 0:01:03.859 ********* 2025-08-29 17:44:45.909192 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:44:45.909211 | orchestrator | 2025-08-29 17:44:45.909229 | orchestrator | PLAY [Restart ceph manager services] ******************************************* 2025-08-29 17:44:45.909248 | orchestrator | 2025-08-29 17:44:45.909265 | orchestrator | TASK [Restart ceph manager service] ******************************************** 2025-08-29 17:44:45.909283 | orchestrator | Friday 29 August 2025 17:44:04 +0000 (0:00:11.589) 0:01:15.449 ********* 2025-08-29 17:44:45.909302 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:44:45.909321 | orchestrator | 2025-08-29 17:44:45.909339 | orchestrator | PLAY [Restart ceph manager services] ******************************************* 2025-08-29 17:44:45.909358 | orchestrator | 2025-08-29 17:44:45.909376 | orchestrator | TASK [Restart ceph manager service] ******************************************** 2025-08-29 17:44:45.909395 | orchestrator | Friday 29 August 2025 17:44:05 +0000 (0:00:01.252) 0:01:16.702 ********* 2025-08-29 17:44:45.909413 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:44:45.909431 | orchestrator | 2025-08-29 17:44:45.909449 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:44:45.909520 | orchestrator | testbed-manager : ok=9  changed=9  unreachable=0 failed=0 skipped=1  rescued=0 ignored=0 2025-08-29 17:44:45.909582 | orchestrator | testbed-node-0 : ok=1  changed=1  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:44:45.909602 | orchestrator | testbed-node-1 : ok=1  changed=1  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:44:45.909620 | orchestrator | testbed-node-2 : ok=1  changed=1  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:44:45.909637 | orchestrator | 2025-08-29 17:44:45.909656 | orchestrator | 2025-08-29 17:44:45.909674 | orchestrator | 2025-08-29 17:44:45.909693 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:44:45.909710 | orchestrator | Friday 29 August 2025 17:44:16 +0000 (0:00:11.161) 0:01:27.863 ********* 2025-08-29 17:44:45.909758 | orchestrator | =============================================================================== 2025-08-29 17:44:45.909776 | orchestrator | Create admin user ------------------------------------------------------ 52.61s 2025-08-29 17:44:45.909794 | orchestrator | Restart ceph manager service ------------------------------------------- 24.00s 2025-08-29 17:44:45.909830 | orchestrator | Enable the ceph dashboard ----------------------------------------------- 2.09s 2025-08-29 17:44:45.909848 | orchestrator | Disable the ceph dashboard ---------------------------------------------- 1.97s 2025-08-29 17:44:45.909867 | orchestrator | Set mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.24s 2025-08-29 17:44:45.909885 | orchestrator | Set mgr/dashboard/standby_behaviour to error ---------------------------- 1.17s 2025-08-29 17:44:45.909901 | orchestrator | Write ceph_dashboard_password to temporary file ------------------------- 1.14s 2025-08-29 17:44:45.909919 | orchestrator | Set mgr/dashboard/server_port to 7000 ----------------------------------- 1.13s 2025-08-29 17:44:45.909938 | orchestrator | Set mgr/dashboard/ssl to false ------------------------------------------ 1.05s 2025-08-29 17:44:45.909957 | orchestrator | Set mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.01s 2025-08-29 17:44:45.909975 | orchestrator | Remove temporary file for ceph_dashboard_password ----------------------- 0.13s 2025-08-29 17:44:45.909993 | orchestrator | 2025-08-29 17:44:45.910011 | orchestrator | 2025-08-29 17:44:45.910084 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:44:45.910103 | orchestrator | 2025-08-29 17:44:45.910121 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:44:45.910139 | orchestrator | Friday 29 August 2025 17:43:20 +0000 (0:00:00.375) 0:00:00.375 ********* 2025-08-29 17:44:45.910157 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:44:45.910175 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:44:45.910192 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:44:45.910211 | orchestrator | 2025-08-29 17:44:45.910231 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:44:45.910248 | orchestrator | Friday 29 August 2025 17:43:21 +0000 (0:00:00.792) 0:00:01.168 ********* 2025-08-29 17:44:45.910266 | orchestrator | ok: [testbed-node-0] => (item=enable_placement_True) 2025-08-29 17:44:45.910284 | orchestrator | ok: [testbed-node-1] => (item=enable_placement_True) 2025-08-29 17:44:45.910302 | orchestrator | ok: [testbed-node-2] => (item=enable_placement_True) 2025-08-29 17:44:45.910319 | orchestrator | 2025-08-29 17:44:45.910338 | orchestrator | PLAY [Apply role placement] **************************************************** 2025-08-29 17:44:45.910356 | orchestrator | 2025-08-29 17:44:45.910374 | orchestrator | TASK [placement : include_tasks] *********************************************** 2025-08-29 17:44:45.910392 | orchestrator | Friday 29 August 2025 17:43:22 +0000 (0:00:00.783) 0:00:01.951 ********* 2025-08-29 17:44:45.910410 | orchestrator | included: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:44:45.910429 | orchestrator | 2025-08-29 17:44:45.910446 | orchestrator | TASK [service-ks-register : placement | Creating services] ********************* 2025-08-29 17:44:45.910526 | orchestrator | Friday 29 August 2025 17:43:22 +0000 (0:00:00.412) 0:00:02.364 ********* 2025-08-29 17:44:45.910545 | orchestrator | changed: [testbed-node-0] => (item=placement (placement)) 2025-08-29 17:44:45.910563 | orchestrator | 2025-08-29 17:44:45.910581 | orchestrator | TASK [service-ks-register : placement | Creating endpoints] ******************** 2025-08-29 17:44:45.910598 | orchestrator | Friday 29 August 2025 17:43:25 +0000 (0:00:03.145) 0:00:05.509 ********* 2025-08-29 17:44:45.910616 | orchestrator | changed: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal) 2025-08-29 17:44:45.910634 | orchestrator | changed: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public) 2025-08-29 17:44:45.910652 | orchestrator | 2025-08-29 17:44:45.910670 | orchestrator | TASK [service-ks-register : placement | Creating projects] ********************* 2025-08-29 17:44:45.910687 | orchestrator | Friday 29 August 2025 17:43:31 +0000 (0:00:05.859) 0:00:11.369 ********* 2025-08-29 17:44:45.910720 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:44:45.910737 | orchestrator | 2025-08-29 17:44:45.910753 | orchestrator | TASK [service-ks-register : placement | Creating users] ************************ 2025-08-29 17:44:45.910769 | orchestrator | Friday 29 August 2025 17:43:34 +0000 (0:00:03.170) 0:00:14.540 ********* 2025-08-29 17:44:45.910784 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:44:45.910800 | orchestrator | changed: [testbed-node-0] => (item=placement -> service) 2025-08-29 17:44:45.910816 | orchestrator | 2025-08-29 17:44:45.910832 | orchestrator | TASK [service-ks-register : placement | Creating roles] ************************ 2025-08-29 17:44:45.910847 | orchestrator | Friday 29 August 2025 17:43:39 +0000 (0:00:04.302) 0:00:18.843 ********* 2025-08-29 17:44:45.910863 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:44:45.910879 | orchestrator | 2025-08-29 17:44:45.910895 | orchestrator | TASK [service-ks-register : placement | Granting user roles] ******************* 2025-08-29 17:44:45.910910 | orchestrator | Friday 29 August 2025 17:43:42 +0000 (0:00:03.270) 0:00:22.114 ********* 2025-08-29 17:44:45.910934 | orchestrator | changed: [testbed-node-0] => (item=placement -> service -> admin) 2025-08-29 17:44:45.910950 | orchestrator | 2025-08-29 17:44:45.910966 | orchestrator | TASK [placement : include_tasks] *********************************************** 2025-08-29 17:44:45.910982 | orchestrator | Friday 29 August 2025 17:43:47 +0000 (0:00:04.548) 0:00:26.663 ********* 2025-08-29 17:44:45.910998 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:44:45.911014 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:44:45.911030 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:44:45.911045 | orchestrator | 2025-08-29 17:44:45.911061 | orchestrator | TASK [placement : Ensuring config directories exist] *************************** 2025-08-29 17:44:45.911077 | orchestrator | Friday 29 August 2025 17:43:47 +0000 (0:00:00.382) 0:00:27.045 ********* 2025-08-29 17:44:45.911110 | orchestrator | changed: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911132 | orchestrator | changed: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911149 | orchestrator | changed: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911174 | orchestrator | 2025-08-29 17:44:45.911191 | orchestrator | TASK [placement : Check if policies shall be overwritten] ********************** 2025-08-29 17:44:45.911207 | orchestrator | Friday 29 August 2025 17:43:48 +0000 (0:00:01.373) 0:00:28.419 ********* 2025-08-29 17:44:45.911222 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:44:45.911238 | orchestrator | 2025-08-29 17:44:45.911254 | orchestrator | TASK [placement : Set placement policy file] *********************************** 2025-08-29 17:44:45.911269 | orchestrator | Friday 29 August 2025 17:43:49 +0000 (0:00:00.198) 0:00:28.618 ********* 2025-08-29 17:44:45.911285 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:44:45.911300 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:44:45.911316 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:44:45.911332 | orchestrator | 2025-08-29 17:44:45.911349 | orchestrator | TASK [placement : include_tasks] *********************************************** 2025-08-29 17:44:45.911364 | orchestrator | Friday 29 August 2025 17:43:49 +0000 (0:00:00.523) 0:00:29.141 ********* 2025-08-29 17:44:45.911380 | orchestrator | included: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:44:45.911396 | orchestrator | 2025-08-29 17:44:45.911418 | orchestrator | TASK [service-cert-copy : placement | Copying over extra CA certificates] ****** 2025-08-29 17:44:45.911434 | orchestrator | Friday 29 August 2025 17:43:50 +0000 (0:00:00.446) 0:00:29.588 ********* 2025-08-29 17:44:45.911483 | orchestrator | changed: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911502 | orchestrator | changed: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911519 | orchestrator | changed: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911567 | orchestrator | 2025-08-29 17:44:45.911584 | orchestrator | TASK [service-cert-copy : placement | Copying over backend internal TLS certificate] *** 2025-08-29 17:44:45.911599 | orchestrator | Friday 29 August 2025 17:43:52 +0000 (0:00:02.611) 0:00:32.200 ********* 2025-08-29 17:44:45.911616 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.911632 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:44:45.911655 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.911672 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:44:45.911696 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.911721 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:44:45.911737 | orchestrator | 2025-08-29 17:44:45.911753 | orchestrator | TASK [service-cert-copy : placement | Copying over backend internal TLS key] *** 2025-08-29 17:44:45.911768 | orchestrator | Friday 29 August 2025 17:43:54 +0000 (0:00:02.146) 0:00:34.346 ********* 2025-08-29 17:44:45.911785 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.911801 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:44:45.911817 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.911833 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:44:45.911854 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.911872 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:44:45.911889 | orchestrator | 2025-08-29 17:44:45.911906 | orchestrator | TASK [placement : Copying over config.json files for services] ***************** 2025-08-29 17:44:45.911930 | orchestrator | Friday 29 August 2025 17:43:56 +0000 (0:00:01.809) 0:00:36.156 ********* 2025-08-29 17:44:45.911948 | orchestrator | changed: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911975 | orchestrator | changed: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.911993 | orchestrator | changed: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.912010 | orchestrator | 2025-08-29 17:44:45.912027 | orchestrator | TASK [placement : Copying over placement.conf] ********************************* 2025-08-29 17:44:45.912042 | orchestrator | Friday 29 August 2025 17:43:58 +0000 (0:00:01.576) 0:00:37.732 ********* 2025-08-29 17:44:45.912065 | orchestrator | changed: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.912094 | orchestrator | changed: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.912124 | orchestrator | changed: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.912142 | orchestrator | 2025-08-29 17:44:45.912158 | orchestrator | TASK [placement : Copying over placement-api wsgi configuration] *************** 2025-08-29 17:44:45.912175 | orchestrator | Friday 29 August 2025 17:44:02 +0000 (0:00:04.779) 0:00:42.512 ********* 2025-08-29 17:44:45.912193 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) 2025-08-29 17:44:45.912209 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) 2025-08-29 17:44:45.912226 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2) 2025-08-29 17:44:45.912243 | orchestrator | 2025-08-29 17:44:45.912260 | orchestrator | TASK [placement : Copying over migrate-db.rc.j2 configuration] ***************** 2025-08-29 17:44:45.912276 | orchestrator | Friday 29 August 2025 17:44:05 +0000 (0:00:02.518) 0:00:45.030 ********* 2025-08-29 17:44:45.912291 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:44:45.912301 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:44:45.912310 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:44:45.912320 | orchestrator | 2025-08-29 17:44:45.912329 | orchestrator | TASK [placement : Copying over existing policy file] *************************** 2025-08-29 17:44:45.912338 | orchestrator | Friday 29 August 2025 17:44:07 +0000 (0:00:01.701) 0:00:46.731 ********* 2025-08-29 17:44:45.912364 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.912374 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:44:45.912392 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.912409 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:44:45.912419 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})  2025-08-29 17:44:45.912429 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:44:45.912438 | orchestrator | 2025-08-29 17:44:45.912448 | orchestrator | TASK [placement : Check placement containers] ********************************** 2025-08-29 17:44:45.912485 | orchestrator | Friday 29 August 2025 17:44:08 +0000 (0:00:01.150) 0:00:47.881 ********* 2025-08-29 17:44:45.912495 | orchestrator | changed: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.912510 | orchestrator | changed: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.912526 | orchestrator | changed: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/release/placement-api:12.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) 2025-08-29 17:44:45.912548 | orchestrator | 2025-08-29 17:44:45.912558 | orchestrator | TASK [placement : Creating placement databases] ******************************** 2025-08-29 17:44:45.912568 | orchestrator | Friday 29 August 2025 17:44:10 +0000 (0:00:02.383) 0:00:50.265 ********* 2025-08-29 17:44:45.912577 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:44:45.912587 | orchestrator | 2025-08-29 17:44:45.912596 | orchestrator | TASK [placement : Creating placement databases user and setting permissions] *** 2025-08-29 17:44:45.912606 | orchestrator | Friday 29 August 2025 17:44:13 +0000 (0:00:02.376) 0:00:52.641 ********* 2025-08-29 17:44:45.912615 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:44:45.912624 | orchestrator | 2025-08-29 17:44:45.912634 | orchestrator | TASK [placement : Running placement bootstrap container] *********************** 2025-08-29 17:44:45.912643 | orchestrator | Friday 29 August 2025 17:44:15 +0000 (0:00:02.409) 0:00:55.050 ********* 2025-08-29 17:44:45.912653 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:44:45.912662 | orchestrator | 2025-08-29 17:44:45.912671 | orchestrator | TASK [placement : Flush handlers] ********************************************** 2025-08-29 17:44:45.912681 | orchestrator | Friday 29 August 2025 17:44:31 +0000 (0:00:15.603) 0:01:10.654 ********* 2025-08-29 17:44:45.912690 | orchestrator | 2025-08-29 17:44:45.912700 | orchestrator | TASK [placement : Flush handlers] ********************************************** 2025-08-29 17:44:45.912709 | orchestrator | Friday 29 August 2025 17:44:31 +0000 (0:00:00.124) 0:01:10.778 ********* 2025-08-29 17:44:45.912719 | orchestrator | 2025-08-29 17:44:45.912728 | orchestrator | TASK [placement : Flush handlers] ********************************************** 2025-08-29 17:44:45.912737 | orchestrator | Friday 29 August 2025 17:44:31 +0000 (0:00:00.064) 0:01:10.842 ********* 2025-08-29 17:44:45.912747 | orchestrator | 2025-08-29 17:44:45.912756 | orchestrator | RUNNING HANDLER [placement : Restart placement-api container] ****************** 2025-08-29 17:44:45.912765 | orchestrator | Friday 29 August 2025 17:44:31 +0000 (0:00:00.065) 0:01:10.908 ********* 2025-08-29 17:44:45.912775 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:44:45.912784 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:44:45.912793 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:44:45.912803 | orchestrator | 2025-08-29 17:44:45.912812 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:44:45.912823 | orchestrator | testbed-node-0 : ok=21  changed=15  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:44:45.912833 | orchestrator | testbed-node-1 : ok=12  changed=8  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:44:45.912842 | orchestrator | testbed-node-2 : ok=12  changed=8  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:44:45.912852 | orchestrator | 2025-08-29 17:44:45.912861 | orchestrator | 2025-08-29 17:44:45.912871 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:44:45.912880 | orchestrator | Friday 29 August 2025 17:44:43 +0000 (0:00:12.156) 0:01:23.065 ********* 2025-08-29 17:44:45.912890 | orchestrator | =============================================================================== 2025-08-29 17:44:45.912899 | orchestrator | placement : Running placement bootstrap container ---------------------- 15.60s 2025-08-29 17:44:45.912908 | orchestrator | placement : Restart placement-api container ---------------------------- 12.16s 2025-08-29 17:44:45.912923 | orchestrator | service-ks-register : placement | Creating endpoints -------------------- 5.86s 2025-08-29 17:44:45.912932 | orchestrator | placement : Copying over placement.conf --------------------------------- 4.78s 2025-08-29 17:44:45.912942 | orchestrator | service-ks-register : placement | Granting user roles ------------------- 4.55s 2025-08-29 17:44:45.912951 | orchestrator | service-ks-register : placement | Creating users ------------------------ 4.31s 2025-08-29 17:44:45.912960 | orchestrator | service-ks-register : placement | Creating roles ------------------------ 3.27s 2025-08-29 17:44:45.912970 | orchestrator | service-ks-register : placement | Creating projects --------------------- 3.17s 2025-08-29 17:44:45.912979 | orchestrator | service-ks-register : placement | Creating services --------------------- 3.15s 2025-08-29 17:44:45.912988 | orchestrator | service-cert-copy : placement | Copying over extra CA certificates ------ 2.61s 2025-08-29 17:44:45.912998 | orchestrator | placement : Copying over placement-api wsgi configuration --------------- 2.52s 2025-08-29 17:44:45.913010 | orchestrator | placement : Creating placement databases user and setting permissions --- 2.41s 2025-08-29 17:44:45.913031 | orchestrator | placement : Check placement containers ---------------------------------- 2.38s 2025-08-29 17:44:45.913049 | orchestrator | placement : Creating placement databases -------------------------------- 2.38s 2025-08-29 17:44:45.913066 | orchestrator | service-cert-copy : placement | Copying over backend internal TLS certificate --- 2.15s 2025-08-29 17:44:45.913082 | orchestrator | service-cert-copy : placement | Copying over backend internal TLS key --- 1.81s 2025-08-29 17:44:45.913098 | orchestrator | placement : Copying over migrate-db.rc.j2 configuration ----------------- 1.70s 2025-08-29 17:44:45.913115 | orchestrator | placement : Copying over config.json files for services ----------------- 1.58s 2025-08-29 17:44:45.913130 | orchestrator | placement : Ensuring config directories exist --------------------------- 1.37s 2025-08-29 17:44:45.913144 | orchestrator | placement : Copying over existing policy file --------------------------- 1.15s 2025-08-29 17:44:45.913162 | orchestrator | 2025-08-29 17:44:45 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:44:45.913179 | orchestrator | 2025-08-29 17:44:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:48.932276 | orchestrator | 2025-08-29 17:44:48 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:48.932961 | orchestrator | 2025-08-29 17:44:48 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:48.934067 | orchestrator | 2025-08-29 17:44:48 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:48.934609 | orchestrator | 2025-08-29 17:44:48 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:44:48.934646 | orchestrator | 2025-08-29 17:44:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:51.959711 | orchestrator | 2025-08-29 17:44:51 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:51.960056 | orchestrator | 2025-08-29 17:44:51 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:51.961701 | orchestrator | 2025-08-29 17:44:51 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:51.962361 | orchestrator | 2025-08-29 17:44:51 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:44:51.962434 | orchestrator | 2025-08-29 17:44:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:54.984862 | orchestrator | 2025-08-29 17:44:54 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:54.986212 | orchestrator | 2025-08-29 17:44:54 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:54.986672 | orchestrator | 2025-08-29 17:44:54 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:54.987225 | orchestrator | 2025-08-29 17:44:54 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:44:54.987249 | orchestrator | 2025-08-29 17:44:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:44:58.033741 | orchestrator | 2025-08-29 17:44:58 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:44:58.034370 | orchestrator | 2025-08-29 17:44:58 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:44:58.036053 | orchestrator | 2025-08-29 17:44:58 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:44:58.036541 | orchestrator | 2025-08-29 17:44:58 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:44:58.036595 | orchestrator | 2025-08-29 17:44:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:01.059425 | orchestrator | 2025-08-29 17:45:01 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:01.061136 | orchestrator | 2025-08-29 17:45:01 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state STARTED 2025-08-29 17:45:01.062405 | orchestrator | 2025-08-29 17:45:01 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:01.062433 | orchestrator | 2025-08-29 17:45:01 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:01.062445 | orchestrator | 2025-08-29 17:45:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:04.087036 | orchestrator | 2025-08-29 17:45:04 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:04.088005 | orchestrator | 2025-08-29 17:45:04 | INFO  | Task bf9f670a-7703-48c2-a0dd-e0b856f7a285 is in state SUCCESS 2025-08-29 17:45:04.091212 | orchestrator | 2025-08-29 17:45:04.091297 | orchestrator | 2025-08-29 17:45:04.091313 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:45:04.091333 | orchestrator | 2025-08-29 17:45:04.091368 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:45:04.091388 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.317) 0:00:00.317 ********* 2025-08-29 17:45:04.091407 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:45:04.091477 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:45:04.091498 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:45:04.091514 | orchestrator | 2025-08-29 17:45:04.091531 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:45:04.091547 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.304) 0:00:00.621 ********* 2025-08-29 17:45:04.091563 | orchestrator | ok: [testbed-node-0] => (item=enable_barbican_True) 2025-08-29 17:45:04.091581 | orchestrator | ok: [testbed-node-1] => (item=enable_barbican_True) 2025-08-29 17:45:04.091598 | orchestrator | ok: [testbed-node-2] => (item=enable_barbican_True) 2025-08-29 17:45:04.091614 | orchestrator | 2025-08-29 17:45:04.091631 | orchestrator | PLAY [Apply role barbican] ***************************************************** 2025-08-29 17:45:04.091649 | orchestrator | 2025-08-29 17:45:04.091667 | orchestrator | TASK [barbican : include_tasks] ************************************************ 2025-08-29 17:45:04.091685 | orchestrator | Friday 29 August 2025 17:42:40 +0000 (0:00:00.636) 0:00:01.258 ********* 2025-08-29 17:45:04.091722 | orchestrator | included: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:45:04.091743 | orchestrator | 2025-08-29 17:45:04.091760 | orchestrator | TASK [service-ks-register : barbican | Creating services] ********************** 2025-08-29 17:45:04.091775 | orchestrator | Friday 29 August 2025 17:42:41 +0000 (0:00:00.629) 0:00:01.887 ********* 2025-08-29 17:45:04.091788 | orchestrator | changed: [testbed-node-0] => (item=barbican (key-manager)) 2025-08-29 17:45:04.091800 | orchestrator | 2025-08-29 17:45:04.091833 | orchestrator | TASK [service-ks-register : barbican | Creating endpoints] ********************* 2025-08-29 17:45:04.091845 | orchestrator | Friday 29 August 2025 17:42:44 +0000 (0:00:03.881) 0:00:05.769 ********* 2025-08-29 17:45:04.091857 | orchestrator | changed: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal) 2025-08-29 17:45:04.091869 | orchestrator | changed: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public) 2025-08-29 17:45:04.091880 | orchestrator | 2025-08-29 17:45:04.091892 | orchestrator | TASK [service-ks-register : barbican | Creating projects] ********************** 2025-08-29 17:45:04.091904 | orchestrator | Friday 29 August 2025 17:42:51 +0000 (0:00:06.495) 0:00:12.265 ********* 2025-08-29 17:45:04.091914 | orchestrator | FAILED - RETRYING: [testbed-node-0]: barbican | Creating projects (5 retries left). 2025-08-29 17:45:04.091924 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:45:04.091934 | orchestrator | 2025-08-29 17:45:04.091944 | orchestrator | TASK [service-ks-register : barbican | Creating users] ************************* 2025-08-29 17:45:04.091954 | orchestrator | Friday 29 August 2025 17:43:08 +0000 (0:00:16.778) 0:00:29.043 ********* 2025-08-29 17:45:04.091963 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:45:04.091973 | orchestrator | changed: [testbed-node-0] => (item=barbican -> service) 2025-08-29 17:45:04.091988 | orchestrator | 2025-08-29 17:45:04.092005 | orchestrator | TASK [service-ks-register : barbican | Creating roles] ************************* 2025-08-29 17:45:04.092021 | orchestrator | Friday 29 August 2025 17:43:12 +0000 (0:00:04.488) 0:00:33.532 ********* 2025-08-29 17:45:04.092038 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:45:04.092054 | orchestrator | changed: [testbed-node-0] => (item=key-manager:service-admin) 2025-08-29 17:45:04.092072 | orchestrator | changed: [testbed-node-0] => (item=creator) 2025-08-29 17:45:04.092088 | orchestrator | changed: [testbed-node-0] => (item=observer) 2025-08-29 17:45:04.092106 | orchestrator | changed: [testbed-node-0] => (item=audit) 2025-08-29 17:45:04.092123 | orchestrator | 2025-08-29 17:45:04.092141 | orchestrator | TASK [service-ks-register : barbican | Granting user roles] ******************** 2025-08-29 17:45:04.092158 | orchestrator | Friday 29 August 2025 17:43:28 +0000 (0:00:15.515) 0:00:49.048 ********* 2025-08-29 17:45:04.092175 | orchestrator | changed: [testbed-node-0] => (item=barbican -> service -> admin) 2025-08-29 17:45:04.092192 | orchestrator | 2025-08-29 17:45:04.092208 | orchestrator | TASK [barbican : Ensuring config directories exist] **************************** 2025-08-29 17:45:04.092224 | orchestrator | Friday 29 August 2025 17:43:32 +0000 (0:00:04.196) 0:00:53.244 ********* 2025-08-29 17:45:04.092244 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.092330 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.092368 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.092387 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.092406 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.092424 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.092487 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.092519 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.092538 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.092556 | orchestrator | 2025-08-29 17:45:04.092574 | orchestrator | TASK [barbican : Ensuring vassals config directories exist] ******************** 2025-08-29 17:45:04.092592 | orchestrator | Friday 29 August 2025 17:43:34 +0000 (0:00:02.327) 0:00:55.572 ********* 2025-08-29 17:45:04.092610 | orchestrator | changed: [testbed-node-0] => (item=barbican-api/vassals) 2025-08-29 17:45:04.092622 | orchestrator | changed: [testbed-node-1] => (item=barbican-api/vassals) 2025-08-29 17:45:04.092685 | orchestrator | changed: [testbed-node-2] => (item=barbican-api/vassals) 2025-08-29 17:45:04.092704 | orchestrator | 2025-08-29 17:45:04.092739 | orchestrator | TASK [barbican : Check if policies shall be overwritten] *********************** 2025-08-29 17:45:04.092776 | orchestrator | Friday 29 August 2025 17:43:36 +0000 (0:00:02.121) 0:00:57.693 ********* 2025-08-29 17:45:04.092795 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:04.092836 | orchestrator | 2025-08-29 17:45:04.092853 | orchestrator | TASK [barbican : Set barbican policy file] ************************************* 2025-08-29 17:45:04.092869 | orchestrator | Friday 29 August 2025 17:43:36 +0000 (0:00:00.083) 0:00:57.776 ********* 2025-08-29 17:45:04.092887 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:04.092903 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:04.092920 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:04.092936 | orchestrator | 2025-08-29 17:45:04.092952 | orchestrator | TASK [barbican : include_tasks] ************************************************ 2025-08-29 17:45:04.092969 | orchestrator | Friday 29 August 2025 17:43:37 +0000 (0:00:00.532) 0:00:58.308 ********* 2025-08-29 17:45:04.092986 | orchestrator | included: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:45:04.093003 | orchestrator | 2025-08-29 17:45:04.093019 | orchestrator | TASK [service-cert-copy : barbican | Copying over extra CA certificates] ******* 2025-08-29 17:45:04.093036 | orchestrator | Friday 29 August 2025 17:43:37 +0000 (0:00:00.399) 0:00:58.708 ********* 2025-08-29 17:45:04.093054 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.093100 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.093119 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.093136 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.093153 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.093171 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.093207 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.093240 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.093257 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.093274 | orchestrator | 2025-08-29 17:45:04.093290 | orchestrator | TASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] *** 2025-08-29 17:45:04.093327 | orchestrator | Friday 29 August 2025 17:43:41 +0000 (0:00:03.693) 0:01:02.402 ********* 2025-08-29 17:45:04.093363 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.093380 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.093417 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.093535 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:04.093612 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.093631 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.093667 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.093703 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:04.093745 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.093781 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.093835 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.093871 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:04.093905 | orchestrator | 2025-08-29 17:45:04.093928 | orchestrator | TASK [service-cert-copy : barbican | Copying over backend internal TLS key] **** 2025-08-29 17:45:04.093982 | orchestrator | Friday 29 August 2025 17:43:42 +0000 (0:00:01.311) 0:01:03.714 ********* 2025-08-29 17:45:04.094068 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.094089 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.094106 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.094123 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:04.094140 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.094166 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.094197 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.094214 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:04.094231 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.094248 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.094265 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.094290 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:04.094306 | orchestrator | 2025-08-29 17:45:04.094323 | orchestrator | TASK [barbican : Copying over config.json files for services] ****************** 2025-08-29 17:45:04.094339 | orchestrator | Friday 29 August 2025 17:43:44 +0000 (0:00:01.767) 0:01:05.481 ********* 2025-08-29 17:45:04.094355 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.094388 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.094406 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.094424 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.094451 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.094490 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.094522 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.094541 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.094559 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.094576 | orchestrator | 2025-08-29 17:45:04.094592 | orchestrator | TASK [barbican : Copying over barbican-api.ini] ******************************** 2025-08-29 17:45:04.094608 | orchestrator | Friday 29 August 2025 17:43:48 +0000 (0:00:03.644) 0:01:09.125 ********* 2025-08-29 17:45:04.094624 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:04.094641 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:04.094656 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:04.094672 | orchestrator | 2025-08-29 17:45:04.094689 | orchestrator | TASK [barbican : Checking whether barbican-api-paste.ini file exists] ********** 2025-08-29 17:45:04.094706 | orchestrator | Friday 29 August 2025 17:43:50 +0000 (0:00:02.193) 0:01:11.318 ********* 2025-08-29 17:45:04.094735 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:45:04.094752 | orchestrator | 2025-08-29 17:45:04.094767 | orchestrator | TASK [barbican : Copying over barbican-api-paste.ini] ************************** 2025-08-29 17:45:04.094783 | orchestrator | Friday 29 August 2025 17:43:53 +0000 (0:00:02.799) 0:01:14.118 ********* 2025-08-29 17:45:04.094799 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:04.094815 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:04.094831 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:04.094847 | orchestrator | 2025-08-29 17:45:04.094863 | orchestrator | TASK [barbican : Copying over barbican.conf] *********************************** 2025-08-29 17:45:04.094879 | orchestrator | Friday 29 August 2025 17:43:54 +0000 (0:00:01.692) 0:01:15.810 ********* 2025-08-29 17:45:04.094894 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.094928 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.094946 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.094964 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.094989 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095006 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095022 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095053 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095070 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095087 | orchestrator | 2025-08-29 17:45:04.095103 | orchestrator | TASK [barbican : Copying over existing policy file] **************************** 2025-08-29 17:45:04.095119 | orchestrator | Friday 29 August 2025 17:44:06 +0000 (0:00:11.081) 0:01:26.892 ********* 2025-08-29 17:45:04.095136 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.095197 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.095215 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.095232 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:04.095257 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.095280 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.095300 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.095325 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:04.095343 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})  2025-08-29 17:45:04.095360 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.095378 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:04.095395 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:04.095411 | orchestrator | 2025-08-29 17:45:04.095428 | orchestrator | TASK [barbican : Check barbican containers] ************************************ 2025-08-29 17:45:04.095445 | orchestrator | Friday 29 August 2025 17:44:06 +0000 (0:00:00.550) 0:01:27.442 ********* 2025-08-29 17:45:04.095539 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.095561 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.095591 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) 2025-08-29 17:45:04.095609 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095627 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095654 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095672 | orchestrator | changed: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095720 | orchestrator | changed: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095783 | orchestrator | changed: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:04.095804 | orchestrator | 2025-08-29 17:45:04.095821 | orchestrator | TASK [barbican : include_tasks] ************************************************ 2025-08-29 17:45:04.095838 | orchestrator | Friday 29 August 2025 17:44:10 +0000 (0:00:03.530) 0:01:30.973 ********* 2025-08-29 17:45:04.095854 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:04.095871 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:04.095888 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:04.095905 | orchestrator | 2025-08-29 17:45:04.095922 | orchestrator | TASK [barbican : Creating barbican database] *********************************** 2025-08-29 17:45:04.095938 | orchestrator | Friday 29 August 2025 17:44:10 +0000 (0:00:00.387) 0:01:31.361 ********* 2025-08-29 17:45:04.095955 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:04.095971 | orchestrator | 2025-08-29 17:45:04.095988 | orchestrator | TASK [barbican : Creating barbican database user and setting permissions] ****** 2025-08-29 17:45:04.096004 | orchestrator | Friday 29 August 2025 17:44:12 +0000 (0:00:02.377) 0:01:33.738 ********* 2025-08-29 17:45:04.096020 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:04.096037 | orchestrator | 2025-08-29 17:45:04.096051 | orchestrator | TASK [barbican : Running barbican bootstrap container] ************************* 2025-08-29 17:45:04.096064 | orchestrator | Friday 29 August 2025 17:44:15 +0000 (0:00:02.569) 0:01:36.308 ********* 2025-08-29 17:45:04.096076 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:04.096088 | orchestrator | 2025-08-29 17:45:04.096100 | orchestrator | TASK [barbican : Flush handlers] *********************************************** 2025-08-29 17:45:04.096112 | orchestrator | Friday 29 August 2025 17:44:29 +0000 (0:00:14.006) 0:01:50.315 ********* 2025-08-29 17:45:04.096124 | orchestrator | 2025-08-29 17:45:04.096137 | orchestrator | TASK [barbican : Flush handlers] *********************************************** 2025-08-29 17:45:04.096150 | orchestrator | Friday 29 August 2025 17:44:29 +0000 (0:00:00.116) 0:01:50.431 ********* 2025-08-29 17:45:04.096162 | orchestrator | 2025-08-29 17:45:04.096175 | orchestrator | TASK [barbican : Flush handlers] *********************************************** 2025-08-29 17:45:04.096187 | orchestrator | Friday 29 August 2025 17:44:29 +0000 (0:00:00.186) 0:01:50.617 ********* 2025-08-29 17:45:04.096199 | orchestrator | 2025-08-29 17:45:04.096212 | orchestrator | RUNNING HANDLER [barbican : Restart barbican-api container] ******************** 2025-08-29 17:45:04.096225 | orchestrator | Friday 29 August 2025 17:44:29 +0000 (0:00:00.144) 0:01:50.762 ********* 2025-08-29 17:45:04.096248 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:04.096262 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:04.096275 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:04.096288 | orchestrator | 2025-08-29 17:45:04.096303 | orchestrator | RUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ****** 2025-08-29 17:45:04.096318 | orchestrator | Friday 29 August 2025 17:44:43 +0000 (0:00:13.460) 0:02:04.222 ********* 2025-08-29 17:45:04.096342 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:04.096356 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:04.096370 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:04.096383 | orchestrator | 2025-08-29 17:45:04.096403 | orchestrator | RUNNING HANDLER [barbican : Restart barbican-worker container] ***************** 2025-08-29 17:45:04.096417 | orchestrator | Friday 29 August 2025 17:44:53 +0000 (0:00:10.630) 0:02:14.853 ********* 2025-08-29 17:45:04.096431 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:04.096444 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:04.096478 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:04.096492 | orchestrator | 2025-08-29 17:45:04.096504 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:45:04.096518 | orchestrator | testbed-node-0 : ok=24  changed=18  unreachable=0 failed=0 skipped=7  rescued=0 ignored=0 2025-08-29 17:45:04.096532 | orchestrator | testbed-node-1 : ok=14  changed=10  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:45:04.096545 | orchestrator | testbed-node-2 : ok=14  changed=10  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:45:04.096558 | orchestrator | 2025-08-29 17:45:04.096571 | orchestrator | 2025-08-29 17:45:04.096584 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:45:04.096598 | orchestrator | Friday 29 August 2025 17:45:02 +0000 (0:00:08.358) 0:02:23.211 ********* 2025-08-29 17:45:04.096611 | orchestrator | =============================================================================== 2025-08-29 17:45:04.096624 | orchestrator | service-ks-register : barbican | Creating projects --------------------- 16.78s 2025-08-29 17:45:04.096638 | orchestrator | service-ks-register : barbican | Creating roles ------------------------ 15.52s 2025-08-29 17:45:04.096651 | orchestrator | barbican : Running barbican bootstrap container ------------------------ 14.01s 2025-08-29 17:45:04.096663 | orchestrator | barbican : Restart barbican-api container ------------------------------ 13.46s 2025-08-29 17:45:04.096676 | orchestrator | barbican : Copying over barbican.conf ---------------------------------- 11.08s 2025-08-29 17:45:04.096688 | orchestrator | barbican : Restart barbican-keystone-listener container ---------------- 10.63s 2025-08-29 17:45:04.096701 | orchestrator | barbican : Restart barbican-worker container ---------------------------- 8.36s 2025-08-29 17:45:04.096714 | orchestrator | service-ks-register : barbican | Creating endpoints --------------------- 6.50s 2025-08-29 17:45:04.096727 | orchestrator | service-ks-register : barbican | Creating users ------------------------- 4.49s 2025-08-29 17:45:04.096740 | orchestrator | service-ks-register : barbican | Granting user roles -------------------- 4.20s 2025-08-29 17:45:04.096753 | orchestrator | service-ks-register : barbican | Creating services ---------------------- 3.88s 2025-08-29 17:45:04.096766 | orchestrator | service-cert-copy : barbican | Copying over extra CA certificates ------- 3.69s 2025-08-29 17:45:04.096779 | orchestrator | barbican : Copying over config.json files for services ------------------ 3.64s 2025-08-29 17:45:04.096791 | orchestrator | barbican : Check barbican containers ------------------------------------ 3.53s 2025-08-29 17:45:04.096804 | orchestrator | barbican : Checking whether barbican-api-paste.ini file exists ---------- 2.80s 2025-08-29 17:45:04.096817 | orchestrator | barbican : Creating barbican database user and setting permissions ------ 2.57s 2025-08-29 17:45:04.096830 | orchestrator | barbican : Creating barbican database ----------------------------------- 2.38s 2025-08-29 17:45:04.096852 | orchestrator | barbican : Ensuring config directories exist ---------------------------- 2.33s 2025-08-29 17:45:04.096865 | orchestrator | barbican : Copying over barbican-api.ini -------------------------------- 2.19s 2025-08-29 17:45:04.096878 | orchestrator | barbican : Ensuring vassals config directories exist -------------------- 2.12s 2025-08-29 17:45:04.096891 | orchestrator | 2025-08-29 17:45:04 | INFO  | Task 78911c18-6f6a-4140-9ac7-584aa6ee024e is in state STARTED 2025-08-29 17:45:04.096905 | orchestrator | 2025-08-29 17:45:04 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:04.096918 | orchestrator | 2025-08-29 17:45:04 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:04.096931 | orchestrator | 2025-08-29 17:45:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:07.128347 | orchestrator | 2025-08-29 17:45:07 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:07.128444 | orchestrator | 2025-08-29 17:45:07 | INFO  | Task 78911c18-6f6a-4140-9ac7-584aa6ee024e is in state STARTED 2025-08-29 17:45:07.129832 | orchestrator | 2025-08-29 17:45:07 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:07.131540 | orchestrator | 2025-08-29 17:45:07 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:07.131580 | orchestrator | 2025-08-29 17:45:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:10.177224 | orchestrator | 2025-08-29 17:45:10 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:10.180558 | orchestrator | 2025-08-29 17:45:10 | INFO  | Task 78911c18-6f6a-4140-9ac7-584aa6ee024e is in state SUCCESS 2025-08-29 17:45:10.186967 | orchestrator | 2025-08-29 17:45:10 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:10.189669 | orchestrator | 2025-08-29 17:45:10 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:10.193453 | orchestrator | 2025-08-29 17:45:10 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:10.194529 | orchestrator | 2025-08-29 17:45:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:13.251191 | orchestrator | 2025-08-29 17:45:13 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:13.252635 | orchestrator | 2025-08-29 17:45:13 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:13.254967 | orchestrator | 2025-08-29 17:45:13 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:13.257398 | orchestrator | 2025-08-29 17:45:13 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:13.258181 | orchestrator | 2025-08-29 17:45:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:16.310932 | orchestrator | 2025-08-29 17:45:16 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:16.311394 | orchestrator | 2025-08-29 17:45:16 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:16.313706 | orchestrator | 2025-08-29 17:45:16 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:16.314504 | orchestrator | 2025-08-29 17:45:16 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:16.314639 | orchestrator | 2025-08-29 17:45:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:19.356165 | orchestrator | 2025-08-29 17:45:19 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:19.359390 | orchestrator | 2025-08-29 17:45:19 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:19.360750 | orchestrator | 2025-08-29 17:45:19 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:19.362366 | orchestrator | 2025-08-29 17:45:19 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:19.362383 | orchestrator | 2025-08-29 17:45:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:22.403947 | orchestrator | 2025-08-29 17:45:22 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:22.404658 | orchestrator | 2025-08-29 17:45:22 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:22.405707 | orchestrator | 2025-08-29 17:45:22 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:22.406666 | orchestrator | 2025-08-29 17:45:22 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:22.406697 | orchestrator | 2025-08-29 17:45:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:25.457654 | orchestrator | 2025-08-29 17:45:25 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:25.458995 | orchestrator | 2025-08-29 17:45:25 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:25.460936 | orchestrator | 2025-08-29 17:45:25 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:25.462253 | orchestrator | 2025-08-29 17:45:25 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:25.462295 | orchestrator | 2025-08-29 17:45:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:28.499992 | orchestrator | 2025-08-29 17:45:28 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:28.500087 | orchestrator | 2025-08-29 17:45:28 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:28.500430 | orchestrator | 2025-08-29 17:45:28 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:28.502123 | orchestrator | 2025-08-29 17:45:28 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:28.502215 | orchestrator | 2025-08-29 17:45:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:31.550826 | orchestrator | 2025-08-29 17:45:31 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:31.551055 | orchestrator | 2025-08-29 17:45:31 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:31.552547 | orchestrator | 2025-08-29 17:45:31 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:31.553340 | orchestrator | 2025-08-29 17:45:31 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:31.554425 | orchestrator | 2025-08-29 17:45:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:34.596398 | orchestrator | 2025-08-29 17:45:34 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:34.596489 | orchestrator | 2025-08-29 17:45:34 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:34.596500 | orchestrator | 2025-08-29 17:45:34 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:34.596507 | orchestrator | 2025-08-29 17:45:34 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:34.596514 | orchestrator | 2025-08-29 17:45:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:37.628286 | orchestrator | 2025-08-29 17:45:37 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:37.629428 | orchestrator | 2025-08-29 17:45:37 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:37.631196 | orchestrator | 2025-08-29 17:45:37 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:37.632926 | orchestrator | 2025-08-29 17:45:37 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:37.632995 | orchestrator | 2025-08-29 17:45:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:40.718677 | orchestrator | 2025-08-29 17:45:40 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:40.718746 | orchestrator | 2025-08-29 17:45:40 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:40.720790 | orchestrator | 2025-08-29 17:45:40 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:40.725893 | orchestrator | 2025-08-29 17:45:40 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:40.726127 | orchestrator | 2025-08-29 17:45:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:43.774821 | orchestrator | 2025-08-29 17:45:43 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:43.774979 | orchestrator | 2025-08-29 17:45:43 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state STARTED 2025-08-29 17:45:43.775622 | orchestrator | 2025-08-29 17:45:43 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:43.776831 | orchestrator | 2025-08-29 17:45:43 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:43.776859 | orchestrator | 2025-08-29 17:45:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:46.801174 | orchestrator | 2025-08-29 17:45:46 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:46.801288 | orchestrator | 2025-08-29 17:45:46 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:45:46.803060 | orchestrator | 2025-08-29 17:45:46 | INFO  | Task 6e25d301-fb90-489b-a591-3a9b3a21ec01 is in state SUCCESS 2025-08-29 17:45:46.806852 | orchestrator | 2025-08-29 17:45:46.806919 | orchestrator | 2025-08-29 17:45:46.806926 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:45:46.806932 | orchestrator | 2025-08-29 17:45:46.806937 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:45:46.806942 | orchestrator | Friday 29 August 2025 17:45:06 +0000 (0:00:00.170) 0:00:00.170 ********* 2025-08-29 17:45:46.806946 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:45:46.806952 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:45:46.806957 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:45:46.806961 | orchestrator | 2025-08-29 17:45:46.806966 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:45:46.806971 | orchestrator | Friday 29 August 2025 17:45:06 +0000 (0:00:00.263) 0:00:00.433 ********* 2025-08-29 17:45:46.806976 | orchestrator | ok: [testbed-node-0] => (item=enable_keystone_True) 2025-08-29 17:45:46.806981 | orchestrator | ok: [testbed-node-1] => (item=enable_keystone_True) 2025-08-29 17:45:46.806985 | orchestrator | ok: [testbed-node-2] => (item=enable_keystone_True) 2025-08-29 17:45:46.806990 | orchestrator | 2025-08-29 17:45:46.806994 | orchestrator | PLAY [Wait for the Keystone service] ******************************************* 2025-08-29 17:45:46.806999 | orchestrator | 2025-08-29 17:45:46.807004 | orchestrator | TASK [Waiting for Keystone public port to be UP] ******************************* 2025-08-29 17:45:46.807032 | orchestrator | Friday 29 August 2025 17:45:07 +0000 (0:00:00.628) 0:00:01.062 ********* 2025-08-29 17:45:46.807038 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:45:46.807042 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:45:46.807047 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:45:46.807051 | orchestrator | 2025-08-29 17:45:46.807056 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:45:46.807071 | orchestrator | testbed-node-0 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:45:46.807077 | orchestrator | testbed-node-1 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:45:46.807082 | orchestrator | testbed-node-2 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:45:46.807086 | orchestrator | 2025-08-29 17:45:46.807091 | orchestrator | 2025-08-29 17:45:46.807096 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:45:46.807100 | orchestrator | Friday 29 August 2025 17:45:08 +0000 (0:00:00.780) 0:00:01.842 ********* 2025-08-29 17:45:46.807105 | orchestrator | =============================================================================== 2025-08-29 17:45:46.807109 | orchestrator | Waiting for Keystone public port to be UP ------------------------------- 0.78s 2025-08-29 17:45:46.807114 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.63s 2025-08-29 17:45:46.807119 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.26s 2025-08-29 17:45:46.807123 | orchestrator | 2025-08-29 17:45:46.807128 | orchestrator | 2025-08-29 17:45:46.807132 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:45:46.807137 | orchestrator | 2025-08-29 17:45:46.807142 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:45:46.807147 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.324) 0:00:00.324 ********* 2025-08-29 17:45:46.807151 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:45:46.807156 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:45:46.807160 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:45:46.807165 | orchestrator | 2025-08-29 17:45:46.807170 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:45:46.807174 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.322) 0:00:00.647 ********* 2025-08-29 17:45:46.807179 | orchestrator | ok: [testbed-node-0] => (item=enable_designate_True) 2025-08-29 17:45:46.807184 | orchestrator | ok: [testbed-node-1] => (item=enable_designate_True) 2025-08-29 17:45:46.807188 | orchestrator | ok: [testbed-node-2] => (item=enable_designate_True) 2025-08-29 17:45:46.807193 | orchestrator | 2025-08-29 17:45:46.807197 | orchestrator | PLAY [Apply role designate] **************************************************** 2025-08-29 17:45:46.807202 | orchestrator | 2025-08-29 17:45:46.807207 | orchestrator | TASK [designate : include_tasks] *********************************************** 2025-08-29 17:45:46.807211 | orchestrator | Friday 29 August 2025 17:42:40 +0000 (0:00:00.621) 0:00:01.269 ********* 2025-08-29 17:45:46.807216 | orchestrator | included: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:45:46.807221 | orchestrator | 2025-08-29 17:45:46.807225 | orchestrator | TASK [service-ks-register : designate | Creating services] ********************* 2025-08-29 17:45:46.807230 | orchestrator | Friday 29 August 2025 17:42:40 +0000 (0:00:00.672) 0:00:01.941 ********* 2025-08-29 17:45:46.807234 | orchestrator | changed: [testbed-node-0] => (item=designate (dns)) 2025-08-29 17:45:46.807239 | orchestrator | 2025-08-29 17:45:46.807244 | orchestrator | TASK [service-ks-register : designate | Creating endpoints] ******************** 2025-08-29 17:45:46.807248 | orchestrator | Friday 29 August 2025 17:42:44 +0000 (0:00:03.799) 0:00:05.740 ********* 2025-08-29 17:45:46.807253 | orchestrator | changed: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal) 2025-08-29 17:45:46.807257 | orchestrator | changed: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public) 2025-08-29 17:45:46.807267 | orchestrator | 2025-08-29 17:45:46.807271 | orchestrator | TASK [service-ks-register : designate | Creating projects] ********************* 2025-08-29 17:45:46.807276 | orchestrator | Friday 29 August 2025 17:42:51 +0000 (0:00:06.656) 0:00:12.397 ********* 2025-08-29 17:45:46.807281 | orchestrator | changed: [testbed-node-0] => (item=service) 2025-08-29 17:45:46.807285 | orchestrator | 2025-08-29 17:45:46.807290 | orchestrator | TASK [service-ks-register : designate | Creating users] ************************ 2025-08-29 17:45:46.807295 | orchestrator | Friday 29 August 2025 17:42:54 +0000 (0:00:03.329) 0:00:15.726 ********* 2025-08-29 17:45:46.807309 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:45:46.807314 | orchestrator | changed: [testbed-node-0] => (item=designate -> service) 2025-08-29 17:45:46.807318 | orchestrator | 2025-08-29 17:45:46.807323 | orchestrator | TASK [service-ks-register : designate | Creating roles] ************************ 2025-08-29 17:45:46.807327 | orchestrator | Friday 29 August 2025 17:42:59 +0000 (0:00:04.405) 0:00:20.132 ********* 2025-08-29 17:45:46.807332 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:45:46.807336 | orchestrator | 2025-08-29 17:45:46.807341 | orchestrator | TASK [service-ks-register : designate | Granting user roles] ******************* 2025-08-29 17:45:46.807345 | orchestrator | Friday 29 August 2025 17:43:02 +0000 (0:00:03.703) 0:00:23.835 ********* 2025-08-29 17:45:46.807350 | orchestrator | changed: [testbed-node-0] => (item=designate -> service -> admin) 2025-08-29 17:45:46.807354 | orchestrator | 2025-08-29 17:45:46.807359 | orchestrator | TASK [designate : Ensuring config directories exist] *************************** 2025-08-29 17:45:46.807363 | orchestrator | Friday 29 August 2025 17:43:07 +0000 (0:00:04.264) 0:00:28.099 ********* 2025-08-29 17:45:46.807373 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.807381 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.807386 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.807394 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807406 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807414 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807419 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807424 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807428 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807437 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807447 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807452 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807459 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807480 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807485 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807493 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807498 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807508 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.807513 | orchestrator | 2025-08-29 17:45:46.807517 | orchestrator | TASK [designate : Check if policies shall be overwritten] ********************** 2025-08-29 17:45:46.807522 | orchestrator | Friday 29 August 2025 17:43:10 +0000 (0:00:03.043) 0:00:31.143 ********* 2025-08-29 17:45:46.807527 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:46.807532 | orchestrator | 2025-08-29 17:45:46.807536 | orchestrator | TASK [designate : Set designate policy file] *********************************** 2025-08-29 17:45:46.807541 | orchestrator | Friday 29 August 2025 17:43:10 +0000 (0:00:00.121) 0:00:31.264 ********* 2025-08-29 17:45:46.807546 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:46.807550 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:46.807555 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:46.807559 | orchestrator | 2025-08-29 17:45:46.807564 | orchestrator | TASK [designate : include_tasks] *********************************************** 2025-08-29 17:45:46.807568 | orchestrator | Friday 29 August 2025 17:43:10 +0000 (0:00:00.282) 0:00:31.546 ********* 2025-08-29 17:45:46.807573 | orchestrator | included: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:45:46.807577 | orchestrator | 2025-08-29 17:45:46.807585 | orchestrator | TASK [service-cert-copy : designate | Copying over extra CA certificates] ****** 2025-08-29 17:45:46.807590 | orchestrator | Friday 29 August 2025 17:43:11 +0000 (0:00:00.736) 0:00:32.283 ********* 2025-08-29 17:45:46.807595 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.808325 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.808343 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.808418 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808427 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808439 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808450 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808457 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808477 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808501 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808509 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808518 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808524 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808538 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808544 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808549 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808571 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808577 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.808583 | orchestrator | 2025-08-29 17:45:46.808588 | orchestrator | TASK [service-cert-copy : designate | Copying over backend internal TLS certificate] *** 2025-08-29 17:45:46.808594 | orchestrator | Friday 29 August 2025 17:43:17 +0000 (0:00:06.607) 0:00:38.890 ********* 2025-08-29 17:45:46.808603 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.808613 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.808618 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808623 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808643 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808650 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808655 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:46.808664 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.808674 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.808680 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808685 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808705 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808711 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808716 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:46.808728 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.808734 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.808739 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808744 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808765 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808771 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808779 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:46.808785 | orchestrator | 2025-08-29 17:45:46.808790 | orchestrator | TASK [service-cert-copy : designate | Copying over backend internal TLS key] *** 2025-08-29 17:45:46.808795 | orchestrator | Friday 29 August 2025 17:43:19 +0000 (0:00:01.466) 0:00:40.357 ********* 2025-08-29 17:45:46.808803 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.808809 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.808814 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808819 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808839 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808845 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808854 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:46.808862 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.808867 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.808873 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808878 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808897 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808903 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808912 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:46.808919 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.808925 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.808930 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808936 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808955 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808965 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.808970 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:46.808975 | orchestrator | 2025-08-29 17:45:46.808980 | orchestrator | TASK [designate : Copying over config.json files for services] ***************** 2025-08-29 17:45:46.808986 | orchestrator | Friday 29 August 2025 17:43:20 +0000 (0:00:01.557) 0:00:41.914 ********* 2025-08-29 17:45:46.808996 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.809005 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.809014 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.809048 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809068 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809082 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809088 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809102 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809108 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809113 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809148 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809155 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809164 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809169 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809175 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809180 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809202 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809212 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809218 | orchestrator | 2025-08-29 17:45:46.809223 | orchestrator | TASK [designate : Copying over designate.conf] ********************************* 2025-08-29 17:45:46.809228 | orchestrator | Friday 29 August 2025 17:43:27 +0000 (0:00:06.261) 0:00:48.176 ********* 2025-08-29 17:45:46.809236 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.809242 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.809249 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.809266 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809278 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809292 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809301 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809309 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809318 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809332 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809348 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809357 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809386 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809397 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809406 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809415 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809428 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809443 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809453 | orchestrator | 2025-08-29 17:45:46.809481 | orchestrator | TASK [designate : Copying over pools.yaml] ************************************* 2025-08-29 17:45:46.809491 | orchestrator | Friday 29 August 2025 17:43:50 +0000 (0:00:22.885) 0:01:11.061 ********* 2025-08-29 17:45:46.809499 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2) 2025-08-29 17:45:46.809505 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2) 2025-08-29 17:45:46.809510 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2) 2025-08-29 17:45:46.809515 | orchestrator | 2025-08-29 17:45:46.809520 | orchestrator | TASK [designate : Copying over named.conf] ************************************* 2025-08-29 17:45:46.809526 | orchestrator | Friday 29 August 2025 17:43:58 +0000 (0:00:07.956) 0:01:19.017 ********* 2025-08-29 17:45:46.809534 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2) 2025-08-29 17:45:46.809539 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2) 2025-08-29 17:45:46.809544 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2) 2025-08-29 17:45:46.809549 | orchestrator | 2025-08-29 17:45:46.809554 | orchestrator | TASK [designate : Copying over rndc.conf] ************************************** 2025-08-29 17:45:46.809560 | orchestrator | Friday 29 August 2025 17:44:02 +0000 (0:00:04.896) 0:01:23.914 ********* 2025-08-29 17:45:46.809565 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.809571 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.809595 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.809601 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809609 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809615 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809620 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809630 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809635 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809644 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809650 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809660 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809666 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809674 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809680 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809688 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809693 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809701 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809707 | orchestrator | 2025-08-29 17:45:46.809712 | orchestrator | TASK [designate : Copying over rndc.key] *************************************** 2025-08-29 17:45:46.809717 | orchestrator | Friday 29 August 2025 17:44:06 +0000 (0:00:03.661) 0:01:27.575 ********* 2025-08-29 17:45:46.809722 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.809731 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.809736 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.809745 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.809753 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.809758 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.810522 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810535 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810542 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810555 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810562 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810573 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.810579 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810589 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810595 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.810601 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811292 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811305 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811312 | orchestrator | 2025-08-29 17:45:46.811318 | orchestrator | TASK [designate : include_tasks] *********************************************** 2025-08-29 17:45:46.811329 | orchestrator | Friday 29 August 2025 17:44:10 +0000 (0:00:03.835) 0:01:31.410 ********* 2025-08-29 17:45:46.811334 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:46.811345 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:46.811351 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:46.811356 | orchestrator | 2025-08-29 17:45:46.811361 | orchestrator | TASK [designate : Copying over existing policy file] *************************** 2025-08-29 17:45:46.811366 | orchestrator | Friday 29 August 2025 17:44:10 +0000 (0:00:00.420) 0:01:31.831 ********* 2025-08-29 17:45:46.811372 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.811378 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.811384 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811395 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811401 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811410 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811419 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:46.811424 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.811430 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.811435 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811444 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811449 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})  2025-08-29 17:45:46.811460 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811481 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})  2025-08-29 17:45:46.811490 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811500 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811514 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811522 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:46.811527 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811542 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})  2025-08-29 17:45:46.811548 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:46.811553 | orchestrator | 2025-08-29 17:45:46.811558 | orchestrator | TASK [designate : Check designate containers] ********************************** 2025-08-29 17:45:46.811563 | orchestrator | Friday 29 August 2025 17:44:11 +0000 (0:00:00.613) 0:01:32.444 ********* 2025-08-29 17:45:46.811569 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.811575 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.811583 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-api:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) 2025-08-29 17:45:46.811589 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811600 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811606 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811611 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811617 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811622 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-central:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811634 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811653 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811670 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811679 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811688 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811697 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811706 | orchestrator | changed: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811717 | orchestrator | changed: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811727 | orchestrator | changed: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) 2025-08-29 17:45:46.811732 | orchestrator | 2025-08-29 17:45:46.811740 | orchestrator | TASK [designate : include_tasks] *********************************************** 2025-08-29 17:45:46.811746 | orchestrator | Friday 29 August 2025 17:44:15 +0000 (0:00:04.354) 0:01:36.798 ********* 2025-08-29 17:45:46.811751 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:45:46.811756 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:45:46.811761 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:45:46.811766 | orchestrator | 2025-08-29 17:45:46.811771 | orchestrator | TASK [designate : Creating Designate databases] ******************************** 2025-08-29 17:45:46.811776 | orchestrator | Friday 29 August 2025 17:44:16 +0000 (0:00:00.331) 0:01:37.130 ********* 2025-08-29 17:45:46.811782 | orchestrator | changed: [testbed-node-0] => (item=designate) 2025-08-29 17:45:46.811787 | orchestrator | 2025-08-29 17:45:46.811792 | orchestrator | TASK [designate : Creating Designate databases user and setting permissions] *** 2025-08-29 17:45:46.811797 | orchestrator | Friday 29 August 2025 17:44:18 +0000 (0:00:02.556) 0:01:39.686 ********* 2025-08-29 17:45:46.811802 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:45:46.811808 | orchestrator | changed: [testbed-node-0 -> {{ groups['designate-central'][0] }}] 2025-08-29 17:45:46.811813 | orchestrator | 2025-08-29 17:45:46.811818 | orchestrator | TASK [designate : Running Designate bootstrap container] *********************** 2025-08-29 17:45:46.811823 | orchestrator | Friday 29 August 2025 17:44:21 +0000 (0:00:02.791) 0:01:42.477 ********* 2025-08-29 17:45:46.811828 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.811833 | orchestrator | 2025-08-29 17:45:46.811838 | orchestrator | TASK [designate : Flush handlers] ********************************************** 2025-08-29 17:45:46.811843 | orchestrator | Friday 29 August 2025 17:44:36 +0000 (0:00:15.188) 0:01:57.666 ********* 2025-08-29 17:45:46.811848 | orchestrator | 2025-08-29 17:45:46.811853 | orchestrator | TASK [designate : Flush handlers] ********************************************** 2025-08-29 17:45:46.811858 | orchestrator | Friday 29 August 2025 17:44:36 +0000 (0:00:00.060) 0:01:57.726 ********* 2025-08-29 17:45:46.811863 | orchestrator | 2025-08-29 17:45:46.811868 | orchestrator | TASK [designate : Flush handlers] ********************************************** 2025-08-29 17:45:46.811873 | orchestrator | Friday 29 August 2025 17:44:36 +0000 (0:00:00.060) 0:01:57.787 ********* 2025-08-29 17:45:46.811878 | orchestrator | 2025-08-29 17:45:46.811883 | orchestrator | RUNNING HANDLER [designate : Restart designate-backend-bind9 container] ******** 2025-08-29 17:45:46.811888 | orchestrator | Friday 29 August 2025 17:44:36 +0000 (0:00:00.062) 0:01:57.850 ********* 2025-08-29 17:45:46.811893 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.811898 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:46.811904 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:46.811909 | orchestrator | 2025-08-29 17:45:46.811914 | orchestrator | RUNNING HANDLER [designate : Restart designate-api container] ****************** 2025-08-29 17:45:46.811923 | orchestrator | Friday 29 August 2025 17:44:50 +0000 (0:00:13.191) 0:02:11.041 ********* 2025-08-29 17:45:46.811928 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:46.811933 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:46.811939 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.811944 | orchestrator | 2025-08-29 17:45:46.811949 | orchestrator | RUNNING HANDLER [designate : Restart designate-central container] ************** 2025-08-29 17:45:46.811954 | orchestrator | Friday 29 August 2025 17:44:59 +0000 (0:00:09.761) 0:02:20.802 ********* 2025-08-29 17:45:46.811959 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.811964 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:46.811969 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:46.811974 | orchestrator | 2025-08-29 17:45:46.811979 | orchestrator | RUNNING HANDLER [designate : Restart designate-producer container] ************* 2025-08-29 17:45:46.811984 | orchestrator | Friday 29 August 2025 17:45:06 +0000 (0:00:07.006) 0:02:27.809 ********* 2025-08-29 17:45:46.811989 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:46.811994 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:46.811999 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.812004 | orchestrator | 2025-08-29 17:45:46.812009 | orchestrator | RUNNING HANDLER [designate : Restart designate-mdns container] ***************** 2025-08-29 17:45:46.812014 | orchestrator | Friday 29 August 2025 17:45:15 +0000 (0:00:08.670) 0:02:36.480 ********* 2025-08-29 17:45:46.812019 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:46.812024 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.812030 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:46.812036 | orchestrator | 2025-08-29 17:45:46.812041 | orchestrator | RUNNING HANDLER [designate : Restart designate-worker container] *************** 2025-08-29 17:45:46.812050 | orchestrator | Friday 29 August 2025 17:45:27 +0000 (0:00:11.994) 0:02:48.474 ********* 2025-08-29 17:45:46.812056 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.812062 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:45:46.812068 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:45:46.812073 | orchestrator | 2025-08-29 17:45:46.812079 | orchestrator | TASK [designate : Non-destructive DNS pools update] **************************** 2025-08-29 17:45:46.812085 | orchestrator | Friday 29 August 2025 17:45:35 +0000 (0:00:08.059) 0:02:56.533 ********* 2025-08-29 17:45:46.812091 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:45:46.812097 | orchestrator | 2025-08-29 17:45:46.812102 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:45:46.812108 | orchestrator | testbed-node-0 : ok=29  changed=24  unreachable=0 failed=0 skipped=7  rescued=0 ignored=0 2025-08-29 17:45:46.812116 | orchestrator | testbed-node-1 : ok=19  changed=15  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:45:46.812122 | orchestrator | testbed-node-2 : ok=19  changed=15  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:45:46.812128 | orchestrator | 2025-08-29 17:45:46.812134 | orchestrator | 2025-08-29 17:45:46.812139 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:45:46.812145 | orchestrator | Friday 29 August 2025 17:45:43 +0000 (0:00:08.335) 0:03:04.869 ********* 2025-08-29 17:45:46.812151 | orchestrator | =============================================================================== 2025-08-29 17:45:46.812160 | orchestrator | designate : Copying over designate.conf -------------------------------- 22.89s 2025-08-29 17:45:46.812165 | orchestrator | designate : Running Designate bootstrap container ---------------------- 15.19s 2025-08-29 17:45:46.812171 | orchestrator | designate : Restart designate-backend-bind9 container ------------------ 13.19s 2025-08-29 17:45:46.812177 | orchestrator | designate : Restart designate-mdns container --------------------------- 11.99s 2025-08-29 17:45:46.812183 | orchestrator | designate : Restart designate-api container ----------------------------- 9.76s 2025-08-29 17:45:46.812188 | orchestrator | designate : Restart designate-producer container ------------------------ 8.67s 2025-08-29 17:45:46.812199 | orchestrator | designate : Non-destructive DNS pools update ---------------------------- 8.34s 2025-08-29 17:45:46.812205 | orchestrator | designate : Restart designate-worker container -------------------------- 8.06s 2025-08-29 17:45:46.812210 | orchestrator | designate : Copying over pools.yaml ------------------------------------- 7.96s 2025-08-29 17:45:46.812216 | orchestrator | designate : Restart designate-central container ------------------------- 7.01s 2025-08-29 17:45:46.812222 | orchestrator | service-ks-register : designate | Creating endpoints -------------------- 6.66s 2025-08-29 17:45:46.812228 | orchestrator | service-cert-copy : designate | Copying over extra CA certificates ------ 6.61s 2025-08-29 17:45:46.812233 | orchestrator | designate : Copying over config.json files for services ----------------- 6.26s 2025-08-29 17:45:46.812239 | orchestrator | designate : Copying over named.conf ------------------------------------- 4.90s 2025-08-29 17:45:46.812245 | orchestrator | service-ks-register : designate | Creating users ------------------------ 4.41s 2025-08-29 17:45:46.812250 | orchestrator | designate : Check designate containers ---------------------------------- 4.35s 2025-08-29 17:45:46.812256 | orchestrator | service-ks-register : designate | Granting user roles ------------------- 4.26s 2025-08-29 17:45:46.812262 | orchestrator | designate : Copying over rndc.key --------------------------------------- 3.84s 2025-08-29 17:45:46.812268 | orchestrator | service-ks-register : designate | Creating services --------------------- 3.80s 2025-08-29 17:45:46.812273 | orchestrator | service-ks-register : designate | Creating roles ------------------------ 3.70s 2025-08-29 17:45:46.812279 | orchestrator | 2025-08-29 17:45:46 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:46.812286 | orchestrator | 2025-08-29 17:45:46 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:46.812292 | orchestrator | 2025-08-29 17:45:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:49.839992 | orchestrator | 2025-08-29 17:45:49 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:49.840126 | orchestrator | 2025-08-29 17:45:49 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:45:49.840141 | orchestrator | 2025-08-29 17:45:49 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:49.840152 | orchestrator | 2025-08-29 17:45:49 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:49.840164 | orchestrator | 2025-08-29 17:45:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:52.866008 | orchestrator | 2025-08-29 17:45:52 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:52.866170 | orchestrator | 2025-08-29 17:45:52 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:45:52.866185 | orchestrator | 2025-08-29 17:45:52 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:52.866198 | orchestrator | 2025-08-29 17:45:52 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:52.866209 | orchestrator | 2025-08-29 17:45:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:55.902136 | orchestrator | 2025-08-29 17:45:55 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:55.902244 | orchestrator | 2025-08-29 17:45:55 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:45:55.902264 | orchestrator | 2025-08-29 17:45:55 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:55.902280 | orchestrator | 2025-08-29 17:45:55 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:55.902294 | orchestrator | 2025-08-29 17:45:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:45:58.926131 | orchestrator | 2025-08-29 17:45:58 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:45:58.926409 | orchestrator | 2025-08-29 17:45:58 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:45:58.927494 | orchestrator | 2025-08-29 17:45:58 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:45:58.928658 | orchestrator | 2025-08-29 17:45:58 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:45:58.928699 | orchestrator | 2025-08-29 17:45:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:01.954695 | orchestrator | 2025-08-29 17:46:01 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:01.956374 | orchestrator | 2025-08-29 17:46:01 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:01.957997 | orchestrator | 2025-08-29 17:46:01 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:01.959507 | orchestrator | 2025-08-29 17:46:01 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:01.959713 | orchestrator | 2025-08-29 17:46:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:04.997240 | orchestrator | 2025-08-29 17:46:04 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:04.997404 | orchestrator | 2025-08-29 17:46:04 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:04.999338 | orchestrator | 2025-08-29 17:46:04 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:05.001633 | orchestrator | 2025-08-29 17:46:04 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:05.001734 | orchestrator | 2025-08-29 17:46:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:08.049173 | orchestrator | 2025-08-29 17:46:08 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:08.050197 | orchestrator | 2025-08-29 17:46:08 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:08.051606 | orchestrator | 2025-08-29 17:46:08 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:08.053123 | orchestrator | 2025-08-29 17:46:08 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:08.053156 | orchestrator | 2025-08-29 17:46:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:11.091896 | orchestrator | 2025-08-29 17:46:11 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:11.092868 | orchestrator | 2025-08-29 17:46:11 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:11.094650 | orchestrator | 2025-08-29 17:46:11 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:11.096436 | orchestrator | 2025-08-29 17:46:11 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:11.096465 | orchestrator | 2025-08-29 17:46:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:14.151716 | orchestrator | 2025-08-29 17:46:14 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:14.153089 | orchestrator | 2025-08-29 17:46:14 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:14.155266 | orchestrator | 2025-08-29 17:46:14 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:14.156784 | orchestrator | 2025-08-29 17:46:14 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:14.156821 | orchestrator | 2025-08-29 17:46:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:17.202742 | orchestrator | 2025-08-29 17:46:17 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:17.204662 | orchestrator | 2025-08-29 17:46:17 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:17.206309 | orchestrator | 2025-08-29 17:46:17 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:17.208147 | orchestrator | 2025-08-29 17:46:17 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:17.208168 | orchestrator | 2025-08-29 17:46:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:20.260700 | orchestrator | 2025-08-29 17:46:20 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:20.262393 | orchestrator | 2025-08-29 17:46:20 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:20.265430 | orchestrator | 2025-08-29 17:46:20 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:20.268028 | orchestrator | 2025-08-29 17:46:20 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:20.268657 | orchestrator | 2025-08-29 17:46:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:23.307183 | orchestrator | 2025-08-29 17:46:23 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:23.307299 | orchestrator | 2025-08-29 17:46:23 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:23.308135 | orchestrator | 2025-08-29 17:46:23 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:23.308776 | orchestrator | 2025-08-29 17:46:23 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:23.308801 | orchestrator | 2025-08-29 17:46:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:26.341033 | orchestrator | 2025-08-29 17:46:26 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:26.341142 | orchestrator | 2025-08-29 17:46:26 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state STARTED 2025-08-29 17:46:26.342939 | orchestrator | 2025-08-29 17:46:26 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:26.343729 | orchestrator | 2025-08-29 17:46:26 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:26.343755 | orchestrator | 2025-08-29 17:46:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:29.375361 | orchestrator | 2025-08-29 17:46:29 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:29.375433 | orchestrator | 2025-08-29 17:46:29 | INFO  | Task a0277c31-b24a-40e3-9565-f06c2a7ff409 is in state SUCCESS 2025-08-29 17:46:29.375753 | orchestrator | 2025-08-29 17:46:29 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:29.377870 | orchestrator | 2025-08-29 17:46:29 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:29.377910 | orchestrator | 2025-08-29 17:46:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:32.402413 | orchestrator | 2025-08-29 17:46:32 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:32.402568 | orchestrator | 2025-08-29 17:46:32 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:32.403538 | orchestrator | 2025-08-29 17:46:32 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:32.404070 | orchestrator | 2025-08-29 17:46:32 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:32.404165 | orchestrator | 2025-08-29 17:46:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:35.495469 | orchestrator | 2025-08-29 17:46:35 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:35.499183 | orchestrator | 2025-08-29 17:46:35 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:35.501062 | orchestrator | 2025-08-29 17:46:35 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:35.503100 | orchestrator | 2025-08-29 17:46:35 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:35.503315 | orchestrator | 2025-08-29 17:46:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:38.550133 | orchestrator | 2025-08-29 17:46:38 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:38.551573 | orchestrator | 2025-08-29 17:46:38 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:38.552288 | orchestrator | 2025-08-29 17:46:38 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:38.553522 | orchestrator | 2025-08-29 17:46:38 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:38.553585 | orchestrator | 2025-08-29 17:46:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:41.584710 | orchestrator | 2025-08-29 17:46:41 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:41.586596 | orchestrator | 2025-08-29 17:46:41 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:41.587928 | orchestrator | 2025-08-29 17:46:41 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:41.589319 | orchestrator | 2025-08-29 17:46:41 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:41.589661 | orchestrator | 2025-08-29 17:46:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:44.630785 | orchestrator | 2025-08-29 17:46:44 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:44.631301 | orchestrator | 2025-08-29 17:46:44 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:44.632507 | orchestrator | 2025-08-29 17:46:44 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:44.634125 | orchestrator | 2025-08-29 17:46:44 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:44.634366 | orchestrator | 2025-08-29 17:46:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:47.673445 | orchestrator | 2025-08-29 17:46:47 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:47.675538 | orchestrator | 2025-08-29 17:46:47 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:47.677918 | orchestrator | 2025-08-29 17:46:47 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:47.680028 | orchestrator | 2025-08-29 17:46:47 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state STARTED 2025-08-29 17:46:47.680058 | orchestrator | 2025-08-29 17:46:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:50.723450 | orchestrator | 2025-08-29 17:46:50 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:50.724402 | orchestrator | 2025-08-29 17:46:50 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:50.726145 | orchestrator | 2025-08-29 17:46:50 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:46:50.727287 | orchestrator | 2025-08-29 17:46:50 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:50.729625 | orchestrator | 2025-08-29 17:46:50 | INFO  | Task 2a9deabe-c9cc-4144-826b-f9cc2ce3af2e is in state SUCCESS 2025-08-29 17:46:50.730249 | orchestrator | 2025-08-29 17:46:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:50.731867 | orchestrator | 2025-08-29 17:46:50.731901 | orchestrator | 2025-08-29 17:46:50.731913 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:46:50.731924 | orchestrator | 2025-08-29 17:46:50.731936 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:46:50.731949 | orchestrator | Friday 29 August 2025 17:45:54 +0000 (0:00:00.627) 0:00:00.627 ********* 2025-08-29 17:46:50.731960 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:46:50.731972 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:46:50.731983 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:46:50.731994 | orchestrator | ok: [testbed-manager] 2025-08-29 17:46:50.732005 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:46:50.732015 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:46:50.732117 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:46:50.732356 | orchestrator | 2025-08-29 17:46:50.732368 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:46:50.732379 | orchestrator | Friday 29 August 2025 17:45:55 +0000 (0:00:01.648) 0:00:02.276 ********* 2025-08-29 17:46:50.732390 | orchestrator | ok: [testbed-node-0] => (item=enable_ceph_rgw_True) 2025-08-29 17:46:50.732401 | orchestrator | ok: [testbed-node-1] => (item=enable_ceph_rgw_True) 2025-08-29 17:46:50.732412 | orchestrator | ok: [testbed-node-2] => (item=enable_ceph_rgw_True) 2025-08-29 17:46:50.732423 | orchestrator | ok: [testbed-manager] => (item=enable_ceph_rgw_True) 2025-08-29 17:46:50.732434 | orchestrator | ok: [testbed-node-3] => (item=enable_ceph_rgw_True) 2025-08-29 17:46:50.732444 | orchestrator | ok: [testbed-node-4] => (item=enable_ceph_rgw_True) 2025-08-29 17:46:50.732455 | orchestrator | ok: [testbed-node-5] => (item=enable_ceph_rgw_True) 2025-08-29 17:46:50.732466 | orchestrator | 2025-08-29 17:46:50.732501 | orchestrator | PLAY [Apply role ceph-rgw] ***************************************************** 2025-08-29 17:46:50.732512 | orchestrator | 2025-08-29 17:46:50.732523 | orchestrator | TASK [ceph-rgw : include_tasks] ************************************************ 2025-08-29 17:46:50.732534 | orchestrator | Friday 29 August 2025 17:45:56 +0000 (0:00:00.862) 0:00:03.139 ********* 2025-08-29 17:46:50.732546 | orchestrator | included: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-manager, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:46:50.732558 | orchestrator | 2025-08-29 17:46:50.732569 | orchestrator | TASK [service-ks-register : ceph-rgw | Creating services] ********************** 2025-08-29 17:46:50.732580 | orchestrator | Friday 29 August 2025 17:45:58 +0000 (0:00:02.095) 0:00:05.234 ********* 2025-08-29 17:46:50.732591 | orchestrator | changed: [testbed-node-0] => (item=swift (object-store)) 2025-08-29 17:46:50.732602 | orchestrator | 2025-08-29 17:46:50.732613 | orchestrator | TASK [service-ks-register : ceph-rgw | Creating endpoints] ********************* 2025-08-29 17:46:50.732624 | orchestrator | Friday 29 August 2025 17:46:02 +0000 (0:00:03.315) 0:00:08.550 ********* 2025-08-29 17:46:50.732636 | orchestrator | changed: [testbed-node-0] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal) 2025-08-29 17:46:50.732649 | orchestrator | changed: [testbed-node-0] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public) 2025-08-29 17:46:50.732689 | orchestrator | 2025-08-29 17:46:50.732701 | orchestrator | TASK [service-ks-register : ceph-rgw | Creating projects] ********************** 2025-08-29 17:46:50.732724 | orchestrator | Friday 29 August 2025 17:46:08 +0000 (0:00:06.408) 0:00:14.959 ********* 2025-08-29 17:46:50.732735 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:46:50.732746 | orchestrator | 2025-08-29 17:46:50.732757 | orchestrator | TASK [service-ks-register : ceph-rgw | Creating users] ************************* 2025-08-29 17:46:50.732768 | orchestrator | Friday 29 August 2025 17:46:11 +0000 (0:00:03.175) 0:00:18.135 ********* 2025-08-29 17:46:50.732779 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:46:50.732789 | orchestrator | changed: [testbed-node-0] => (item=ceph_rgw -> service) 2025-08-29 17:46:50.732800 | orchestrator | 2025-08-29 17:46:50.732811 | orchestrator | TASK [service-ks-register : ceph-rgw | Creating roles] ************************* 2025-08-29 17:46:50.732821 | orchestrator | Friday 29 August 2025 17:46:15 +0000 (0:00:03.902) 0:00:22.037 ********* 2025-08-29 17:46:50.732832 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:46:50.732843 | orchestrator | changed: [testbed-node-0] => (item=ResellerAdmin) 2025-08-29 17:46:50.732853 | orchestrator | 2025-08-29 17:46:50.732864 | orchestrator | TASK [service-ks-register : ceph-rgw | Granting user roles] ******************** 2025-08-29 17:46:50.732875 | orchestrator | Friday 29 August 2025 17:46:21 +0000 (0:00:06.287) 0:00:28.325 ********* 2025-08-29 17:46:50.732886 | orchestrator | changed: [testbed-node-0] => (item=ceph_rgw -> service -> admin) 2025-08-29 17:46:50.732896 | orchestrator | 2025-08-29 17:46:50.732907 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:46:50.732918 | orchestrator | testbed-manager : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:46:50.732929 | orchestrator | testbed-node-0 : ok=9  changed=5  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:46:50.732940 | orchestrator | testbed-node-1 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:46:50.732951 | orchestrator | testbed-node-2 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:46:50.732962 | orchestrator | testbed-node-3 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:46:50.732983 | orchestrator | testbed-node-4 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:46:50.732995 | orchestrator | testbed-node-5 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:46:50.733006 | orchestrator | 2025-08-29 17:46:50.733016 | orchestrator | 2025-08-29 17:46:50.733027 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:46:50.733038 | orchestrator | Friday 29 August 2025 17:46:27 +0000 (0:00:05.983) 0:00:34.309 ********* 2025-08-29 17:46:50.733049 | orchestrator | =============================================================================== 2025-08-29 17:46:50.733060 | orchestrator | service-ks-register : ceph-rgw | Creating endpoints --------------------- 6.41s 2025-08-29 17:46:50.733070 | orchestrator | service-ks-register : ceph-rgw | Creating roles ------------------------- 6.29s 2025-08-29 17:46:50.733081 | orchestrator | service-ks-register : ceph-rgw | Granting user roles -------------------- 5.98s 2025-08-29 17:46:50.733092 | orchestrator | service-ks-register : ceph-rgw | Creating users ------------------------- 3.90s 2025-08-29 17:46:50.733102 | orchestrator | service-ks-register : ceph-rgw | Creating services ---------------------- 3.32s 2025-08-29 17:46:50.733113 | orchestrator | service-ks-register : ceph-rgw | Creating projects ---------------------- 3.18s 2025-08-29 17:46:50.733124 | orchestrator | ceph-rgw : include_tasks ------------------------------------------------ 2.10s 2025-08-29 17:46:50.733153 | orchestrator | Group hosts based on Kolla action --------------------------------------- 1.65s 2025-08-29 17:46:50.733164 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.86s 2025-08-29 17:46:50.733175 | orchestrator | 2025-08-29 17:46:50.733186 | orchestrator | 2025-08-29 17:46:50.733197 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:46:50.733208 | orchestrator | 2025-08-29 17:46:50.733218 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:46:50.733229 | orchestrator | Friday 29 August 2025 17:44:49 +0000 (0:00:00.509) 0:00:00.509 ********* 2025-08-29 17:46:50.733240 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:46:50.733251 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:46:50.733261 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:46:50.733272 | orchestrator | 2025-08-29 17:46:50.733283 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:46:50.733294 | orchestrator | Friday 29 August 2025 17:44:50 +0000 (0:00:00.494) 0:00:01.004 ********* 2025-08-29 17:46:50.733305 | orchestrator | ok: [testbed-node-0] => (item=enable_magnum_True) 2025-08-29 17:46:50.733316 | orchestrator | ok: [testbed-node-1] => (item=enable_magnum_True) 2025-08-29 17:46:50.733326 | orchestrator | ok: [testbed-node-2] => (item=enable_magnum_True) 2025-08-29 17:46:50.733337 | orchestrator | 2025-08-29 17:46:50.733348 | orchestrator | PLAY [Apply role magnum] ******************************************************* 2025-08-29 17:46:50.733359 | orchestrator | 2025-08-29 17:46:50.733369 | orchestrator | TASK [magnum : include_tasks] ************************************************** 2025-08-29 17:46:50.733380 | orchestrator | Friday 29 August 2025 17:44:50 +0000 (0:00:00.644) 0:00:01.649 ********* 2025-08-29 17:46:50.733391 | orchestrator | included: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:46:50.733402 | orchestrator | 2025-08-29 17:46:50.733412 | orchestrator | TASK [service-ks-register : magnum | Creating services] ************************ 2025-08-29 17:46:50.733428 | orchestrator | Friday 29 August 2025 17:44:52 +0000 (0:00:01.288) 0:00:02.937 ********* 2025-08-29 17:46:50.733439 | orchestrator | changed: [testbed-node-0] => (item=magnum (container-infra)) 2025-08-29 17:46:50.733450 | orchestrator | 2025-08-29 17:46:50.733461 | orchestrator | TASK [service-ks-register : magnum | Creating endpoints] *********************** 2025-08-29 17:46:50.733511 | orchestrator | Friday 29 August 2025 17:44:56 +0000 (0:00:04.007) 0:00:06.945 ********* 2025-08-29 17:46:50.733524 | orchestrator | changed: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal) 2025-08-29 17:46:50.733535 | orchestrator | changed: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public) 2025-08-29 17:46:50.733545 | orchestrator | 2025-08-29 17:46:50.733556 | orchestrator | TASK [service-ks-register : magnum | Creating projects] ************************ 2025-08-29 17:46:50.733567 | orchestrator | Friday 29 August 2025 17:45:03 +0000 (0:00:06.957) 0:00:13.902 ********* 2025-08-29 17:46:50.733577 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:46:50.733588 | orchestrator | 2025-08-29 17:46:50.733599 | orchestrator | TASK [service-ks-register : magnum | Creating users] *************************** 2025-08-29 17:46:50.733609 | orchestrator | Friday 29 August 2025 17:45:06 +0000 (0:00:03.255) 0:00:17.157 ********* 2025-08-29 17:46:50.733620 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:46:50.733631 | orchestrator | changed: [testbed-node-0] => (item=magnum -> service) 2025-08-29 17:46:50.733641 | orchestrator | 2025-08-29 17:46:50.733652 | orchestrator | TASK [service-ks-register : magnum | Creating roles] *************************** 2025-08-29 17:46:50.733663 | orchestrator | Friday 29 August 2025 17:45:10 +0000 (0:00:04.042) 0:00:21.200 ********* 2025-08-29 17:46:50.733674 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:46:50.733684 | orchestrator | 2025-08-29 17:46:50.733695 | orchestrator | TASK [service-ks-register : magnum | Granting user roles] ********************** 2025-08-29 17:46:50.733706 | orchestrator | Friday 29 August 2025 17:45:13 +0000 (0:00:03.444) 0:00:24.645 ********* 2025-08-29 17:46:50.733724 | orchestrator | changed: [testbed-node-0] => (item=magnum -> service -> admin) 2025-08-29 17:46:50.733735 | orchestrator | 2025-08-29 17:46:50.733746 | orchestrator | TASK [magnum : Creating Magnum trustee domain] ********************************* 2025-08-29 17:46:50.733756 | orchestrator | Friday 29 August 2025 17:45:17 +0000 (0:00:04.151) 0:00:28.796 ********* 2025-08-29 17:46:50.733767 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.733778 | orchestrator | 2025-08-29 17:46:50.733789 | orchestrator | TASK [magnum : Creating Magnum trustee user] *********************************** 2025-08-29 17:46:50.733808 | orchestrator | Friday 29 August 2025 17:45:21 +0000 (0:00:03.747) 0:00:32.544 ********* 2025-08-29 17:46:50.733819 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.733830 | orchestrator | 2025-08-29 17:46:50.733840 | orchestrator | TASK [magnum : Creating Magnum trustee user role] ****************************** 2025-08-29 17:46:50.733851 | orchestrator | Friday 29 August 2025 17:45:25 +0000 (0:00:04.264) 0:00:36.809 ********* 2025-08-29 17:46:50.733862 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.733873 | orchestrator | 2025-08-29 17:46:50.733884 | orchestrator | TASK [magnum : Ensuring config directories exist] ****************************** 2025-08-29 17:46:50.733895 | orchestrator | Friday 29 August 2025 17:45:29 +0000 (0:00:03.935) 0:00:40.744 ********* 2025-08-29 17:46:50.733909 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.733926 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.733943 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.733963 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.733983 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.733995 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.734007 | orchestrator | 2025-08-29 17:46:50.734065 | orchestrator | TASK [magnum : Check if policies shall be overwritten] ************************* 2025-08-29 17:46:50.734079 | orchestrator | Friday 29 August 2025 17:45:32 +0000 (0:00:02.193) 0:00:42.937 ********* 2025-08-29 17:46:50.734090 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:46:50.734101 | orchestrator | 2025-08-29 17:46:50.734112 | orchestrator | TASK [magnum : Set magnum policy file] ***************************************** 2025-08-29 17:46:50.734123 | orchestrator | Friday 29 August 2025 17:45:32 +0000 (0:00:00.111) 0:00:43.049 ********* 2025-08-29 17:46:50.734134 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:46:50.734144 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:46:50.734155 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:46:50.734165 | orchestrator | 2025-08-29 17:46:50.734176 | orchestrator | TASK [magnum : Check if kubeconfig file is supplied] *************************** 2025-08-29 17:46:50.734187 | orchestrator | Friday 29 August 2025 17:45:32 +0000 (0:00:00.712) 0:00:43.761 ********* 2025-08-29 17:46:50.734197 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:46:50.734208 | orchestrator | 2025-08-29 17:46:50.734219 | orchestrator | TASK [magnum : Copying over kubeconfig file] *********************************** 2025-08-29 17:46:50.734229 | orchestrator | Friday 29 August 2025 17:45:34 +0000 (0:00:01.126) 0:00:44.888 ********* 2025-08-29 17:46:50.734246 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.734266 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.734286 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.734299 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.734367 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.734381 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.734400 | orchestrator | 2025-08-29 17:46:50.734411 | orchestrator | TASK [magnum : Set magnum kubeconfig file's path] ****************************** 2025-08-29 17:46:50.734422 | orchestrator | Friday 29 August 2025 17:45:36 +0000 (0:00:02.579) 0:00:47.467 ********* 2025-08-29 17:46:50.734433 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:46:50.734443 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:46:50.734454 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:46:50.734465 | orchestrator | 2025-08-29 17:46:50.734522 | orchestrator | TASK [magnum : include_tasks] ************************************************** 2025-08-29 17:46:50.734535 | orchestrator | Friday 29 August 2025 17:45:36 +0000 (0:00:00.298) 0:00:47.766 ********* 2025-08-29 17:46:50.734546 | orchestrator | included: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:46:50.734557 | orchestrator | 2025-08-29 17:46:50.734568 | orchestrator | TASK [service-cert-copy : magnum | Copying over extra CA certificates] ********* 2025-08-29 17:46:50.734578 | orchestrator | Friday 29 August 2025 17:45:37 +0000 (0:00:00.566) 0:00:48.332 ********* 2025-08-29 17:46:50.734598 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.734611 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.734629 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.734648 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.734659 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.734678 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.734689 | orchestrator | 2025-08-29 17:46:50.734700 | orchestrator | TASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] *** 2025-08-29 17:46:50.734711 | orchestrator | Friday 29 August 2025 17:45:39 +0000 (0:00:02.477) 0:00:50.810 ********* 2025-08-29 17:46:50.734723 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.734746 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.734757 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:46:50.734769 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.734786 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.734797 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:46:50.734809 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.734820 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.734837 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:46:50.734848 | orchestrator | 2025-08-29 17:46:50.734859 | orchestrator | TASK [service-cert-copy : magnum | Copying over backend internal TLS key] ****** 2025-08-29 17:46:50.734870 | orchestrator | Friday 29 August 2025 17:45:41 +0000 (0:00:01.785) 0:00:52.595 ********* 2025-08-29 17:46:50.734887 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.734899 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.734910 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:46:50.734930 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.734942 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.734963 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:46:50.734975 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.734991 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.735002 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:46:50.735013 | orchestrator | 2025-08-29 17:46:50.735024 | orchestrator | TASK [magnum : Copying over config.json files for services] ******************** 2025-08-29 17:46:50.735035 | orchestrator | Friday 29 August 2025 17:45:45 +0000 (0:00:03.246) 0:00:55.842 ********* 2025-08-29 17:46:50.735046 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735065 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735077 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735101 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735113 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735125 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735136 | orchestrator | 2025-08-29 17:46:50.735152 | orchestrator | TASK [magnum : Copying over magnum.conf] *************************************** 2025-08-29 17:46:50.735163 | orchestrator | Friday 29 August 2025 17:45:48 +0000 (0:00:03.301) 0:00:59.143 ********* 2025-08-29 17:46:50.735175 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735192 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735208 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735220 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735239 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735250 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735286 | orchestrator | 2025-08-29 17:46:50.735298 | orchestrator | TASK [magnum : Copying over existing policy file] ****************************** 2025-08-29 17:46:50.735309 | orchestrator | Friday 29 August 2025 17:45:56 +0000 (0:00:08.549) 0:01:07.693 ********* 2025-08-29 17:46:50.735321 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.735338 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.735350 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.735361 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:46:50.735380 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.735398 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:46:50.735410 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})  2025-08-29 17:46:50.735422 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:46:50.735433 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:46:50.735444 | orchestrator | 2025-08-29 17:46:50.735459 | orchestrator | TASK [magnum : Check magnum containers] **************************************** 2025-08-29 17:46:50.735471 | orchestrator | Friday 29 August 2025 17:45:58 +0000 (0:00:01.160) 0:01:08.853 ********* 2025-08-29 17:46:50.735502 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735520 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735539 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) 2025-08-29 17:46:50.735550 | orchestrator | changed: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735567 | orchestrator | changed: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735578 | orchestrator | changed: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:46:50.735589 | orchestrator | 2025-08-29 17:46:50.735601 | orchestrator | TASK [magnum : include_tasks] ************************************************** 2025-08-29 17:46:50.735612 | orchestrator | Friday 29 August 2025 17:46:00 +0000 (0:00:02.510) 0:01:11.364 ********* 2025-08-29 17:46:50.735623 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:46:50.735633 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:46:50.735644 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:46:50.735655 | orchestrator | 2025-08-29 17:46:50.735666 | orchestrator | TASK [magnum : Creating Magnum database] *************************************** 2025-08-29 17:46:50.735677 | orchestrator | Friday 29 August 2025 17:46:00 +0000 (0:00:00.247) 0:01:11.611 ********* 2025-08-29 17:46:50.735687 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.735698 | orchestrator | 2025-08-29 17:46:50.735716 | orchestrator | TASK [magnum : Creating Magnum database user and setting permissions] ********** 2025-08-29 17:46:50.735727 | orchestrator | Friday 29 August 2025 17:46:02 +0000 (0:00:02.186) 0:01:13.797 ********* 2025-08-29 17:46:50.735738 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.735748 | orchestrator | 2025-08-29 17:46:50.735759 | orchestrator | TASK [magnum : Running Magnum bootstrap container] ***************************** 2025-08-29 17:46:50.735775 | orchestrator | Friday 29 August 2025 17:46:05 +0000 (0:00:02.087) 0:01:15.885 ********* 2025-08-29 17:46:50.735787 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.735797 | orchestrator | 2025-08-29 17:46:50.735808 | orchestrator | TASK [magnum : Flush handlers] ************************************************* 2025-08-29 17:46:50.735819 | orchestrator | Friday 29 August 2025 17:46:20 +0000 (0:00:15.554) 0:01:31.439 ********* 2025-08-29 17:46:50.735830 | orchestrator | 2025-08-29 17:46:50.735840 | orchestrator | TASK [magnum : Flush handlers] ************************************************* 2025-08-29 17:46:50.735851 | orchestrator | Friday 29 August 2025 17:46:20 +0000 (0:00:00.066) 0:01:31.506 ********* 2025-08-29 17:46:50.735862 | orchestrator | 2025-08-29 17:46:50.735873 | orchestrator | TASK [magnum : Flush handlers] ************************************************* 2025-08-29 17:46:50.735883 | orchestrator | Friday 29 August 2025 17:46:20 +0000 (0:00:00.064) 0:01:31.570 ********* 2025-08-29 17:46:50.735894 | orchestrator | 2025-08-29 17:46:50.735905 | orchestrator | RUNNING HANDLER [magnum : Restart magnum-api container] ************************ 2025-08-29 17:46:50.735915 | orchestrator | Friday 29 August 2025 17:46:20 +0000 (0:00:00.066) 0:01:31.637 ********* 2025-08-29 17:46:50.735926 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.735937 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:46:50.735948 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:46:50.735958 | orchestrator | 2025-08-29 17:46:50.735969 | orchestrator | RUNNING HANDLER [magnum : Restart magnum-conductor container] ****************** 2025-08-29 17:46:50.735980 | orchestrator | Friday 29 August 2025 17:46:37 +0000 (0:00:16.499) 0:01:48.136 ********* 2025-08-29 17:46:50.735991 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:46:50.736001 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:46:50.736012 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:46:50.736023 | orchestrator | 2025-08-29 17:46:50.736034 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:46:50.736044 | orchestrator | testbed-node-0 : ok=26  changed=18  unreachable=0 failed=0 skipped=6  rescued=0 ignored=0 2025-08-29 17:46:50.736056 | orchestrator | testbed-node-1 : ok=13  changed=8  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:46:50.736067 | orchestrator | testbed-node-2 : ok=13  changed=8  unreachable=0 failed=0 skipped=5  rescued=0 ignored=0 2025-08-29 17:46:50.736078 | orchestrator | 2025-08-29 17:46:50.736088 | orchestrator | 2025-08-29 17:46:50.736099 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:46:50.736110 | orchestrator | Friday 29 August 2025 17:46:48 +0000 (0:00:11.559) 0:01:59.696 ********* 2025-08-29 17:46:50.736121 | orchestrator | =============================================================================== 2025-08-29 17:46:50.736131 | orchestrator | magnum : Restart magnum-api container ---------------------------------- 16.50s 2025-08-29 17:46:50.736142 | orchestrator | magnum : Running Magnum bootstrap container ---------------------------- 15.55s 2025-08-29 17:46:50.736153 | orchestrator | magnum : Restart magnum-conductor container ---------------------------- 11.56s 2025-08-29 17:46:50.736163 | orchestrator | magnum : Copying over magnum.conf --------------------------------------- 8.55s 2025-08-29 17:46:50.736179 | orchestrator | service-ks-register : magnum | Creating endpoints ----------------------- 6.96s 2025-08-29 17:46:50.736190 | orchestrator | magnum : Creating Magnum trustee user ----------------------------------- 4.26s 2025-08-29 17:46:50.736200 | orchestrator | service-ks-register : magnum | Granting user roles ---------------------- 4.15s 2025-08-29 17:46:50.736218 | orchestrator | service-ks-register : magnum | Creating users --------------------------- 4.04s 2025-08-29 17:46:50.736229 | orchestrator | service-ks-register : magnum | Creating services ------------------------ 4.01s 2025-08-29 17:46:50.736239 | orchestrator | magnum : Creating Magnum trustee user role ------------------------------ 3.94s 2025-08-29 17:46:50.736250 | orchestrator | magnum : Creating Magnum trustee domain --------------------------------- 3.75s 2025-08-29 17:46:50.736261 | orchestrator | service-ks-register : magnum | Creating roles --------------------------- 3.44s 2025-08-29 17:46:50.736272 | orchestrator | magnum : Copying over config.json files for services -------------------- 3.30s 2025-08-29 17:46:50.736283 | orchestrator | service-ks-register : magnum | Creating projects ------------------------ 3.26s 2025-08-29 17:46:50.736293 | orchestrator | service-cert-copy : magnum | Copying over backend internal TLS key ------ 3.25s 2025-08-29 17:46:50.736304 | orchestrator | magnum : Copying over kubeconfig file ----------------------------------- 2.58s 2025-08-29 17:46:50.736315 | orchestrator | magnum : Check magnum containers ---------------------------------------- 2.51s 2025-08-29 17:46:50.736326 | orchestrator | service-cert-copy : magnum | Copying over extra CA certificates --------- 2.48s 2025-08-29 17:46:50.736336 | orchestrator | magnum : Ensuring config directories exist ------------------------------ 2.19s 2025-08-29 17:46:50.736347 | orchestrator | magnum : Creating Magnum database --------------------------------------- 2.19s 2025-08-29 17:46:53.783186 | orchestrator | 2025-08-29 17:46:53 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:53.783917 | orchestrator | 2025-08-29 17:46:53 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:53.785615 | orchestrator | 2025-08-29 17:46:53 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:46:53.786627 | orchestrator | 2025-08-29 17:46:53 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:53.786658 | orchestrator | 2025-08-29 17:46:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:56.826828 | orchestrator | 2025-08-29 17:46:56 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:56.826944 | orchestrator | 2025-08-29 17:46:56 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:56.828152 | orchestrator | 2025-08-29 17:46:56 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:46:56.829137 | orchestrator | 2025-08-29 17:46:56 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:56.829392 | orchestrator | 2025-08-29 17:46:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:46:59.872790 | orchestrator | 2025-08-29 17:46:59 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:46:59.873416 | orchestrator | 2025-08-29 17:46:59 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:46:59.875116 | orchestrator | 2025-08-29 17:46:59 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:46:59.877071 | orchestrator | 2025-08-29 17:46:59 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:46:59.877104 | orchestrator | 2025-08-29 17:46:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:02.916376 | orchestrator | 2025-08-29 17:47:02 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:47:02.916992 | orchestrator | 2025-08-29 17:47:02 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:02.918880 | orchestrator | 2025-08-29 17:47:02 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:02.919608 | orchestrator | 2025-08-29 17:47:02 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:02.919774 | orchestrator | 2025-08-29 17:47:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:05.971131 | orchestrator | 2025-08-29 17:47:05 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:47:05.971391 | orchestrator | 2025-08-29 17:47:05 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:05.972188 | orchestrator | 2025-08-29 17:47:05 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:05.972704 | orchestrator | 2025-08-29 17:47:05 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:05.972770 | orchestrator | 2025-08-29 17:47:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:09.022180 | orchestrator | 2025-08-29 17:47:09 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:47:09.024775 | orchestrator | 2025-08-29 17:47:09 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:09.028648 | orchestrator | 2025-08-29 17:47:09 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:09.032185 | orchestrator | 2025-08-29 17:47:09 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:09.032547 | orchestrator | 2025-08-29 17:47:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:12.097966 | orchestrator | 2025-08-29 17:47:12 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:47:12.099414 | orchestrator | 2025-08-29 17:47:12 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:12.100300 | orchestrator | 2025-08-29 17:47:12 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:12.101154 | orchestrator | 2025-08-29 17:47:12 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:12.101198 | orchestrator | 2025-08-29 17:47:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:15.136601 | orchestrator | 2025-08-29 17:47:15 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:47:15.138995 | orchestrator | 2025-08-29 17:47:15 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:15.140981 | orchestrator | 2025-08-29 17:47:15 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:15.142859 | orchestrator | 2025-08-29 17:47:15 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:15.143047 | orchestrator | 2025-08-29 17:47:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:18.175894 | orchestrator | 2025-08-29 17:47:18 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:47:18.175981 | orchestrator | 2025-08-29 17:47:18 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:18.176711 | orchestrator | 2025-08-29 17:47:18 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:18.177637 | orchestrator | 2025-08-29 17:47:18 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:18.177663 | orchestrator | 2025-08-29 17:47:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:21.208045 | orchestrator | 2025-08-29 17:47:21 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state STARTED 2025-08-29 17:47:21.208257 | orchestrator | 2025-08-29 17:47:21 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:21.210324 | orchestrator | 2025-08-29 17:47:21 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:21.210791 | orchestrator | 2025-08-29 17:47:21 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:21.210819 | orchestrator | 2025-08-29 17:47:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:24.246237 | orchestrator | 2025-08-29 17:47:24 | INFO  | Task c6d6cc31-6976-4b2c-9a15-81b0e01287a6 is in state SUCCESS 2025-08-29 17:47:24.249222 | orchestrator | 2025-08-29 17:47:24.249325 | orchestrator | 2025-08-29 17:47:24.249341 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:47:24.249353 | orchestrator | 2025-08-29 17:47:24.249365 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:47:24.249376 | orchestrator | Friday 29 August 2025 17:42:39 +0000 (0:00:00.396) 0:00:00.396 ********* 2025-08-29 17:47:24.249387 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:47:24.249399 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:47:24.249411 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:47:24.249421 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:47:24.249440 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:47:24.249459 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:47:24.249535 | orchestrator | 2025-08-29 17:47:24.249558 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:47:24.249578 | orchestrator | Friday 29 August 2025 17:42:40 +0000 (0:00:00.793) 0:00:01.190 ********* 2025-08-29 17:47:24.249598 | orchestrator | ok: [testbed-node-0] => (item=enable_neutron_True) 2025-08-29 17:47:24.249617 | orchestrator | ok: [testbed-node-1] => (item=enable_neutron_True) 2025-08-29 17:47:24.249637 | orchestrator | ok: [testbed-node-2] => (item=enable_neutron_True) 2025-08-29 17:47:24.249653 | orchestrator | ok: [testbed-node-3] => (item=enable_neutron_True) 2025-08-29 17:47:24.249664 | orchestrator | ok: [testbed-node-4] => (item=enable_neutron_True) 2025-08-29 17:47:24.249675 | orchestrator | ok: [testbed-node-5] => (item=enable_neutron_True) 2025-08-29 17:47:24.249686 | orchestrator | 2025-08-29 17:47:24.249696 | orchestrator | PLAY [Apply role neutron] ****************************************************** 2025-08-29 17:47:24.249707 | orchestrator | 2025-08-29 17:47:24.249734 | orchestrator | TASK [neutron : include_tasks] ************************************************* 2025-08-29 17:47:24.249745 | orchestrator | Friday 29 August 2025 17:42:41 +0000 (0:00:00.626) 0:00:01.816 ********* 2025-08-29 17:47:24.249757 | orchestrator | included: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:47:24.249771 | orchestrator | 2025-08-29 17:47:24.249877 | orchestrator | TASK [neutron : Get container facts] ******************************************* 2025-08-29 17:47:24.249892 | orchestrator | Friday 29 August 2025 17:42:42 +0000 (0:00:01.080) 0:00:02.896 ********* 2025-08-29 17:47:24.249905 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:47:24.249916 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:47:24.249927 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:47:24.249938 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:47:24.249949 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:47:24.249959 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:47:24.249970 | orchestrator | 2025-08-29 17:47:24.249981 | orchestrator | TASK [neutron : Get container volume facts] ************************************ 2025-08-29 17:47:24.249992 | orchestrator | Friday 29 August 2025 17:42:43 +0000 (0:00:01.170) 0:00:04.067 ********* 2025-08-29 17:47:24.250002 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:47:24.250013 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:47:24.250082 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:47:24.250104 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:47:24.250125 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:47:24.250146 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:47:24.250166 | orchestrator | 2025-08-29 17:47:24.251131 | orchestrator | TASK [neutron : Check for ML2/OVN presence] ************************************ 2025-08-29 17:47:24.251340 | orchestrator | Friday 29 August 2025 17:42:44 +0000 (0:00:00.973) 0:00:05.040 ********* 2025-08-29 17:47:24.251368 | orchestrator | ok: [testbed-node-0] => { 2025-08-29 17:47:24.251389 | orchestrator |  "changed": false, 2025-08-29 17:47:24.251410 | orchestrator |  "msg": "All assertions passed" 2025-08-29 17:47:24.251430 | orchestrator | } 2025-08-29 17:47:24.251451 | orchestrator | ok: [testbed-node-1] => { 2025-08-29 17:47:24.251501 | orchestrator |  "changed": false, 2025-08-29 17:47:24.251524 | orchestrator |  "msg": "All assertions passed" 2025-08-29 17:47:24.251545 | orchestrator | } 2025-08-29 17:47:24.251565 | orchestrator | ok: [testbed-node-2] => { 2025-08-29 17:47:24.251585 | orchestrator |  "changed": false, 2025-08-29 17:47:24.251605 | orchestrator |  "msg": "All assertions passed" 2025-08-29 17:47:24.251658 | orchestrator | } 2025-08-29 17:47:24.251678 | orchestrator | ok: [testbed-node-3] => { 2025-08-29 17:47:24.251696 | orchestrator |  "changed": false, 2025-08-29 17:47:24.251715 | orchestrator |  "msg": "All assertions passed" 2025-08-29 17:47:24.251734 | orchestrator | } 2025-08-29 17:47:24.251753 | orchestrator | ok: [testbed-node-4] => { 2025-08-29 17:47:24.251773 | orchestrator |  "changed": false, 2025-08-29 17:47:24.251791 | orchestrator |  "msg": "All assertions passed" 2025-08-29 17:47:24.251808 | orchestrator | } 2025-08-29 17:47:24.251827 | orchestrator | ok: [testbed-node-5] => { 2025-08-29 17:47:24.251846 | orchestrator |  "changed": false, 2025-08-29 17:47:24.251864 | orchestrator |  "msg": "All assertions passed" 2025-08-29 17:47:24.251880 | orchestrator | } 2025-08-29 17:47:24.251895 | orchestrator | 2025-08-29 17:47:24.251906 | orchestrator | TASK [neutron : Check for ML2/OVS presence] ************************************ 2025-08-29 17:47:24.251918 | orchestrator | Friday 29 August 2025 17:42:44 +0000 (0:00:00.636) 0:00:05.677 ********* 2025-08-29 17:47:24.251928 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.251939 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.251952 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.251964 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.251977 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.251989 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.252001 | orchestrator | 2025-08-29 17:47:24.252014 | orchestrator | TASK [service-ks-register : neutron | Creating services] *********************** 2025-08-29 17:47:24.252027 | orchestrator | Friday 29 August 2025 17:42:45 +0000 (0:00:00.577) 0:00:06.254 ********* 2025-08-29 17:47:24.252039 | orchestrator | changed: [testbed-node-0] => (item=neutron (network)) 2025-08-29 17:47:24.252052 | orchestrator | 2025-08-29 17:47:24.252064 | orchestrator | TASK [service-ks-register : neutron | Creating endpoints] ********************** 2025-08-29 17:47:24.252077 | orchestrator | Friday 29 August 2025 17:42:49 +0000 (0:00:03.473) 0:00:09.728 ********* 2025-08-29 17:47:24.252090 | orchestrator | changed: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal) 2025-08-29 17:47:24.252104 | orchestrator | changed: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public) 2025-08-29 17:47:24.252207 | orchestrator | 2025-08-29 17:47:24.252305 | orchestrator | TASK [service-ks-register : neutron | Creating projects] *********************** 2025-08-29 17:47:24.252320 | orchestrator | Friday 29 August 2025 17:42:55 +0000 (0:00:06.666) 0:00:16.395 ********* 2025-08-29 17:47:24.252331 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:47:24.252342 | orchestrator | 2025-08-29 17:47:24.252353 | orchestrator | TASK [service-ks-register : neutron | Creating users] ************************** 2025-08-29 17:47:24.252364 | orchestrator | Friday 29 August 2025 17:42:58 +0000 (0:00:03.250) 0:00:19.645 ********* 2025-08-29 17:47:24.252375 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:47:24.252386 | orchestrator | changed: [testbed-node-0] => (item=neutron -> service) 2025-08-29 17:47:24.252397 | orchestrator | 2025-08-29 17:47:24.252408 | orchestrator | TASK [service-ks-register : neutron | Creating roles] ************************** 2025-08-29 17:47:24.252419 | orchestrator | Friday 29 August 2025 17:43:03 +0000 (0:00:04.295) 0:00:23.941 ********* 2025-08-29 17:47:24.252443 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:47:24.252455 | orchestrator | 2025-08-29 17:47:24.252465 | orchestrator | TASK [service-ks-register : neutron | Granting user roles] ********************* 2025-08-29 17:47:24.252540 | orchestrator | Friday 29 August 2025 17:43:06 +0000 (0:00:03.504) 0:00:27.446 ********* 2025-08-29 17:47:24.252562 | orchestrator | changed: [testbed-node-0] => (item=neutron -> service -> admin) 2025-08-29 17:47:24.252582 | orchestrator | changed: [testbed-node-0] => (item=neutron -> service -> service) 2025-08-29 17:47:24.252601 | orchestrator | 2025-08-29 17:47:24.252621 | orchestrator | TASK [neutron : include_tasks] ************************************************* 2025-08-29 17:47:24.252652 | orchestrator | Friday 29 August 2025 17:43:15 +0000 (0:00:08.469) 0:00:35.915 ********* 2025-08-29 17:47:24.252671 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.252682 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.252693 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.252704 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.252714 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.252725 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.252735 | orchestrator | 2025-08-29 17:47:24.252746 | orchestrator | TASK [Load and persist kernel modules] ***************************************** 2025-08-29 17:47:24.252757 | orchestrator | Friday 29 August 2025 17:43:16 +0000 (0:00:00.810) 0:00:36.725 ********* 2025-08-29 17:47:24.252767 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.252778 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.252788 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.252799 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.252810 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.252820 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.252831 | orchestrator | 2025-08-29 17:47:24.252841 | orchestrator | TASK [neutron : Check IPv6 support] ******************************************** 2025-08-29 17:47:24.252852 | orchestrator | Friday 29 August 2025 17:43:18 +0000 (0:00:02.904) 0:00:39.630 ********* 2025-08-29 17:47:24.252863 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:47:24.252874 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:47:24.252884 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:47:24.252895 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:47:24.252906 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:47:24.252916 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:47:24.252927 | orchestrator | 2025-08-29 17:47:24.252938 | orchestrator | TASK [Setting sysctl values] *************************************************** 2025-08-29 17:47:24.252949 | orchestrator | Friday 29 August 2025 17:43:20 +0000 (0:00:01.370) 0:00:41.000 ********* 2025-08-29 17:47:24.252962 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.252975 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.252987 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.253000 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.253012 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.253024 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.253036 | orchestrator | 2025-08-29 17:47:24.253049 | orchestrator | TASK [neutron : Ensuring config directories exist] ***************************** 2025-08-29 17:47:24.253062 | orchestrator | Friday 29 August 2025 17:43:22 +0000 (0:00:02.504) 0:00:43.505 ********* 2025-08-29 17:47:24.253078 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.253199 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.253223 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.253239 | orchestrator | changed: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.253253 | orchestrator | changed: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.253266 | orchestrator | changed: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.253286 | orchestrator | 2025-08-29 17:47:24.253299 | orchestrator | TASK [neutron : Check if extra ml2 plugins exists] ***************************** 2025-08-29 17:47:24.253312 | orchestrator | Friday 29 August 2025 17:43:25 +0000 (0:00:02.467) 0:00:45.973 ********* 2025-08-29 17:47:24.253324 | orchestrator | [WARNING]: Skipped 2025-08-29 17:47:24.253335 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path 2025-08-29 17:47:24.253346 | orchestrator | due to this access issue: 2025-08-29 17:47:24.253357 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not 2025-08-29 17:47:24.253368 | orchestrator | a directory 2025-08-29 17:47:24.253379 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:47:24.253390 | orchestrator | 2025-08-29 17:47:24.253431 | orchestrator | TASK [neutron : include_tasks] ************************************************* 2025-08-29 17:47:24.253444 | orchestrator | Friday 29 August 2025 17:43:26 +0000 (0:00:00.756) 0:00:46.729 ********* 2025-08-29 17:47:24.253456 | orchestrator | included: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:47:24.253469 | orchestrator | 2025-08-29 17:47:24.253498 | orchestrator | TASK [service-cert-copy : neutron | Copying over extra CA certificates] ******** 2025-08-29 17:47:24.253509 | orchestrator | Friday 29 August 2025 17:43:27 +0000 (0:00:01.082) 0:00:47.812 ********* 2025-08-29 17:47:24.253522 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.253644 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.253670 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.253691 | orchestrator | changed: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.253742 | orchestrator | changed: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.253761 | orchestrator | changed: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.253772 | orchestrator | 2025-08-29 17:47:24.253784 | orchestrator | TASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] *** 2025-08-29 17:47:24.253795 | orchestrator | Friday 29 August 2025 17:43:30 +0000 (0:00:03.781) 0:00:51.594 ********* 2025-08-29 17:47:24.253807 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.253825 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.253837 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.253848 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.253891 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.253904 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.253915 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.253927 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.253943 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.253954 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.253966 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.253984 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.253995 | orchestrator | 2025-08-29 17:47:24.254006 | orchestrator | TASK [service-cert-copy : neutron | Copying over backend internal TLS key] ***** 2025-08-29 17:47:24.254072 | orchestrator | Friday 29 August 2025 17:43:34 +0000 (0:00:03.332) 0:00:54.926 ********* 2025-08-29 17:47:24.254087 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.254099 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.254147 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.254161 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.254178 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.254190 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.254201 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.254220 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.254231 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.254243 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.254254 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.254265 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.254276 | orchestrator | 2025-08-29 17:47:24.254287 | orchestrator | TASK [neutron : Creating TLS backend PEM File] ********************************* 2025-08-29 17:47:24.254303 | orchestrator | Friday 29 August 2025 17:43:38 +0000 (0:00:03.909) 0:00:58.836 ********* 2025-08-29 17:47:24.254315 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.254326 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.254336 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.254347 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.254358 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.254369 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.254379 | orchestrator | 2025-08-29 17:47:24.254390 | orchestrator | TASK [neutron : Check if policies shall be overwritten] ************************ 2025-08-29 17:47:24.254401 | orchestrator | Friday 29 August 2025 17:43:40 +0000 (0:00:02.782) 0:01:01.619 ********* 2025-08-29 17:47:24.254412 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.254423 | orchestrator | 2025-08-29 17:47:24.254434 | orchestrator | TASK [neutron : Set neutron policy file] *************************************** 2025-08-29 17:47:24.254445 | orchestrator | Friday 29 August 2025 17:43:41 +0000 (0:00:00.117) 0:01:01.736 ********* 2025-08-29 17:47:24.254456 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.254466 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.254497 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.254508 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.254519 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.254530 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.254540 | orchestrator | 2025-08-29 17:47:24.254551 | orchestrator | TASK [neutron : Copying over existing policy file] ***************************** 2025-08-29 17:47:24.254569 | orchestrator | Friday 29 August 2025 17:43:41 +0000 (0:00:00.618) 0:01:02.355 ********* 2025-08-29 17:47:24.254585 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.254597 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.254609 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.254620 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.254631 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.254642 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.254663 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.254675 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.254692 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.254710 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.254721 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.254732 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.254743 | orchestrator | 2025-08-29 17:47:24.254754 | orchestrator | TASK [neutron : Copying over config.json files for services] ******************* 2025-08-29 17:47:24.254765 | orchestrator | Friday 29 August 2025 17:43:45 +0000 (0:00:03.454) 0:01:05.809 ********* 2025-08-29 17:47:24.254776 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.254794 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.254806 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.254829 | orchestrator | changed: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.254841 | orchestrator | changed: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.254852 | orchestrator | changed: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.254863 | orchestrator | 2025-08-29 17:47:24.254875 | orchestrator | TASK [neutron : Copying over neutron.conf] ************************************* 2025-08-29 17:47:24.254886 | orchestrator | Friday 29 August 2025 17:43:50 +0000 (0:00:04.967) 0:01:10.776 ********* 2025-08-29 17:47:24.254902 | orchestrator | changed: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.254924 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.254968 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.254990 | orchestrator | changed: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.255013 | orchestrator | changed: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.255045 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.255077 | orchestrator | 2025-08-29 17:47:24.255099 | orchestrator | TASK [neutron : Copying over neutron_vpnaas.conf] ****************************** 2025-08-29 17:47:24.255120 | orchestrator | Friday 29 August 2025 17:43:58 +0000 (0:00:08.195) 0:01:18.972 ********* 2025-08-29 17:47:24.255177 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.255203 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.255225 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.255247 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.255268 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.255291 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.255313 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.255346 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.255387 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.255410 | orchestrator | 2025-08-29 17:47:24.255431 | orchestrator | TASK [neutron : Copying over ssh key] ****************************************** 2025-08-29 17:47:24.255452 | orchestrator | Friday 29 August 2025 17:44:03 +0000 (0:00:04.782) 0:01:23.754 ********* 2025-08-29 17:47:24.255541 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.255566 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.255587 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.255606 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:47:24.255623 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:47:24.255643 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:47:24.255661 | orchestrator | 2025-08-29 17:47:24.255679 | orchestrator | TASK [neutron : Copying over ml2_conf.ini] ************************************* 2025-08-29 17:47:24.255697 | orchestrator | Friday 29 August 2025 17:44:06 +0000 (0:00:03.625) 0:01:27.380 ********* 2025-08-29 17:47:24.255715 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.255734 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.255752 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.255782 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.255812 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.255831 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.255856 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.255875 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.255893 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.255910 | orchestrator | 2025-08-29 17:47:24.255927 | orchestrator | TASK [neutron : Copying over linuxbridge_agent.ini] **************************** 2025-08-29 17:47:24.255945 | orchestrator | Friday 29 August 2025 17:44:11 +0000 (0:00:05.044) 0:01:32.425 ********* 2025-08-29 17:47:24.255991 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256004 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256014 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256023 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256033 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256043 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.256053 | orchestrator | 2025-08-29 17:47:24.256063 | orchestrator | TASK [neutron : Copying over openvswitch_agent.ini] **************************** 2025-08-29 17:47:24.256073 | orchestrator | Friday 29 August 2025 17:44:13 +0000 (0:00:01.926) 0:01:34.352 ********* 2025-08-29 17:47:24.256083 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256093 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256102 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256112 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256121 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256131 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.256141 | orchestrator | 2025-08-29 17:47:24.256150 | orchestrator | TASK [neutron : Copying over sriov_agent.ini] ********************************** 2025-08-29 17:47:24.256160 | orchestrator | Friday 29 August 2025 17:44:16 +0000 (0:00:02.566) 0:01:36.918 ********* 2025-08-29 17:47:24.256170 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256179 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256189 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256205 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256215 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256225 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.256234 | orchestrator | 2025-08-29 17:47:24.256244 | orchestrator | TASK [neutron : Copying over mlnx_agent.ini] *********************************** 2025-08-29 17:47:24.256253 | orchestrator | Friday 29 August 2025 17:44:18 +0000 (0:00:01.997) 0:01:38.916 ********* 2025-08-29 17:47:24.256263 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256272 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256282 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256291 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256301 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256310 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.256320 | orchestrator | 2025-08-29 17:47:24.256329 | orchestrator | TASK [neutron : Copying over eswitchd.conf] ************************************ 2025-08-29 17:47:24.256339 | orchestrator | Friday 29 August 2025 17:44:20 +0000 (0:00:01.995) 0:01:40.912 ********* 2025-08-29 17:47:24.256349 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256358 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256368 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256377 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256387 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.256396 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256406 | orchestrator | 2025-08-29 17:47:24.256416 | orchestrator | TASK [neutron : Copying over dhcp_agent.ini] *********************************** 2025-08-29 17:47:24.256425 | orchestrator | Friday 29 August 2025 17:44:22 +0000 (0:00:01.885) 0:01:42.797 ********* 2025-08-29 17:47:24.256435 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256450 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256460 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256469 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256495 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.256504 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256517 | orchestrator | 2025-08-29 17:47:24.256534 | orchestrator | TASK [neutron : Copying over dnsmasq.conf] ************************************* 2025-08-29 17:47:24.256550 | orchestrator | Friday 29 August 2025 17:44:24 +0000 (0:00:02.395) 0:01:45.193 ********* 2025-08-29 17:47:24.256568 | orchestrator | skipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2)  2025-08-29 17:47:24.256596 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256615 | orchestrator | skipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2)  2025-08-29 17:47:24.256633 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256649 | orchestrator | skipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2)  2025-08-29 17:47:24.256664 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256674 | orchestrator | skipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2)  2025-08-29 17:47:24.256683 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256693 | orchestrator | skipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2)  2025-08-29 17:47:24.256702 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256712 | orchestrator | skipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2)  2025-08-29 17:47:24.256721 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.256731 | orchestrator | 2025-08-29 17:47:24.256740 | orchestrator | TASK [neutron : Copying over l3_agent.ini] ************************************* 2025-08-29 17:47:24.256750 | orchestrator | Friday 29 August 2025 17:44:26 +0000 (0:00:02.135) 0:01:47.328 ********* 2025-08-29 17:47:24.256760 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.256778 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.256806 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.256825 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.256850 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.256880 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.256900 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.256918 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.256937 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.256956 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.256974 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.256993 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.257011 | orchestrator | 2025-08-29 17:47:24.257029 | orchestrator | TASK [neutron : Copying over fwaas_driver.ini] ********************************* 2025-08-29 17:47:24.257047 | orchestrator | Friday 29 August 2025 17:44:28 +0000 (0:00:02.107) 0:01:49.435 ********* 2025-08-29 17:47:24.257073 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.257092 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.257116 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.257144 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.257164 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.257182 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.257200 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.257219 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.257237 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.257256 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.257282 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.257315 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.257333 | orchestrator | 2025-08-29 17:47:24.257351 | orchestrator | TASK [neutron : Copying over metadata_agent.ini] ******************************* 2025-08-29 17:47:24.257369 | orchestrator | Friday 29 August 2025 17:44:31 +0000 (0:00:02.941) 0:01:52.377 ********* 2025-08-29 17:47:24.257386 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.257404 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.257422 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.257445 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.257463 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.257504 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.257523 | orchestrator | 2025-08-29 17:47:24.257541 | orchestrator | TASK [neutron : Copying over neutron_ovn_metadata_agent.ini] ******************* 2025-08-29 17:47:24.257559 | orchestrator | Friday 29 August 2025 17:44:34 +0000 (0:00:03.111) 0:01:55.488 ********* 2025-08-29 17:47:24.257577 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.257594 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.257612 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.257630 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:47:24.257647 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:47:24.257665 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:47:24.257682 | orchestrator | 2025-08-29 17:47:24.257700 | orchestrator | TASK [neutron : Copying over neutron_ovn_vpn_agent.ini] ************************ 2025-08-29 17:47:24.257718 | orchestrator | Friday 29 August 2025 17:44:38 +0000 (0:00:03.476) 0:01:58.965 ********* 2025-08-29 17:47:24.257736 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.257753 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.257770 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.257788 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.257806 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.257823 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.257841 | orchestrator | 2025-08-29 17:47:24.257859 | orchestrator | TASK [neutron : Copying over metering_agent.ini] ******************************* 2025-08-29 17:47:24.257877 | orchestrator | Friday 29 August 2025 17:44:41 +0000 (0:00:03.067) 0:02:02.033 ********* 2025-08-29 17:47:24.257895 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.257913 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.257931 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.257948 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.257966 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.257984 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.258001 | orchestrator | 2025-08-29 17:47:24.258055 | orchestrator | TASK [neutron : Copying over ironic_neutron_agent.ini] ************************* 2025-08-29 17:47:24.258077 | orchestrator | Friday 29 August 2025 17:44:43 +0000 (0:00:01.818) 0:02:03.852 ********* 2025-08-29 17:47:24.258096 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.258114 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.258132 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.258151 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.258169 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.258187 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.258205 | orchestrator | 2025-08-29 17:47:24.258224 | orchestrator | TASK [neutron : Copying over bgp_dragent.ini] ********************************** 2025-08-29 17:47:24.258242 | orchestrator | Friday 29 August 2025 17:44:46 +0000 (0:00:03.052) 0:02:06.905 ********* 2025-08-29 17:47:24.258260 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.258278 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.258308 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.258326 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.258344 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.258362 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.258380 | orchestrator | 2025-08-29 17:47:24.258398 | orchestrator | TASK [neutron : Copying over ovn_agent.ini] ************************************ 2025-08-29 17:47:24.258416 | orchestrator | Friday 29 August 2025 17:44:48 +0000 (0:00:02.182) 0:02:09.088 ********* 2025-08-29 17:47:24.258434 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.258452 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.258525 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.258546 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.258564 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.258581 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.258599 | orchestrator | 2025-08-29 17:47:24.258616 | orchestrator | TASK [neutron : Copying over nsx.ini] ****************************************** 2025-08-29 17:47:24.258634 | orchestrator | Friday 29 August 2025 17:44:50 +0000 (0:00:02.064) 0:02:11.153 ********* 2025-08-29 17:47:24.258652 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.258669 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.258687 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.258704 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.258722 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.258739 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.258757 | orchestrator | 2025-08-29 17:47:24.258774 | orchestrator | TASK [neutron : Copy neutron-l3-agent-wrapper script] ************************** 2025-08-29 17:47:24.258792 | orchestrator | Friday 29 August 2025 17:44:53 +0000 (0:00:02.722) 0:02:13.875 ********* 2025-08-29 17:47:24.258809 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.258836 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.258853 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.258867 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.258882 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.258896 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.258910 | orchestrator | 2025-08-29 17:47:24.258924 | orchestrator | TASK [neutron : Copying over extra ml2 plugins] ******************************** 2025-08-29 17:47:24.258939 | orchestrator | Friday 29 August 2025 17:44:55 +0000 (0:00:02.166) 0:02:16.042 ********* 2025-08-29 17:47:24.258953 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.258966 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.258979 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.258992 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.259004 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.259012 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.259019 | orchestrator | 2025-08-29 17:47:24.259027 | orchestrator | TASK [neutron : Copying over neutron-tls-proxy.cfg] **************************** 2025-08-29 17:47:24.259035 | orchestrator | Friday 29 August 2025 17:44:57 +0000 (0:00:01.811) 0:02:17.854 ********* 2025-08-29 17:47:24.259043 | orchestrator | skipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2)  2025-08-29 17:47:24.259051 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.259059 | orchestrator | skipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2)  2025-08-29 17:47:24.259067 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.259079 | orchestrator | skipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2)  2025-08-29 17:47:24.259091 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.259104 | orchestrator | skipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2)  2025-08-29 17:47:24.259117 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.259131 | orchestrator | skipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2)  2025-08-29 17:47:24.259146 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.259170 | orchestrator | skipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2)  2025-08-29 17:47:24.259184 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.259198 | orchestrator | 2025-08-29 17:47:24.259206 | orchestrator | TASK [neutron : Copying over neutron_taas.conf] ******************************** 2025-08-29 17:47:24.259214 | orchestrator | Friday 29 August 2025 17:44:59 +0000 (0:00:01.923) 0:02:19.778 ********* 2025-08-29 17:47:24.259223 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.259231 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.259240 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.259248 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.259265 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})  2025-08-29 17:47:24.259273 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.259285 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.259300 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.259309 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.259317 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.259325 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})  2025-08-29 17:47:24.259333 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.259340 | orchestrator | 2025-08-29 17:47:24.259348 | orchestrator | TASK [neutron : Check neutron containers] ************************************** 2025-08-29 17:47:24.259356 | orchestrator | Friday 29 August 2025 17:45:01 +0000 (0:00:02.550) 0:02:22.328 ********* 2025-08-29 17:47:24.259364 | orchestrator | changed: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.259378 | orchestrator | changed: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.259396 | orchestrator | changed: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) 2025-08-29 17:47:24.259405 | orchestrator | changed: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.259413 | orchestrator | changed: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.259422 | orchestrator | changed: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) 2025-08-29 17:47:24.259430 | orchestrator | 2025-08-29 17:47:24.259438 | orchestrator | TASK [neutron : include_tasks] ************************************************* 2025-08-29 17:47:24.259450 | orchestrator | Friday 29 August 2025 17:45:04 +0000 (0:00:03.208) 0:02:25.536 ********* 2025-08-29 17:47:24.259459 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:47:24.259466 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:47:24.259493 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:47:24.259502 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:47:24.259509 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:47:24.259517 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:47:24.259525 | orchestrator | 2025-08-29 17:47:24.259533 | orchestrator | TASK [neutron : Creating Neutron database] ************************************* 2025-08-29 17:47:24.259546 | orchestrator | Friday 29 August 2025 17:45:05 +0000 (0:00:00.608) 0:02:26.145 ********* 2025-08-29 17:47:24.259554 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:47:24.259562 | orchestrator | 2025-08-29 17:47:24.259569 | orchestrator | TASK [neutron : Creating Neutron database user and setting permissions] ******** 2025-08-29 17:47:24.259577 | orchestrator | Friday 29 August 2025 17:45:07 +0000 (0:00:02.145) 0:02:28.290 ********* 2025-08-29 17:47:24.259585 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:47:24.259593 | orchestrator | 2025-08-29 17:47:24.259601 | orchestrator | TASK [neutron : Running Neutron bootstrap container] *************************** 2025-08-29 17:47:24.259609 | orchestrator | Friday 29 August 2025 17:45:09 +0000 (0:00:02.335) 0:02:30.626 ********* 2025-08-29 17:47:24.259617 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:47:24.259625 | orchestrator | 2025-08-29 17:47:24.259632 | orchestrator | TASK [neutron : Flush Handlers] ************************************************ 2025-08-29 17:47:24.259640 | orchestrator | Friday 29 August 2025 17:45:53 +0000 (0:00:43.089) 0:03:13.715 ********* 2025-08-29 17:47:24.259648 | orchestrator | 2025-08-29 17:47:24.259673 | orchestrator | TASK [neutron : Flush Handlers] ************************************************ 2025-08-29 17:47:24.259682 | orchestrator | Friday 29 August 2025 17:45:53 +0000 (0:00:00.131) 0:03:13.847 ********* 2025-08-29 17:47:24.259690 | orchestrator | 2025-08-29 17:47:24.259698 | orchestrator | TASK [neutron : Flush Handlers] ************************************************ 2025-08-29 17:47:24.259706 | orchestrator | Friday 29 August 2025 17:45:53 +0000 (0:00:00.067) 0:03:13.915 ********* 2025-08-29 17:47:24.259714 | orchestrator | 2025-08-29 17:47:24.259722 | orchestrator | TASK [neutron : Flush Handlers] ************************************************ 2025-08-29 17:47:24.259729 | orchestrator | Friday 29 August 2025 17:45:53 +0000 (0:00:00.050) 0:03:13.966 ********* 2025-08-29 17:47:24.259737 | orchestrator | 2025-08-29 17:47:24.259745 | orchestrator | TASK [neutron : Flush Handlers] ************************************************ 2025-08-29 17:47:24.259753 | orchestrator | Friday 29 August 2025 17:45:53 +0000 (0:00:00.168) 0:03:14.134 ********* 2025-08-29 17:47:24.259761 | orchestrator | 2025-08-29 17:47:24.259769 | orchestrator | TASK [neutron : Flush Handlers] ************************************************ 2025-08-29 17:47:24.259777 | orchestrator | Friday 29 August 2025 17:45:53 +0000 (0:00:00.061) 0:03:14.196 ********* 2025-08-29 17:47:24.259784 | orchestrator | 2025-08-29 17:47:24.259792 | orchestrator | RUNNING HANDLER [neutron : Restart neutron-server container] ******************* 2025-08-29 17:47:24.259800 | orchestrator | Friday 29 August 2025 17:45:53 +0000 (0:00:00.061) 0:03:14.257 ********* 2025-08-29 17:47:24.259808 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:47:24.259816 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:47:24.259824 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:47:24.259832 | orchestrator | 2025-08-29 17:47:24.259840 | orchestrator | RUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] ******* 2025-08-29 17:47:24.259847 | orchestrator | Friday 29 August 2025 17:46:21 +0000 (0:00:28.305) 0:03:42.562 ********* 2025-08-29 17:47:24.259855 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:47:24.259863 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:47:24.259871 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:47:24.259878 | orchestrator | 2025-08-29 17:47:24.259886 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:47:24.259894 | orchestrator | testbed-node-0 : ok=27  changed=16  unreachable=0 failed=0 skipped=32  rescued=0 ignored=0 2025-08-29 17:47:24.259903 | orchestrator | testbed-node-1 : ok=17  changed=9  unreachable=0 failed=0 skipped=31  rescued=0 ignored=0 2025-08-29 17:47:24.259911 | orchestrator | testbed-node-2 : ok=17  changed=9  unreachable=0 failed=0 skipped=31  rescued=0 ignored=0 2025-08-29 17:47:24.259919 | orchestrator | testbed-node-3 : ok=15  changed=7  unreachable=0 failed=0 skipped=33  rescued=0 ignored=0 2025-08-29 17:47:24.259933 | orchestrator | testbed-node-4 : ok=15  changed=7  unreachable=0 failed=0 skipped=33  rescued=0 ignored=0 2025-08-29 17:47:24.259941 | orchestrator | testbed-node-5 : ok=15  changed=7  unreachable=0 failed=0 skipped=33  rescued=0 ignored=0 2025-08-29 17:47:24.259949 | orchestrator | 2025-08-29 17:47:24.259957 | orchestrator | 2025-08-29 17:47:24.259965 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:47:24.259973 | orchestrator | Friday 29 August 2025 17:47:23 +0000 (0:01:01.541) 0:04:44.104 ********* 2025-08-29 17:47:24.259981 | orchestrator | =============================================================================== 2025-08-29 17:47:24.259989 | orchestrator | neutron : Restart neutron-ovn-metadata-agent container ----------------- 61.54s 2025-08-29 17:47:24.259996 | orchestrator | neutron : Running Neutron bootstrap container -------------------------- 43.09s 2025-08-29 17:47:24.260004 | orchestrator | neutron : Restart neutron-server container ----------------------------- 28.31s 2025-08-29 17:47:24.260012 | orchestrator | service-ks-register : neutron | Granting user roles --------------------- 8.47s 2025-08-29 17:47:24.260027 | orchestrator | neutron : Copying over neutron.conf ------------------------------------- 8.20s 2025-08-29 17:47:24.260035 | orchestrator | service-ks-register : neutron | Creating endpoints ---------------------- 6.67s 2025-08-29 17:47:24.260043 | orchestrator | neutron : Copying over ml2_conf.ini ------------------------------------- 5.04s 2025-08-29 17:47:24.260051 | orchestrator | neutron : Copying over config.json files for services ------------------- 4.97s 2025-08-29 17:47:24.260059 | orchestrator | neutron : Copying over neutron_vpnaas.conf ------------------------------ 4.78s 2025-08-29 17:47:24.260067 | orchestrator | service-ks-register : neutron | Creating users -------------------------- 4.30s 2025-08-29 17:47:24.260075 | orchestrator | service-cert-copy : neutron | Copying over backend internal TLS key ----- 3.91s 2025-08-29 17:47:24.260082 | orchestrator | service-cert-copy : neutron | Copying over extra CA certificates -------- 3.78s 2025-08-29 17:47:24.260090 | orchestrator | neutron : Copying over ssh key ------------------------------------------ 3.63s 2025-08-29 17:47:24.260098 | orchestrator | service-ks-register : neutron | Creating roles -------------------------- 3.50s 2025-08-29 17:47:24.260106 | orchestrator | neutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 3.48s 2025-08-29 17:47:24.260114 | orchestrator | service-ks-register : neutron | Creating services ----------------------- 3.47s 2025-08-29 17:47:24.260121 | orchestrator | neutron : Copying over existing policy file ----------------------------- 3.45s 2025-08-29 17:47:24.260129 | orchestrator | service-cert-copy : neutron | Copying over backend internal TLS certificate --- 3.33s 2025-08-29 17:47:24.260151 | orchestrator | service-ks-register : neutron | Creating projects ----------------------- 3.25s 2025-08-29 17:47:24.260160 | orchestrator | neutron : Check neutron containers -------------------------------------- 3.21s 2025-08-29 17:47:24.260168 | orchestrator | 2025-08-29 17:47:24 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:24.260176 | orchestrator | 2025-08-29 17:47:24 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:24.260184 | orchestrator | 2025-08-29 17:47:24 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:24.260192 | orchestrator | 2025-08-29 17:47:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:27.281827 | orchestrator | 2025-08-29 17:47:27 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:27.282172 | orchestrator | 2025-08-29 17:47:27 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:27.283157 | orchestrator | 2025-08-29 17:47:27 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:27.283610 | orchestrator | 2025-08-29 17:47:27 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:27.283681 | orchestrator | 2025-08-29 17:47:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:30.312526 | orchestrator | 2025-08-29 17:47:30 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:30.312785 | orchestrator | 2025-08-29 17:47:30 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:30.313434 | orchestrator | 2025-08-29 17:47:30 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:30.314270 | orchestrator | 2025-08-29 17:47:30 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:30.314348 | orchestrator | 2025-08-29 17:47:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:33.343451 | orchestrator | 2025-08-29 17:47:33 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:33.343679 | orchestrator | 2025-08-29 17:47:33 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:33.344291 | orchestrator | 2025-08-29 17:47:33 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:33.345033 | orchestrator | 2025-08-29 17:47:33 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:33.345171 | orchestrator | 2025-08-29 17:47:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:36.374395 | orchestrator | 2025-08-29 17:47:36 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:36.374654 | orchestrator | 2025-08-29 17:47:36 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:36.375397 | orchestrator | 2025-08-29 17:47:36 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:36.376533 | orchestrator | 2025-08-29 17:47:36 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:36.376560 | orchestrator | 2025-08-29 17:47:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:39.406764 | orchestrator | 2025-08-29 17:47:39 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:39.407647 | orchestrator | 2025-08-29 17:47:39 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:39.408219 | orchestrator | 2025-08-29 17:47:39 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:39.408919 | orchestrator | 2025-08-29 17:47:39 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:39.408947 | orchestrator | 2025-08-29 17:47:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:42.443951 | orchestrator | 2025-08-29 17:47:42 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:42.445700 | orchestrator | 2025-08-29 17:47:42 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:42.446345 | orchestrator | 2025-08-29 17:47:42 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:42.447224 | orchestrator | 2025-08-29 17:47:42 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:42.447249 | orchestrator | 2025-08-29 17:47:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:45.484402 | orchestrator | 2025-08-29 17:47:45 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:45.484699 | orchestrator | 2025-08-29 17:47:45 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:45.485575 | orchestrator | 2025-08-29 17:47:45 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:45.486211 | orchestrator | 2025-08-29 17:47:45 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:45.486235 | orchestrator | 2025-08-29 17:47:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:48.516252 | orchestrator | 2025-08-29 17:47:48 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:48.516690 | orchestrator | 2025-08-29 17:47:48 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:48.517104 | orchestrator | 2025-08-29 17:47:48 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:48.518069 | orchestrator | 2025-08-29 17:47:48 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:48.518094 | orchestrator | 2025-08-29 17:47:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:51.568777 | orchestrator | 2025-08-29 17:47:51 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:51.569412 | orchestrator | 2025-08-29 17:47:51 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:51.569914 | orchestrator | 2025-08-29 17:47:51 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:51.570580 | orchestrator | 2025-08-29 17:47:51 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:51.570612 | orchestrator | 2025-08-29 17:47:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:54.597587 | orchestrator | 2025-08-29 17:47:54 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:54.599392 | orchestrator | 2025-08-29 17:47:54 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:54.602196 | orchestrator | 2025-08-29 17:47:54 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:54.603543 | orchestrator | 2025-08-29 17:47:54 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:54.603567 | orchestrator | 2025-08-29 17:47:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:47:57.632950 | orchestrator | 2025-08-29 17:47:57 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:47:57.635580 | orchestrator | 2025-08-29 17:47:57 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:47:57.638315 | orchestrator | 2025-08-29 17:47:57 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:47:57.641625 | orchestrator | 2025-08-29 17:47:57 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:47:57.642071 | orchestrator | 2025-08-29 17:47:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:00.689380 | orchestrator | 2025-08-29 17:48:00 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:00.689544 | orchestrator | 2025-08-29 17:48:00 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:00.689947 | orchestrator | 2025-08-29 17:48:00 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:00.690686 | orchestrator | 2025-08-29 17:48:00 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:00.690715 | orchestrator | 2025-08-29 17:48:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:03.721343 | orchestrator | 2025-08-29 17:48:03 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:03.722681 | orchestrator | 2025-08-29 17:48:03 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:03.722720 | orchestrator | 2025-08-29 17:48:03 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:03.722736 | orchestrator | 2025-08-29 17:48:03 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:03.722956 | orchestrator | 2025-08-29 17:48:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:06.748179 | orchestrator | 2025-08-29 17:48:06 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:06.748269 | orchestrator | 2025-08-29 17:48:06 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:06.748284 | orchestrator | 2025-08-29 17:48:06 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:06.748716 | orchestrator | 2025-08-29 17:48:06 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:06.748744 | orchestrator | 2025-08-29 17:48:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:09.786178 | orchestrator | 2025-08-29 17:48:09 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:09.787162 | orchestrator | 2025-08-29 17:48:09 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:09.788142 | orchestrator | 2025-08-29 17:48:09 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:09.789096 | orchestrator | 2025-08-29 17:48:09 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:09.789125 | orchestrator | 2025-08-29 17:48:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:12.833881 | orchestrator | 2025-08-29 17:48:12 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:12.834659 | orchestrator | 2025-08-29 17:48:12 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:12.836063 | orchestrator | 2025-08-29 17:48:12 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:12.837087 | orchestrator | 2025-08-29 17:48:12 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:12.837751 | orchestrator | 2025-08-29 17:48:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:15.890365 | orchestrator | 2025-08-29 17:48:15 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:15.891691 | orchestrator | 2025-08-29 17:48:15 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:15.891909 | orchestrator | 2025-08-29 17:48:15 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:15.893159 | orchestrator | 2025-08-29 17:48:15 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:15.893177 | orchestrator | 2025-08-29 17:48:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:18.972010 | orchestrator | 2025-08-29 17:48:18 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:18.972626 | orchestrator | 2025-08-29 17:48:18 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:18.973781 | orchestrator | 2025-08-29 17:48:18 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:18.974732 | orchestrator | 2025-08-29 17:48:18 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:18.974790 | orchestrator | 2025-08-29 17:48:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:22.024653 | orchestrator | 2025-08-29 17:48:22 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:22.025235 | orchestrator | 2025-08-29 17:48:22 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:22.026260 | orchestrator | 2025-08-29 17:48:22 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:22.027592 | orchestrator | 2025-08-29 17:48:22 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:22.027655 | orchestrator | 2025-08-29 17:48:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:25.074916 | orchestrator | 2025-08-29 17:48:25 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:25.076091 | orchestrator | 2025-08-29 17:48:25 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:25.077574 | orchestrator | 2025-08-29 17:48:25 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:25.079169 | orchestrator | 2025-08-29 17:48:25 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:25.079212 | orchestrator | 2025-08-29 17:48:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:28.121921 | orchestrator | 2025-08-29 17:48:28 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:28.122431 | orchestrator | 2025-08-29 17:48:28 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:28.124548 | orchestrator | 2025-08-29 17:48:28 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:28.126407 | orchestrator | 2025-08-29 17:48:28 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:28.127378 | orchestrator | 2025-08-29 17:48:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:31.183909 | orchestrator | 2025-08-29 17:48:31 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:31.186167 | orchestrator | 2025-08-29 17:48:31 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:31.188977 | orchestrator | 2025-08-29 17:48:31 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state STARTED 2025-08-29 17:48:31.190885 | orchestrator | 2025-08-29 17:48:31 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:31.191269 | orchestrator | 2025-08-29 17:48:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:34.238745 | orchestrator | 2025-08-29 17:48:34 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:34.239379 | orchestrator | 2025-08-29 17:48:34 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:34.242865 | orchestrator | 2025-08-29 17:48:34 | INFO  | Task 596f371d-8df8-4549-9794-e3d2eca11665 is in state SUCCESS 2025-08-29 17:48:34.245197 | orchestrator | 2025-08-29 17:48:34.245546 | orchestrator | 2025-08-29 17:48:34.245568 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:48:34.245582 | orchestrator | 2025-08-29 17:48:34.245594 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:48:34.245606 | orchestrator | Friday 29 August 2025 17:45:12 +0000 (0:00:00.284) 0:00:00.284 ********* 2025-08-29 17:48:34.245617 | orchestrator | ok: [testbed-manager] 2025-08-29 17:48:34.245631 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:48:34.245642 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:48:34.245682 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:48:34.245697 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:48:34.245709 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:48:34.245720 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:48:34.245730 | orchestrator | 2025-08-29 17:48:34.245742 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:48:34.245753 | orchestrator | Friday 29 August 2025 17:45:13 +0000 (0:00:00.872) 0:00:01.156 ********* 2025-08-29 17:48:34.246635 | orchestrator | ok: [testbed-manager] => (item=enable_prometheus_True) 2025-08-29 17:48:34.247603 | orchestrator | ok: [testbed-node-0] => (item=enable_prometheus_True) 2025-08-29 17:48:34.247620 | orchestrator | ok: [testbed-node-1] => (item=enable_prometheus_True) 2025-08-29 17:48:34.247631 | orchestrator | ok: [testbed-node-2] => (item=enable_prometheus_True) 2025-08-29 17:48:34.247642 | orchestrator | ok: [testbed-node-3] => (item=enable_prometheus_True) 2025-08-29 17:48:34.247653 | orchestrator | ok: [testbed-node-4] => (item=enable_prometheus_True) 2025-08-29 17:48:34.247663 | orchestrator | ok: [testbed-node-5] => (item=enable_prometheus_True) 2025-08-29 17:48:34.247675 | orchestrator | 2025-08-29 17:48:34.247686 | orchestrator | PLAY [Apply role prometheus] *************************************************** 2025-08-29 17:48:34.247697 | orchestrator | 2025-08-29 17:48:34.247708 | orchestrator | TASK [prometheus : include_tasks] ********************************************** 2025-08-29 17:48:34.247719 | orchestrator | Friday 29 August 2025 17:45:14 +0000 (0:00:00.753) 0:00:01.910 ********* 2025-08-29 17:48:34.247764 | orchestrator | included: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:48:34.247778 | orchestrator | 2025-08-29 17:48:34.247789 | orchestrator | TASK [prometheus : Ensuring config directories exist] ************************** 2025-08-29 17:48:34.247799 | orchestrator | Friday 29 August 2025 17:45:16 +0000 (0:00:01.458) 0:00:03.368 ********* 2025-08-29 17:48:34.247813 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.247843 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) 2025-08-29 17:48:34.247856 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.247867 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248004 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248022 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248033 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248044 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248054 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248087 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248098 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248189 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248219 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248231 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248243 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248255 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248272 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248283 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248303 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248429 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) 2025-08-29 17:48:34.248547 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248563 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248574 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248592 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248607 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.248629 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248720 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248734 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248758 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.248770 | orchestrator | 2025-08-29 17:48:34.248781 | orchestrator | TASK [prometheus : include_tasks] ********************************************** 2025-08-29 17:48:34.248791 | orchestrator | Friday 29 August 2025 17:45:19 +0000 (0:00:03.655) 0:00:07.024 ********* 2025-08-29 17:48:34.248802 | orchestrator | included: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:48:34.248814 | orchestrator | 2025-08-29 17:48:34.248824 | orchestrator | TASK [service-cert-copy : prometheus | Copying over extra CA certificates] ***** 2025-08-29 17:48:34.248835 | orchestrator | Friday 29 August 2025 17:45:21 +0000 (0:00:01.348) 0:00:08.372 ********* 2025-08-29 17:48:34.248853 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) 2025-08-29 17:48:34.248874 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248885 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248957 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248969 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.248991 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.249001 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.249011 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.249026 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249043 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249053 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249137 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249161 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249172 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249202 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249212 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249239 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249250 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249317 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) 2025-08-29 17:48:34.249345 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249356 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249367 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249377 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249400 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249411 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249447 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.249459 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249523 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249534 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.249578 | orchestrator | 2025-08-29 17:48:34.249587 | orchestrator | TASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] *** 2025-08-29 17:48:34.249600 | orchestrator | Friday 29 August 2025 17:45:26 +0000 (0:00:05.897) 0:00:14.270 ********* 2025-08-29 17:48:34.249607 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})  2025-08-29 17:48:34.249617 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.249625 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249657 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})  2025-08-29 17:48:34.249666 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249672 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.249683 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249693 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249700 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249706 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249713 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.249737 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.249745 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249751 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249758 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249769 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249778 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.249785 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249792 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249816 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249824 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.249830 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.249837 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.249843 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.249854 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.249860 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249887 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249895 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.249901 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.249908 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249934 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249943 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.249953 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.249970 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249980 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.249991 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.250003 | orchestrator | 2025-08-29 17:48:34.250040 | orchestrator | TASK [service-cert-copy : prometheus | Copying over backend internal TLS key] *** 2025-08-29 17:48:34.250050 | orchestrator | Friday 29 August 2025 17:45:28 +0000 (0:00:01.956) 0:00:16.226 ********* 2025-08-29 17:48:34.250058 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.250069 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250075 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250100 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250107 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250118 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.250125 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})  2025-08-29 17:48:34.250131 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.250138 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250148 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})  2025-08-29 17:48:34.250171 | orchestrator | skipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250178 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.250184 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.250198 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250204 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250210 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250231 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250238 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.250244 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.250250 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250273 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250285 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250291 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})  2025-08-29 17:48:34.250297 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.250303 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.250310 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250328 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250334 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.250339 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.250360 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250372 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250378 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.250384 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})  2025-08-29 17:48:34.250389 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250395 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})  2025-08-29 17:48:34.250400 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.250406 | orchestrator | 2025-08-29 17:48:34.250411 | orchestrator | TASK [prometheus : Copying over config.json files] ***************************** 2025-08-29 17:48:34.250417 | orchestrator | Friday 29 August 2025 17:45:31 +0000 (0:00:03.064) 0:00:19.291 ********* 2025-08-29 17:48:34.250426 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) 2025-08-29 17:48:34.250431 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.250455 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.250462 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.250481 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.250487 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.250493 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.250505 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250515 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.250522 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250549 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250556 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250562 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250568 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250574 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250585 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250596 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) 2025-08-29 17:48:34.250625 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250636 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250642 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250648 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250653 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250662 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250672 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250699 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250706 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.250711 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250717 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250726 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.250736 | orchestrator | 2025-08-29 17:48:34.250742 | orchestrator | TASK [prometheus : Find custom prometheus alert rules files] ******************* 2025-08-29 17:48:34.250747 | orchestrator | Friday 29 August 2025 17:45:38 +0000 (0:00:06.345) 0:00:25.636 ********* 2025-08-29 17:48:34.250756 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:48:34.250764 | orchestrator | 2025-08-29 17:48:34.250770 | orchestrator | TASK [prometheus : Copying over custom prometheus alert rules files] *********** 2025-08-29 17:48:34.250775 | orchestrator | Friday 29 August 2025 17:45:39 +0000 (0:00:01.110) 0:00:26.746 ********* 2025-08-29 17:48:34.250784 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1094084, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6006212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250794 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1094084, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6006212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250817 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1094084, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6006212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250824 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1094084, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6006212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250830 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1094084, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6006212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.250835 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1094111, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6083915, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250841 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1094111, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6083915, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250855 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1094084, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6006212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250861 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1094111, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6083915, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250880 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1094084, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6006212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250887 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1094111, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6083915, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250893 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1094079, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5988238, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250898 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1094079, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5988238, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250904 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1094111, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6083915, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250920 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1094111, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6083915, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250926 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1094104, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6061723, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250947 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1094079, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5988238, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250953 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12980, 'inode': 1094111, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6083915, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.250959 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1094079, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5988238, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250964 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1094079, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5988238, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250970 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1094104, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6061723, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.250994 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1094079, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5988238, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251000 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1094104, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6061723, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251022 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1094075, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5979831, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251029 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1094075, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5979831, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251035 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1094086, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6009414, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251041 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1094104, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6061723, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251060 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1094075, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5979831, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251069 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1094104, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6061723, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251075 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1094099, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.603871, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251096 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1094104, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6061723, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251103 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1094086, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6009414, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251109 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1094075, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5979831, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251114 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 55956, 'inode': 1094079, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5988238, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251124 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1094075, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5979831, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251141 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1094086, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6009414, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251147 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1094088, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.602203, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251168 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1094086, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6009414, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251175 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1094086, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6009414, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251181 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1094075, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5979831, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251186 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1094099, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.603871, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251196 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1094082, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5998523, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251205 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1094099, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.603871, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251211 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1094086, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6009414, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251231 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1094099, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.603871, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251238 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1094088, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.602203, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251244 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1094099, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.603871, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251249 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094108, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6077433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251259 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1094088, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.602203, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251267 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1094082, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5998523, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251273 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1094099, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.603871, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251293 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1094088, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.602203, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251300 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1094104, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6061723, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251305 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094108, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6077433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251316 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1094088, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.602203, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251322 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1094082, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5998523, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251332 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094073, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5971754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251338 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1094088, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.602203, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251344 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1094122, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.610821, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251364 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094073, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5971754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251371 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1094082, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5998523, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251381 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1094082, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5998523, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251387 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094108, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6077433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251395 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1094107, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6064646, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251401 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1094122, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.610821, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251407 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094108, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6077433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251427 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1094082, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5998523, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251434 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1094075, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5979831, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251444 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094108, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6077433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251450 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1094107, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6064646, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251459 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094073, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5971754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251478 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094077, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5985248, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251485 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094073, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5971754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251507 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094073, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5971754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251522 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094108, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6077433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251528 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1094122, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.610821, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251533 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094077, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5985248, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251542 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1094074, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5976844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251547 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1094086, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6009414, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251553 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1094122, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.610821, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251562 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1094122, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.610821, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251571 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1094107, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6064646, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251577 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094073, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5971754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251583 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1094107, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6064646, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251591 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1094074, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5976844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251597 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1094097, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6032066, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251603 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094077, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5985248, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251612 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094077, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5985248, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251622 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1094107, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6064646, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251627 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1094095, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6028595, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251633 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13522, 'inode': 1094099, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.603871, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251641 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1094122, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.610821, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251647 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1094097, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6032066, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251653 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1094074, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5976844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251667 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1094074, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5976844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251674 | orchestrator | skipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1094120, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6104074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251679 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.251685 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1094095, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6028595, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251690 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1094107, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6064646, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251699 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1094097, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6032066, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251705 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094077, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5985248, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251710 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1094097, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6032066, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251723 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094077, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5985248, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251729 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1094095, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6028595, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251734 | orchestrator | skipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1094120, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6104074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251740 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.251745 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1094095, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6028595, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251754 | orchestrator | skipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1094120, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6104074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251760 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.251765 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1094074, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5976844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251771 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1094074, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5976844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251785 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1094088, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.602203, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251791 | orchestrator | skipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1094120, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6104074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251797 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.251803 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1094097, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6032066, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251808 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1094097, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6032066, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251817 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1094082, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5998523, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251822 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1094095, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6028595, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251832 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1094095, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6028595, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251840 | orchestrator | skipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1094120, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6104074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251846 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.251851 | orchestrator | skipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1094120, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6104074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})  2025-08-29 17:48:34.251857 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.251862 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094108, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6077433, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251868 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094073, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5971754, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251877 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1094122, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.610821, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251882 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1094107, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6064646, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251892 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1094077, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5985248, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251900 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5051, 'inode': 1094074, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5976844, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251906 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1094097, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6032066, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251912 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1094095, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6028595, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251917 | orchestrator | changed: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1094120, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.6104074, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) 2025-08-29 17:48:34.251923 | orchestrator | 2025-08-29 17:48:34.251928 | orchestrator | TASK [prometheus : Find prometheus common config overrides] ******************** 2025-08-29 17:48:34.251934 | orchestrator | Friday 29 August 2025 17:46:05 +0000 (0:00:26.192) 0:00:52.939 ********* 2025-08-29 17:48:34.251939 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:48:34.251945 | orchestrator | 2025-08-29 17:48:34.251950 | orchestrator | TASK [prometheus : Find prometheus host config overrides] ********************** 2025-08-29 17:48:34.251956 | orchestrator | Friday 29 August 2025 17:46:06 +0000 (0:00:00.802) 0:00:53.742 ********* 2025-08-29 17:48:34.251965 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.251973 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.251979 | orchestrator | manager/prometheus.yml.d' path due to this access issue: 2025-08-29 17:48:34.251985 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.251990 | orchestrator | manager/prometheus.yml.d' is not a directory 2025-08-29 17:48:34.251995 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:48:34.252001 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.252006 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252011 | orchestrator | node-0/prometheus.yml.d' path due to this access issue: 2025-08-29 17:48:34.252017 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252022 | orchestrator | node-0/prometheus.yml.d' is not a directory 2025-08-29 17:48:34.252028 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:48:34.252033 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.252038 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252044 | orchestrator | node-1/prometheus.yml.d' path due to this access issue: 2025-08-29 17:48:34.252049 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252054 | orchestrator | node-1/prometheus.yml.d' is not a directory 2025-08-29 17:48:34.252060 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.252065 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252070 | orchestrator | node-2/prometheus.yml.d' path due to this access issue: 2025-08-29 17:48:34.252076 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252081 | orchestrator | node-2/prometheus.yml.d' is not a directory 2025-08-29 17:48:34.252087 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.252095 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252100 | orchestrator | node-3/prometheus.yml.d' path due to this access issue: 2025-08-29 17:48:34.252106 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252111 | orchestrator | node-3/prometheus.yml.d' is not a directory 2025-08-29 17:48:34.252116 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.252122 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252127 | orchestrator | node-4/prometheus.yml.d' path due to this access issue: 2025-08-29 17:48:34.252132 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252138 | orchestrator | node-4/prometheus.yml.d' is not a directory 2025-08-29 17:48:34.252143 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.252149 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252154 | orchestrator | node-5/prometheus.yml.d' path due to this access issue: 2025-08-29 17:48:34.252159 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/testbed- 2025-08-29 17:48:34.252165 | orchestrator | node-5/prometheus.yml.d' is not a directory 2025-08-29 17:48:34.252170 | orchestrator | ok: [testbed-node-1 -> localhost] 2025-08-29 17:48:34.252175 | orchestrator | ok: [testbed-node-2 -> localhost] 2025-08-29 17:48:34.252180 | orchestrator | ok: [testbed-node-3 -> localhost] 2025-08-29 17:48:34.252186 | orchestrator | ok: [testbed-node-4 -> localhost] 2025-08-29 17:48:34.252191 | orchestrator | ok: [testbed-node-5 -> localhost] 2025-08-29 17:48:34.252197 | orchestrator | 2025-08-29 17:48:34.252202 | orchestrator | TASK [prometheus : Copying over prometheus config file] ************************ 2025-08-29 17:48:34.252207 | orchestrator | Friday 29 August 2025 17:46:08 +0000 (0:00:01.720) 0:00:55.463 ********* 2025-08-29 17:48:34.252213 | orchestrator | skipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)  2025-08-29 17:48:34.252222 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252228 | orchestrator | skipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)  2025-08-29 17:48:34.252233 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252239 | orchestrator | skipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)  2025-08-29 17:48:34.252244 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252250 | orchestrator | skipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)  2025-08-29 17:48:34.252255 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252260 | orchestrator | skipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)  2025-08-29 17:48:34.252266 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252271 | orchestrator | skipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)  2025-08-29 17:48:34.252277 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252282 | orchestrator | changed: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) 2025-08-29 17:48:34.252287 | orchestrator | 2025-08-29 17:48:34.252293 | orchestrator | TASK [prometheus : Copying over prometheus web config file] ******************** 2025-08-29 17:48:34.252298 | orchestrator | Friday 29 August 2025 17:46:23 +0000 (0:00:15.040) 0:01:10.504 ********* 2025-08-29 17:48:34.252303 | orchestrator | skipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)  2025-08-29 17:48:34.252309 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252314 | orchestrator | skipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)  2025-08-29 17:48:34.252319 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252328 | orchestrator | skipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)  2025-08-29 17:48:34.252333 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252339 | orchestrator | skipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)  2025-08-29 17:48:34.252344 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252349 | orchestrator | skipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)  2025-08-29 17:48:34.252355 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252360 | orchestrator | skipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)  2025-08-29 17:48:34.252365 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252371 | orchestrator | changed: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) 2025-08-29 17:48:34.252376 | orchestrator | 2025-08-29 17:48:34.252381 | orchestrator | TASK [prometheus : Copying over prometheus alertmanager config file] *********** 2025-08-29 17:48:34.252387 | orchestrator | Friday 29 August 2025 17:46:27 +0000 (0:00:04.716) 0:01:15.220 ********* 2025-08-29 17:48:34.252392 | orchestrator | skipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)  2025-08-29 17:48:34.252398 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252403 | orchestrator | skipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)  2025-08-29 17:48:34.252409 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252414 | orchestrator | skipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)  2025-08-29 17:48:34.252420 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252427 | orchestrator | skipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)  2025-08-29 17:48:34.252433 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252439 | orchestrator | skipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)  2025-08-29 17:48:34.252448 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252453 | orchestrator | skipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)  2025-08-29 17:48:34.252458 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252478 | orchestrator | changed: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) 2025-08-29 17:48:34.252484 | orchestrator | 2025-08-29 17:48:34.252489 | orchestrator | TASK [prometheus : Find custom Alertmanager alert notification templates] ****** 2025-08-29 17:48:34.252495 | orchestrator | Friday 29 August 2025 17:46:30 +0000 (0:00:02.398) 0:01:17.618 ********* 2025-08-29 17:48:34.252500 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:48:34.252505 | orchestrator | 2025-08-29 17:48:34.252511 | orchestrator | TASK [prometheus : Copying over custom Alertmanager alert notification templates] *** 2025-08-29 17:48:34.252516 | orchestrator | Friday 29 August 2025 17:46:31 +0000 (0:00:00.892) 0:01:18.511 ********* 2025-08-29 17:48:34.252521 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.252527 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252532 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252537 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252543 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252548 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252553 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252559 | orchestrator | 2025-08-29 17:48:34.252564 | orchestrator | TASK [prometheus : Copying over my.cnf for mysqld_exporter] ******************** 2025-08-29 17:48:34.252569 | orchestrator | Friday 29 August 2025 17:46:31 +0000 (0:00:00.597) 0:01:19.108 ********* 2025-08-29 17:48:34.252575 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.252580 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252585 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252591 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252596 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:48:34.252601 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:48:34.252606 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:48:34.252612 | orchestrator | 2025-08-29 17:48:34.252617 | orchestrator | TASK [prometheus : Copying cloud config file for openstack exporter] *********** 2025-08-29 17:48:34.252622 | orchestrator | Friday 29 August 2025 17:46:34 +0000 (0:00:02.262) 0:01:21.371 ********* 2025-08-29 17:48:34.252628 | orchestrator | skipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2)  2025-08-29 17:48:34.252633 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.252638 | orchestrator | skipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2)  2025-08-29 17:48:34.252644 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252649 | orchestrator | skipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2)  2025-08-29 17:48:34.252654 | orchestrator | skipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2)  2025-08-29 17:48:34.252660 | orchestrator | skipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2)  2025-08-29 17:48:34.252665 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252670 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252676 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252681 | orchestrator | skipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2)  2025-08-29 17:48:34.252686 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252707 | orchestrator | skipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2)  2025-08-29 17:48:34.252713 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252718 | orchestrator | 2025-08-29 17:48:34.252723 | orchestrator | TASK [prometheus : Copying config file for blackbox exporter] ****************** 2025-08-29 17:48:34.252733 | orchestrator | Friday 29 August 2025 17:46:35 +0000 (0:00:01.799) 0:01:23.170 ********* 2025-08-29 17:48:34.252738 | orchestrator | skipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)  2025-08-29 17:48:34.252744 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252749 | orchestrator | skipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)  2025-08-29 17:48:34.252754 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252760 | orchestrator | skipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)  2025-08-29 17:48:34.252765 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252770 | orchestrator | skipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)  2025-08-29 17:48:34.252776 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252781 | orchestrator | skipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)  2025-08-29 17:48:34.252786 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252792 | orchestrator | skipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)  2025-08-29 17:48:34.252797 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252802 | orchestrator | changed: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) 2025-08-29 17:48:34.252808 | orchestrator | 2025-08-29 17:48:34.252816 | orchestrator | TASK [prometheus : Find extra prometheus server config files] ****************** 2025-08-29 17:48:34.252822 | orchestrator | Friday 29 August 2025 17:46:37 +0000 (0:00:01.742) 0:01:24.913 ********* 2025-08-29 17:48:34.252827 | orchestrator | [WARNING]: Skipped 2025-08-29 17:48:34.252832 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path 2025-08-29 17:48:34.252838 | orchestrator | due to this access issue: 2025-08-29 17:48:34.252843 | orchestrator | '/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is 2025-08-29 17:48:34.252848 | orchestrator | not a directory 2025-08-29 17:48:34.252854 | orchestrator | ok: [testbed-manager -> localhost] 2025-08-29 17:48:34.252859 | orchestrator | 2025-08-29 17:48:34.252864 | orchestrator | TASK [prometheus : Create subdirectories for extra config files] *************** 2025-08-29 17:48:34.252870 | orchestrator | Friday 29 August 2025 17:46:38 +0000 (0:00:01.312) 0:01:26.226 ********* 2025-08-29 17:48:34.252875 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.252881 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252886 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252891 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252896 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252902 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252907 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252912 | orchestrator | 2025-08-29 17:48:34.252918 | orchestrator | TASK [prometheus : Template extra prometheus server config files] ************** 2025-08-29 17:48:34.252923 | orchestrator | Friday 29 August 2025 17:46:40 +0000 (0:00:01.576) 0:01:27.802 ********* 2025-08-29 17:48:34.252929 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.252934 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:48:34.252939 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:48:34.252945 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:48:34.252950 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:48:34.252955 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:48:34.252960 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:48:34.252971 | orchestrator | 2025-08-29 17:48:34.252981 | orchestrator | TASK [prometheus : Check prometheus containers] ******************************** 2025-08-29 17:48:34.252986 | orchestrator | Friday 29 August 2025 17:46:41 +0000 (0:00:00.775) 0:01:28.578 ********* 2025-08-29 17:48:34.252992 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.253001 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.253010 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) 2025-08-29 17:48:34.253016 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.253025 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.253031 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.253037 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.253046 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) 2025-08-29 17:48:34.253052 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253063 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253069 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253075 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253084 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253090 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253095 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253104 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253110 | orchestrator | changed: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253116 | orchestrator | changed: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253121 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253159 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) 2025-08-29 17:48:34.253166 | orchestrator | changed: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253175 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253181 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253187 | orchestrator | changed: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253195 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253201 | orchestrator | changed: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253211 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) 2025-08-29 17:48:34.253217 | orchestrator | changed: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253227 | orchestrator | changed: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) 2025-08-29 17:48:34.253232 | orchestrator | 2025-08-29 17:48:34.253238 | orchestrator | TASK [prometheus : Creating prometheus database user and setting permissions] *** 2025-08-29 17:48:34.253243 | orchestrator | Friday 29 August 2025 17:46:45 +0000 (0:00:04.707) 0:01:33.285 ********* 2025-08-29 17:48:34.253249 | orchestrator | skipping: [testbed-manager] => (item=testbed-node-0)  2025-08-29 17:48:34.253254 | orchestrator | skipping: [testbed-manager] 2025-08-29 17:48:34.253260 | orchestrator | 2025-08-29 17:48:34.253265 | orchestrator | TASK [prometheus : Flush handlers] ********************************************* 2025-08-29 17:48:34.253270 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:01.181) 0:01:34.467 ********* 2025-08-29 17:48:34.253276 | orchestrator | 2025-08-29 17:48:34.253281 | orchestrator | TASK [prometheus : Flush handlers] ********************************************* 2025-08-29 17:48:34.253287 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:00.074) 0:01:34.541 ********* 2025-08-29 17:48:34.253292 | orchestrator | 2025-08-29 17:48:34.253297 | orchestrator | TASK [prometheus : Flush handlers] ********************************************* 2025-08-29 17:48:34.253303 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:00.062) 0:01:34.604 ********* 2025-08-29 17:48:34.253308 | orchestrator | 2025-08-29 17:48:34.253313 | orchestrator | TASK [prometheus : Flush handlers] ********************************************* 2025-08-29 17:48:34.253319 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:00.218) 0:01:34.823 ********* 2025-08-29 17:48:34.253324 | orchestrator | 2025-08-29 17:48:34.253329 | orchestrator | TASK [prometheus : Flush handlers] ********************************************* 2025-08-29 17:48:34.253335 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:00.063) 0:01:34.887 ********* 2025-08-29 17:48:34.253340 | orchestrator | 2025-08-29 17:48:34.253345 | orchestrator | TASK [prometheus : Flush handlers] ********************************************* 2025-08-29 17:48:34.253351 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:00.076) 0:01:34.963 ********* 2025-08-29 17:48:34.253356 | orchestrator | 2025-08-29 17:48:34.253361 | orchestrator | TASK [prometheus : Flush handlers] ********************************************* 2025-08-29 17:48:34.253367 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:00.064) 0:01:35.027 ********* 2025-08-29 17:48:34.253372 | orchestrator | 2025-08-29 17:48:34.253380 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-server container] ************* 2025-08-29 17:48:34.253386 | orchestrator | Friday 29 August 2025 17:46:47 +0000 (0:00:00.083) 0:01:35.111 ********* 2025-08-29 17:48:34.253391 | orchestrator | changed: [testbed-manager] 2025-08-29 17:48:34.253396 | orchestrator | 2025-08-29 17:48:34.253402 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ****** 2025-08-29 17:48:34.253407 | orchestrator | Friday 29 August 2025 17:47:10 +0000 (0:00:22.469) 0:01:57.580 ********* 2025-08-29 17:48:34.253413 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:48:34.253418 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:48:34.253423 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:48:34.253429 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:48:34.253434 | orchestrator | changed: [testbed-manager] 2025-08-29 17:48:34.253439 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:48:34.253445 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:48:34.253450 | orchestrator | 2025-08-29 17:48:34.253456 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] **** 2025-08-29 17:48:34.253461 | orchestrator | Friday 29 August 2025 17:47:25 +0000 (0:00:14.980) 0:02:12.561 ********* 2025-08-29 17:48:34.253506 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:48:34.253512 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:48:34.253522 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:48:34.253527 | orchestrator | 2025-08-29 17:48:34.253532 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] *** 2025-08-29 17:48:34.253538 | orchestrator | Friday 29 August 2025 17:47:31 +0000 (0:00:06.534) 0:02:19.096 ********* 2025-08-29 17:48:34.253543 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:48:34.253549 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:48:34.253554 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:48:34.253560 | orchestrator | 2025-08-29 17:48:34.253565 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] *********** 2025-08-29 17:48:34.253571 | orchestrator | Friday 29 August 2025 17:47:42 +0000 (0:00:10.778) 0:02:29.875 ********* 2025-08-29 17:48:34.253576 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:48:34.253581 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:48:34.253590 | orchestrator | changed: [testbed-manager] 2025-08-29 17:48:34.253595 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:48:34.253601 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:48:34.253606 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:48:34.253612 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:48:34.253617 | orchestrator | 2025-08-29 17:48:34.253622 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] ******* 2025-08-29 17:48:34.253628 | orchestrator | Friday 29 August 2025 17:48:00 +0000 (0:00:18.053) 0:02:47.928 ********* 2025-08-29 17:48:34.253633 | orchestrator | changed: [testbed-manager] 2025-08-29 17:48:34.253639 | orchestrator | 2025-08-29 17:48:34.253644 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] *** 2025-08-29 17:48:34.253649 | orchestrator | Friday 29 August 2025 17:48:17 +0000 (0:00:16.918) 0:03:04.847 ********* 2025-08-29 17:48:34.253655 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:48:34.253660 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:48:34.253666 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:48:34.253671 | orchestrator | 2025-08-29 17:48:34.253676 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] *** 2025-08-29 17:48:34.253682 | orchestrator | Friday 29 August 2025 17:48:22 +0000 (0:00:05.261) 0:03:10.108 ********* 2025-08-29 17:48:34.253687 | orchestrator | changed: [testbed-manager] 2025-08-29 17:48:34.253692 | orchestrator | 2025-08-29 17:48:34.253698 | orchestrator | RUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] *** 2025-08-29 17:48:34.253703 | orchestrator | Friday 29 August 2025 17:48:27 +0000 (0:00:05.119) 0:03:15.228 ********* 2025-08-29 17:48:34.253709 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:48:34.253714 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:48:34.253719 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:48:34.253725 | orchestrator | 2025-08-29 17:48:34.253730 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:48:34.253736 | orchestrator | testbed-manager : ok=23  changed=14  unreachable=0 failed=0 skipped=8  rescued=0 ignored=0 2025-08-29 17:48:34.253742 | orchestrator | testbed-node-0 : ok=15  changed=10  unreachable=0 failed=0 skipped=11  rescued=0 ignored=0 2025-08-29 17:48:34.253747 | orchestrator | testbed-node-1 : ok=15  changed=10  unreachable=0 failed=0 skipped=11  rescued=0 ignored=0 2025-08-29 17:48:34.253753 | orchestrator | testbed-node-2 : ok=15  changed=10  unreachable=0 failed=0 skipped=11  rescued=0 ignored=0 2025-08-29 17:48:34.253758 | orchestrator | testbed-node-3 : ok=12  changed=7  unreachable=0 failed=0 skipped=12  rescued=0 ignored=0 2025-08-29 17:48:34.253764 | orchestrator | testbed-node-4 : ok=12  changed=7  unreachable=0 failed=0 skipped=12  rescued=0 ignored=0 2025-08-29 17:48:34.253773 | orchestrator | testbed-node-5 : ok=12  changed=7  unreachable=0 failed=0 skipped=12  rescued=0 ignored=0 2025-08-29 17:48:34.253778 | orchestrator | 2025-08-29 17:48:34.253784 | orchestrator | 2025-08-29 17:48:34.253789 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:48:34.253795 | orchestrator | Friday 29 August 2025 17:48:33 +0000 (0:00:05.329) 0:03:20.557 ********* 2025-08-29 17:48:34.253800 | orchestrator | =============================================================================== 2025-08-29 17:48:34.253806 | orchestrator | prometheus : Copying over custom prometheus alert rules files ---------- 26.19s 2025-08-29 17:48:34.253814 | orchestrator | prometheus : Restart prometheus-server container ----------------------- 22.47s 2025-08-29 17:48:34.253820 | orchestrator | prometheus : Restart prometheus-cadvisor container --------------------- 18.05s 2025-08-29 17:48:34.253825 | orchestrator | prometheus : Restart prometheus-alertmanager container ----------------- 16.92s 2025-08-29 17:48:34.253830 | orchestrator | prometheus : Copying over prometheus config file ----------------------- 15.04s 2025-08-29 17:48:34.253836 | orchestrator | prometheus : Restart prometheus-node-exporter container ---------------- 14.98s 2025-08-29 17:48:34.253841 | orchestrator | prometheus : Restart prometheus-memcached-exporter container ----------- 10.78s 2025-08-29 17:48:34.253847 | orchestrator | prometheus : Restart prometheus-mysqld-exporter container --------------- 6.53s 2025-08-29 17:48:34.253852 | orchestrator | prometheus : Copying over config.json files ----------------------------- 6.35s 2025-08-29 17:48:34.253857 | orchestrator | service-cert-copy : prometheus | Copying over extra CA certificates ----- 5.90s 2025-08-29 17:48:34.253863 | orchestrator | prometheus : Restart prometheus-libvirt-exporter container -------------- 5.33s 2025-08-29 17:48:34.253868 | orchestrator | prometheus : Restart prometheus-elasticsearch-exporter container -------- 5.26s 2025-08-29 17:48:34.253874 | orchestrator | prometheus : Restart prometheus-blackbox-exporter container ------------- 5.12s 2025-08-29 17:48:34.253879 | orchestrator | prometheus : Copying over prometheus web config file -------------------- 4.72s 2025-08-29 17:48:34.253884 | orchestrator | prometheus : Check prometheus containers -------------------------------- 4.71s 2025-08-29 17:48:34.253890 | orchestrator | prometheus : Ensuring config directories exist -------------------------- 3.66s 2025-08-29 17:48:34.253895 | orchestrator | service-cert-copy : prometheus | Copying over backend internal TLS key --- 3.06s 2025-08-29 17:48:34.253901 | orchestrator | prometheus : Copying over prometheus alertmanager config file ----------- 2.40s 2025-08-29 17:48:34.253909 | orchestrator | prometheus : Copying over my.cnf for mysqld_exporter -------------------- 2.26s 2025-08-29 17:48:34.253914 | orchestrator | service-cert-copy : prometheus | Copying over backend internal TLS certificate --- 1.96s 2025-08-29 17:48:34.253920 | orchestrator | 2025-08-29 17:48:34 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:34.253925 | orchestrator | 2025-08-29 17:48:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:37.305147 | orchestrator | 2025-08-29 17:48:37 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:37.307037 | orchestrator | 2025-08-29 17:48:37 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:37.309569 | orchestrator | 2025-08-29 17:48:37 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:37.312024 | orchestrator | 2025-08-29 17:48:37 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:37.312332 | orchestrator | 2025-08-29 17:48:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:40.358677 | orchestrator | 2025-08-29 17:48:40 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:40.359382 | orchestrator | 2025-08-29 17:48:40 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:40.360507 | orchestrator | 2025-08-29 17:48:40 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:40.361595 | orchestrator | 2025-08-29 17:48:40 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:40.361623 | orchestrator | 2025-08-29 17:48:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:43.400915 | orchestrator | 2025-08-29 17:48:43 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:43.401015 | orchestrator | 2025-08-29 17:48:43 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:43.401366 | orchestrator | 2025-08-29 17:48:43 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:43.402248 | orchestrator | 2025-08-29 17:48:43 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:43.402271 | orchestrator | 2025-08-29 17:48:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:46.448281 | orchestrator | 2025-08-29 17:48:46 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:46.448905 | orchestrator | 2025-08-29 17:48:46 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:46.449830 | orchestrator | 2025-08-29 17:48:46 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:46.452839 | orchestrator | 2025-08-29 17:48:46 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:46.453123 | orchestrator | 2025-08-29 17:48:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:49.493942 | orchestrator | 2025-08-29 17:48:49 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:49.494110 | orchestrator | 2025-08-29 17:48:49 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:49.495140 | orchestrator | 2025-08-29 17:48:49 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:49.495290 | orchestrator | 2025-08-29 17:48:49 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:49.495310 | orchestrator | 2025-08-29 17:48:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:52.545810 | orchestrator | 2025-08-29 17:48:52 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:52.545911 | orchestrator | 2025-08-29 17:48:52 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:52.548769 | orchestrator | 2025-08-29 17:48:52 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:52.551023 | orchestrator | 2025-08-29 17:48:52 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:52.551135 | orchestrator | 2025-08-29 17:48:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:55.613249 | orchestrator | 2025-08-29 17:48:55 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:55.614213 | orchestrator | 2025-08-29 17:48:55 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:55.616323 | orchestrator | 2025-08-29 17:48:55 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:55.617288 | orchestrator | 2025-08-29 17:48:55 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:55.617801 | orchestrator | 2025-08-29 17:48:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:48:58.658789 | orchestrator | 2025-08-29 17:48:58 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:48:58.658896 | orchestrator | 2025-08-29 17:48:58 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:48:58.660539 | orchestrator | 2025-08-29 17:48:58 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:48:58.662125 | orchestrator | 2025-08-29 17:48:58 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:48:58.662165 | orchestrator | 2025-08-29 17:48:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:01.715170 | orchestrator | 2025-08-29 17:49:01 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:01.715276 | orchestrator | 2025-08-29 17:49:01 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:49:01.715975 | orchestrator | 2025-08-29 17:49:01 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:01.717523 | orchestrator | 2025-08-29 17:49:01 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:01.717548 | orchestrator | 2025-08-29 17:49:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:04.769049 | orchestrator | 2025-08-29 17:49:04 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:04.771531 | orchestrator | 2025-08-29 17:49:04 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:49:04.773385 | orchestrator | 2025-08-29 17:49:04 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:04.774956 | orchestrator | 2025-08-29 17:49:04 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:04.775266 | orchestrator | 2025-08-29 17:49:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:07.819435 | orchestrator | 2025-08-29 17:49:07 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:07.820895 | orchestrator | 2025-08-29 17:49:07 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:49:07.822781 | orchestrator | 2025-08-29 17:49:07 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:07.824236 | orchestrator | 2025-08-29 17:49:07 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:07.824259 | orchestrator | 2025-08-29 17:49:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:10.870557 | orchestrator | 2025-08-29 17:49:10 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:10.873140 | orchestrator | 2025-08-29 17:49:10 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:49:10.875263 | orchestrator | 2025-08-29 17:49:10 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:10.877767 | orchestrator | 2025-08-29 17:49:10 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:10.878120 | orchestrator | 2025-08-29 17:49:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:13.936915 | orchestrator | 2025-08-29 17:49:13 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:13.938308 | orchestrator | 2025-08-29 17:49:13 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:49:13.940359 | orchestrator | 2025-08-29 17:49:13 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:13.942091 | orchestrator | 2025-08-29 17:49:13 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:13.942147 | orchestrator | 2025-08-29 17:49:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:16.983387 | orchestrator | 2025-08-29 17:49:16 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:16.986774 | orchestrator | 2025-08-29 17:49:16 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state STARTED 2025-08-29 17:49:16.988670 | orchestrator | 2025-08-29 17:49:16 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:16.990776 | orchestrator | 2025-08-29 17:49:16 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:16.991144 | orchestrator | 2025-08-29 17:49:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:20.031855 | orchestrator | 2025-08-29 17:49:20 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:20.033145 | orchestrator | 2025-08-29 17:49:20 | INFO  | Task 9b3a6d71-051c-4291-9115-742c16137196 is in state SUCCESS 2025-08-29 17:49:20.034394 | orchestrator | 2025-08-29 17:49:20.034428 | orchestrator | 2025-08-29 17:49:20.034440 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:49:20.034452 | orchestrator | 2025-08-29 17:49:20.034488 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:49:20.034501 | orchestrator | Friday 29 August 2025 17:46:33 +0000 (0:00:00.266) 0:00:00.266 ********* 2025-08-29 17:49:20.034512 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:49:20.034524 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:49:20.034535 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:49:20.034546 | orchestrator | 2025-08-29 17:49:20.034557 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:49:20.034568 | orchestrator | Friday 29 August 2025 17:46:33 +0000 (0:00:00.280) 0:00:00.547 ********* 2025-08-29 17:49:20.034579 | orchestrator | ok: [testbed-node-0] => (item=enable_glance_True) 2025-08-29 17:49:20.034590 | orchestrator | ok: [testbed-node-1] => (item=enable_glance_True) 2025-08-29 17:49:20.034601 | orchestrator | ok: [testbed-node-2] => (item=enable_glance_True) 2025-08-29 17:49:20.034612 | orchestrator | 2025-08-29 17:49:20.034623 | orchestrator | PLAY [Apply role glance] ******************************************************* 2025-08-29 17:49:20.034634 | orchestrator | 2025-08-29 17:49:20.034822 | orchestrator | TASK [glance : include_tasks] ************************************************** 2025-08-29 17:49:20.034846 | orchestrator | Friday 29 August 2025 17:46:34 +0000 (0:00:00.377) 0:00:00.924 ********* 2025-08-29 17:49:20.034866 | orchestrator | included: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:49:20.034885 | orchestrator | 2025-08-29 17:49:20.034903 | orchestrator | TASK [service-ks-register : glance | Creating services] ************************ 2025-08-29 17:49:20.034914 | orchestrator | Friday 29 August 2025 17:46:34 +0000 (0:00:00.593) 0:00:01.518 ********* 2025-08-29 17:49:20.034925 | orchestrator | changed: [testbed-node-0] => (item=glance (image)) 2025-08-29 17:49:20.034936 | orchestrator | 2025-08-29 17:49:20.034947 | orchestrator | TASK [service-ks-register : glance | Creating endpoints] *********************** 2025-08-29 17:49:20.034958 | orchestrator | Friday 29 August 2025 17:46:38 +0000 (0:00:03.825) 0:00:05.344 ********* 2025-08-29 17:49:20.034969 | orchestrator | changed: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal) 2025-08-29 17:49:20.034980 | orchestrator | changed: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public) 2025-08-29 17:49:20.034991 | orchestrator | 2025-08-29 17:49:20.035002 | orchestrator | TASK [service-ks-register : glance | Creating projects] ************************ 2025-08-29 17:49:20.035013 | orchestrator | Friday 29 August 2025 17:46:45 +0000 (0:00:07.315) 0:00:12.660 ********* 2025-08-29 17:49:20.035024 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:49:20.035035 | orchestrator | 2025-08-29 17:49:20.035046 | orchestrator | TASK [service-ks-register : glance | Creating users] *************************** 2025-08-29 17:49:20.035080 | orchestrator | Friday 29 August 2025 17:46:49 +0000 (0:00:03.454) 0:00:16.115 ********* 2025-08-29 17:49:20.035092 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:49:20.035103 | orchestrator | changed: [testbed-node-0] => (item=glance -> service) 2025-08-29 17:49:20.035114 | orchestrator | 2025-08-29 17:49:20.035125 | orchestrator | TASK [service-ks-register : glance | Creating roles] *************************** 2025-08-29 17:49:20.035136 | orchestrator | Friday 29 August 2025 17:46:53 +0000 (0:00:03.992) 0:00:20.107 ********* 2025-08-29 17:49:20.035147 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:49:20.035158 | orchestrator | 2025-08-29 17:49:20.035169 | orchestrator | TASK [service-ks-register : glance | Granting user roles] ********************** 2025-08-29 17:49:20.035179 | orchestrator | Friday 29 August 2025 17:46:56 +0000 (0:00:03.407) 0:00:23.514 ********* 2025-08-29 17:49:20.035190 | orchestrator | changed: [testbed-node-0] => (item=glance -> service -> admin) 2025-08-29 17:49:20.035201 | orchestrator | 2025-08-29 17:49:20.035211 | orchestrator | TASK [glance : Ensuring config directories exist] ****************************** 2025-08-29 17:49:20.035222 | orchestrator | Friday 29 August 2025 17:47:00 +0000 (0:00:03.963) 0:00:27.478 ********* 2025-08-29 17:49:20.035251 | orchestrator | changed: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.035268 | orchestrator | changed: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.035290 | orchestrator | changed: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.035303 | orchestrator | 2025-08-29 17:49:20.035314 | orchestrator | TASK [glance : include_tasks] ************************************************** 2025-08-29 17:49:20.035325 | orchestrator | Friday 29 August 2025 17:47:04 +0000 (0:00:03.458) 0:00:30.937 ********* 2025-08-29 17:49:20.035344 | orchestrator | included: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:49:20.035356 | orchestrator | 2025-08-29 17:49:20.035367 | orchestrator | TASK [glance : Ensuring glance service ceph config subdir exists] ************** 2025-08-29 17:49:20.035377 | orchestrator | Friday 29 August 2025 17:47:04 +0000 (0:00:00.682) 0:00:31.620 ********* 2025-08-29 17:49:20.035388 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:49:20.035399 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.035410 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:49:20.035420 | orchestrator | 2025-08-29 17:49:20.035433 | orchestrator | TASK [glance : Copy over multiple ceph configs for Glance] ********************* 2025-08-29 17:49:20.035445 | orchestrator | Friday 29 August 2025 17:47:08 +0000 (0:00:03.454) 0:00:35.074 ********* 2025-08-29 17:49:20.035476 | orchestrator | changed: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:49:20.035490 | orchestrator | changed: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:49:20.035503 | orchestrator | changed: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:49:20.035516 | orchestrator | 2025-08-29 17:49:20.035528 | orchestrator | TASK [glance : Copy over ceph Glance keyrings] ********************************* 2025-08-29 17:49:20.035540 | orchestrator | Friday 29 August 2025 17:47:09 +0000 (0:00:01.538) 0:00:36.612 ********* 2025-08-29 17:49:20.035561 | orchestrator | changed: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:49:20.035574 | orchestrator | changed: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:49:20.035587 | orchestrator | changed: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:49:20.035599 | orchestrator | 2025-08-29 17:49:20.035611 | orchestrator | TASK [glance : Ensuring config directory has correct owner and permission] ***** 2025-08-29 17:49:20.035624 | orchestrator | Friday 29 August 2025 17:47:11 +0000 (0:00:01.341) 0:00:37.953 ********* 2025-08-29 17:49:20.035637 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:49:20.035649 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:49:20.035662 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:49:20.035675 | orchestrator | 2025-08-29 17:49:20.035687 | orchestrator | TASK [glance : Check if policies shall be overwritten] ************************* 2025-08-29 17:49:20.035700 | orchestrator | Friday 29 August 2025 17:47:12 +0000 (0:00:01.310) 0:00:39.264 ********* 2025-08-29 17:49:20.035713 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.035725 | orchestrator | 2025-08-29 17:49:20.035737 | orchestrator | TASK [glance : Set glance policy file] ***************************************** 2025-08-29 17:49:20.035750 | orchestrator | Friday 29 August 2025 17:47:12 +0000 (0:00:00.313) 0:00:39.578 ********* 2025-08-29 17:49:20.035763 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.035774 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.035785 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.035796 | orchestrator | 2025-08-29 17:49:20.035807 | orchestrator | TASK [glance : include_tasks] ************************************************** 2025-08-29 17:49:20.035818 | orchestrator | Friday 29 August 2025 17:47:13 +0000 (0:00:00.706) 0:00:40.284 ********* 2025-08-29 17:49:20.035930 | orchestrator | included: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:49:20.035942 | orchestrator | 2025-08-29 17:49:20.035953 | orchestrator | TASK [service-cert-copy : glance | Copying over extra CA certificates] ********* 2025-08-29 17:49:20.035964 | orchestrator | Friday 29 August 2025 17:47:14 +0000 (0:00:01.183) 0:00:41.467 ********* 2025-08-29 17:49:20.035986 | orchestrator | changed: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.036008 | orchestrator | changed: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.036021 | orchestrator | changed: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.036033 | orchestrator | 2025-08-29 17:49:20.036117 | orchestrator | TASK [service-cert-copy : glance | Copying over backend internal TLS certificate] *** 2025-08-29 17:49:20.036130 | orchestrator | Friday 29 August 2025 17:47:18 +0000 (0:00:03.721) 0:00:45.189 ********* 2025-08-29 17:49:20.036152 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:49:20.036172 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036185 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:49:20.036197 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036217 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:49:20.036235 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036246 | orchestrator | 2025-08-29 17:49:20.036257 | orchestrator | TASK [service-cert-copy : glance | Copying over backend internal TLS key] ****** 2025-08-29 17:49:20.036268 | orchestrator | Friday 29 August 2025 17:47:20 +0000 (0:00:02.456) 0:00:47.645 ********* 2025-08-29 17:49:20.036280 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:49:20.036292 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036309 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:49:20.036329 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036341 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})  2025-08-29 17:49:20.036353 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036364 | orchestrator | 2025-08-29 17:49:20.036374 | orchestrator | TASK [glance : Creating TLS backend PEM File] ********************************** 2025-08-29 17:49:20.036386 | orchestrator | Friday 29 August 2025 17:47:24 +0000 (0:00:03.195) 0:00:50.841 ********* 2025-08-29 17:49:20.036397 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036408 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036419 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036429 | orchestrator | 2025-08-29 17:49:20.036440 | orchestrator | TASK [glance : Copying over config.json files for services] ******************** 2025-08-29 17:49:20.036451 | orchestrator | Friday 29 August 2025 17:47:29 +0000 (0:00:05.476) 0:00:56.318 ********* 2025-08-29 17:49:20.036506 | orchestrator | changed: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.036529 | orchestrator | changed: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.036543 | orchestrator | changed: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.036561 | orchestrator | 2025-08-29 17:49:20.036572 | orchestrator | TASK [glance : Copying over glance-api.conf] *********************************** 2025-08-29 17:49:20.036583 | orchestrator | Friday 29 August 2025 17:47:34 +0000 (0:00:04.468) 0:01:00.787 ********* 2025-08-29 17:49:20.036594 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:49:20.036605 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.036615 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:49:20.036626 | orchestrator | 2025-08-29 17:49:20.036637 | orchestrator | TASK [glance : Copying over glance-cache.conf for glance_api] ****************** 2025-08-29 17:49:20.036653 | orchestrator | Friday 29 August 2025 17:47:39 +0000 (0:00:05.218) 0:01:06.005 ********* 2025-08-29 17:49:20.036665 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036676 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036686 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036698 | orchestrator | 2025-08-29 17:49:20.036709 | orchestrator | TASK [glance : Copying over glance-swift.conf for glance_api] ****************** 2025-08-29 17:49:20.036720 | orchestrator | Friday 29 August 2025 17:47:42 +0000 (0:00:03.159) 0:01:09.165 ********* 2025-08-29 17:49:20.036731 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036742 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036755 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036768 | orchestrator | 2025-08-29 17:49:20.036780 | orchestrator | TASK [glance : Copying over glance-image-import.conf] ************************** 2025-08-29 17:49:20.036793 | orchestrator | Friday 29 August 2025 17:47:50 +0000 (0:00:07.822) 0:01:16.987 ********* 2025-08-29 17:49:20.036805 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036817 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036830 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036842 | orchestrator | 2025-08-29 17:49:20.036855 | orchestrator | TASK [glance : Copying over property-protections-rules.conf] ******************* 2025-08-29 17:49:20.036867 | orchestrator | Friday 29 August 2025 17:47:54 +0000 (0:00:04.489) 0:01:21.477 ********* 2025-08-29 17:49:20.036879 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036891 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036904 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036917 | orchestrator | 2025-08-29 17:49:20.036929 | orchestrator | TASK [glance : Copying over existing policy file] ****************************** 2025-08-29 17:49:20.036941 | orchestrator | Friday 29 August 2025 17:47:58 +0000 (0:00:03.559) 0:01:25.036 ********* 2025-08-29 17:49:20.036954 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.036966 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.036978 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.036991 | orchestrator | 2025-08-29 17:49:20.037003 | orchestrator | TASK [glance : Copying over glance-haproxy-tls.cfg] **************************** 2025-08-29 17:49:20.037016 | orchestrator | Friday 29 August 2025 17:47:58 +0000 (0:00:00.405) 0:01:25.442 ********* 2025-08-29 17:49:20.037028 | orchestrator | skipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2)  2025-08-29 17:49:20.037041 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.037053 | orchestrator | skipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2)  2025-08-29 17:49:20.037066 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.037079 | orchestrator | skipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2)  2025-08-29 17:49:20.037092 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.037105 | orchestrator | 2025-08-29 17:49:20.037116 | orchestrator | TASK [glance : Check glance containers] **************************************** 2025-08-29 17:49:20.037133 | orchestrator | Friday 29 August 2025 17:48:04 +0000 (0:00:06.147) 0:01:31.590 ********* 2025-08-29 17:49:20.037146 | orchestrator | changed: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.037167 | orchestrator | changed: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.037180 | orchestrator | changed: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/release/glance-api:29.0.1.20250711', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', 'iscsi_info:/etc/iscsi', '/dev:/dev'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) 2025-08-29 17:49:20.037198 | orchestrator | 2025-08-29 17:49:20.037209 | orchestrator | TASK [glance : include_tasks] ************************************************** 2025-08-29 17:49:20.037220 | orchestrator | Friday 29 August 2025 17:48:08 +0000 (0:00:03.985) 0:01:35.575 ********* 2025-08-29 17:49:20.037231 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:49:20.037242 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:49:20.037252 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:49:20.037263 | orchestrator | 2025-08-29 17:49:20.037274 | orchestrator | TASK [glance : Creating Glance database] *************************************** 2025-08-29 17:49:20.037285 | orchestrator | Friday 29 August 2025 17:48:09 +0000 (0:00:00.283) 0:01:35.858 ********* 2025-08-29 17:49:20.037295 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.037306 | orchestrator | 2025-08-29 17:49:20.037317 | orchestrator | TASK [glance : Creating Glance database user and setting permissions] ********** 2025-08-29 17:49:20.037328 | orchestrator | Friday 29 August 2025 17:48:11 +0000 (0:00:02.089) 0:01:37.947 ********* 2025-08-29 17:49:20.037338 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.037349 | orchestrator | 2025-08-29 17:49:20.037360 | orchestrator | TASK [glance : Enable log_bin_trust_function_creators function] **************** 2025-08-29 17:49:20.037371 | orchestrator | Friday 29 August 2025 17:48:13 +0000 (0:00:02.195) 0:01:40.143 ********* 2025-08-29 17:49:20.037381 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.037392 | orchestrator | 2025-08-29 17:49:20.037403 | orchestrator | TASK [glance : Running Glance bootstrap container] ***************************** 2025-08-29 17:49:20.037419 | orchestrator | Friday 29 August 2025 17:48:15 +0000 (0:00:02.161) 0:01:42.305 ********* 2025-08-29 17:49:20.037430 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.037441 | orchestrator | 2025-08-29 17:49:20.037452 | orchestrator | TASK [glance : Disable log_bin_trust_function_creators function] *************** 2025-08-29 17:49:20.037508 | orchestrator | Friday 29 August 2025 17:48:46 +0000 (0:00:30.584) 0:02:12.890 ********* 2025-08-29 17:49:20.037520 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.037531 | orchestrator | 2025-08-29 17:49:20.037542 | orchestrator | TASK [glance : Flush handlers] ************************************************* 2025-08-29 17:49:20.037553 | orchestrator | Friday 29 August 2025 17:48:48 +0000 (0:00:02.079) 0:02:14.970 ********* 2025-08-29 17:49:20.037564 | orchestrator | 2025-08-29 17:49:20.037576 | orchestrator | TASK [glance : Flush handlers] ************************************************* 2025-08-29 17:49:20.037586 | orchestrator | Friday 29 August 2025 17:48:48 +0000 (0:00:00.337) 0:02:15.308 ********* 2025-08-29 17:49:20.037597 | orchestrator | 2025-08-29 17:49:20.037608 | orchestrator | TASK [glance : Flush handlers] ************************************************* 2025-08-29 17:49:20.037619 | orchestrator | Friday 29 August 2025 17:48:48 +0000 (0:00:00.073) 0:02:15.381 ********* 2025-08-29 17:49:20.037636 | orchestrator | 2025-08-29 17:49:20.037648 | orchestrator | RUNNING HANDLER [glance : Restart glance-api container] ************************ 2025-08-29 17:49:20.037659 | orchestrator | Friday 29 August 2025 17:48:48 +0000 (0:00:00.072) 0:02:15.453 ********* 2025-08-29 17:49:20.037669 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:49:20.037680 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:49:20.037691 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:49:20.037702 | orchestrator | 2025-08-29 17:49:20.037713 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:49:20.037725 | orchestrator | testbed-node-0 : ok=26  changed=18  unreachable=0 failed=0 skipped=12  rescued=0 ignored=0 2025-08-29 17:49:20.037737 | orchestrator | testbed-node-1 : ok=15  changed=9  unreachable=0 failed=0 skipped=11  rescued=0 ignored=0 2025-08-29 17:49:20.037748 | orchestrator | testbed-node-2 : ok=15  changed=9  unreachable=0 failed=0 skipped=11  rescued=0 ignored=0 2025-08-29 17:49:20.037759 | orchestrator | 2025-08-29 17:49:20.037770 | orchestrator | 2025-08-29 17:49:20.037781 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:49:20.037792 | orchestrator | Friday 29 August 2025 17:49:19 +0000 (0:00:30.761) 0:02:46.215 ********* 2025-08-29 17:49:20.037803 | orchestrator | =============================================================================== 2025-08-29 17:49:20.037814 | orchestrator | glance : Restart glance-api container ---------------------------------- 30.76s 2025-08-29 17:49:20.037825 | orchestrator | glance : Running Glance bootstrap container ---------------------------- 30.59s 2025-08-29 17:49:20.037836 | orchestrator | glance : Copying over glance-swift.conf for glance_api ------------------ 7.82s 2025-08-29 17:49:20.037846 | orchestrator | service-ks-register : glance | Creating endpoints ----------------------- 7.32s 2025-08-29 17:49:20.037855 | orchestrator | glance : Copying over glance-haproxy-tls.cfg ---------------------------- 6.15s 2025-08-29 17:49:20.037865 | orchestrator | glance : Creating TLS backend PEM File ---------------------------------- 5.48s 2025-08-29 17:49:20.037874 | orchestrator | glance : Copying over glance-api.conf ----------------------------------- 5.22s 2025-08-29 17:49:20.037884 | orchestrator | glance : Copying over glance-image-import.conf -------------------------- 4.49s 2025-08-29 17:49:20.037894 | orchestrator | glance : Copying over config.json files for services -------------------- 4.47s 2025-08-29 17:49:20.037903 | orchestrator | service-ks-register : glance | Creating users --------------------------- 3.99s 2025-08-29 17:49:20.037913 | orchestrator | glance : Check glance containers ---------------------------------------- 3.99s 2025-08-29 17:49:20.037923 | orchestrator | service-ks-register : glance | Granting user roles ---------------------- 3.96s 2025-08-29 17:49:20.037932 | orchestrator | service-ks-register : glance | Creating services ------------------------ 3.83s 2025-08-29 17:49:20.037942 | orchestrator | service-cert-copy : glance | Copying over extra CA certificates --------- 3.72s 2025-08-29 17:49:20.037952 | orchestrator | glance : Copying over property-protections-rules.conf ------------------- 3.56s 2025-08-29 17:49:20.037961 | orchestrator | glance : Ensuring config directories exist ------------------------------ 3.46s 2025-08-29 17:49:20.037971 | orchestrator | service-ks-register : glance | Creating projects ------------------------ 3.45s 2025-08-29 17:49:20.037981 | orchestrator | glance : Ensuring glance service ceph config subdir exists -------------- 3.45s 2025-08-29 17:49:20.037990 | orchestrator | service-ks-register : glance | Creating roles --------------------------- 3.41s 2025-08-29 17:49:20.038000 | orchestrator | service-cert-copy : glance | Copying over backend internal TLS key ------ 3.20s 2025-08-29 17:49:20.038010 | orchestrator | 2025-08-29 17:49:20 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:20.038065 | orchestrator | 2025-08-29 17:49:20 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:20.038076 | orchestrator | 2025-08-29 17:49:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:23.082002 | orchestrator | 2025-08-29 17:49:23 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:23.083922 | orchestrator | 2025-08-29 17:49:23 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:23.086056 | orchestrator | 2025-08-29 17:49:23 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:23.088088 | orchestrator | 2025-08-29 17:49:23 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:23.088382 | orchestrator | 2025-08-29 17:49:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:26.129870 | orchestrator | 2025-08-29 17:49:26 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:26.132251 | orchestrator | 2025-08-29 17:49:26 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:26.133896 | orchestrator | 2025-08-29 17:49:26 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:26.135555 | orchestrator | 2025-08-29 17:49:26 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:26.135660 | orchestrator | 2025-08-29 17:49:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:29.171607 | orchestrator | 2025-08-29 17:49:29 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:29.172775 | orchestrator | 2025-08-29 17:49:29 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:29.174175 | orchestrator | 2025-08-29 17:49:29 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:29.175318 | orchestrator | 2025-08-29 17:49:29 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:29.175340 | orchestrator | 2025-08-29 17:49:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:32.209177 | orchestrator | 2025-08-29 17:49:32 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:32.210668 | orchestrator | 2025-08-29 17:49:32 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:32.212986 | orchestrator | 2025-08-29 17:49:32 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:32.215015 | orchestrator | 2025-08-29 17:49:32 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:32.215126 | orchestrator | 2025-08-29 17:49:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:35.261512 | orchestrator | 2025-08-29 17:49:35 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:35.262813 | orchestrator | 2025-08-29 17:49:35 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:35.264491 | orchestrator | 2025-08-29 17:49:35 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:35.266274 | orchestrator | 2025-08-29 17:49:35 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:35.266347 | orchestrator | 2025-08-29 17:49:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:38.308434 | orchestrator | 2025-08-29 17:49:38 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:38.310193 | orchestrator | 2025-08-29 17:49:38 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:38.312131 | orchestrator | 2025-08-29 17:49:38 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:38.313489 | orchestrator | 2025-08-29 17:49:38 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:38.313845 | orchestrator | 2025-08-29 17:49:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:41.345662 | orchestrator | 2025-08-29 17:49:41 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:41.346326 | orchestrator | 2025-08-29 17:49:41 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:41.347493 | orchestrator | 2025-08-29 17:49:41 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:41.348670 | orchestrator | 2025-08-29 17:49:41 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:41.348699 | orchestrator | 2025-08-29 17:49:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:44.392408 | orchestrator | 2025-08-29 17:49:44 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:44.394968 | orchestrator | 2025-08-29 17:49:44 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:44.395932 | orchestrator | 2025-08-29 17:49:44 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:44.397370 | orchestrator | 2025-08-29 17:49:44 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:44.397396 | orchestrator | 2025-08-29 17:49:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:47.437803 | orchestrator | 2025-08-29 17:49:47 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:47.439489 | orchestrator | 2025-08-29 17:49:47 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:47.441711 | orchestrator | 2025-08-29 17:49:47 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:47.443522 | orchestrator | 2025-08-29 17:49:47 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:47.443556 | orchestrator | 2025-08-29 17:49:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:50.481405 | orchestrator | 2025-08-29 17:49:50 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:50.481941 | orchestrator | 2025-08-29 17:49:50 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:50.482681 | orchestrator | 2025-08-29 17:49:50 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:50.483858 | orchestrator | 2025-08-29 17:49:50 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:50.483934 | orchestrator | 2025-08-29 17:49:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:53.525820 | orchestrator | 2025-08-29 17:49:53 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:53.528582 | orchestrator | 2025-08-29 17:49:53 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:53.530112 | orchestrator | 2025-08-29 17:49:53 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:53.531642 | orchestrator | 2025-08-29 17:49:53 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:53.532020 | orchestrator | 2025-08-29 17:49:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:56.571280 | orchestrator | 2025-08-29 17:49:56 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:56.574095 | orchestrator | 2025-08-29 17:49:56 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:56.576251 | orchestrator | 2025-08-29 17:49:56 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:56.577274 | orchestrator | 2025-08-29 17:49:56 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:56.577817 | orchestrator | 2025-08-29 17:49:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:49:59.624906 | orchestrator | 2025-08-29 17:49:59 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:49:59.626414 | orchestrator | 2025-08-29 17:49:59 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:49:59.629169 | orchestrator | 2025-08-29 17:49:59 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:49:59.630634 | orchestrator | 2025-08-29 17:49:59 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:49:59.630667 | orchestrator | 2025-08-29 17:49:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:02.678659 | orchestrator | 2025-08-29 17:50:02 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:02.680153 | orchestrator | 2025-08-29 17:50:02 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:02.681825 | orchestrator | 2025-08-29 17:50:02 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:02.684038 | orchestrator | 2025-08-29 17:50:02 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:50:02.684078 | orchestrator | 2025-08-29 17:50:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:05.733449 | orchestrator | 2025-08-29 17:50:05 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:05.735107 | orchestrator | 2025-08-29 17:50:05 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:05.736825 | orchestrator | 2025-08-29 17:50:05 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:05.738434 | orchestrator | 2025-08-29 17:50:05 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:50:05.738506 | orchestrator | 2025-08-29 17:50:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:08.794265 | orchestrator | 2025-08-29 17:50:08 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:08.796551 | orchestrator | 2025-08-29 17:50:08 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:08.798260 | orchestrator | 2025-08-29 17:50:08 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:08.800822 | orchestrator | 2025-08-29 17:50:08 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:50:08.800862 | orchestrator | 2025-08-29 17:50:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:11.842685 | orchestrator | 2025-08-29 17:50:11 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:11.844135 | orchestrator | 2025-08-29 17:50:11 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:11.845676 | orchestrator | 2025-08-29 17:50:11 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:11.847043 | orchestrator | 2025-08-29 17:50:11 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:50:11.847184 | orchestrator | 2025-08-29 17:50:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:14.891767 | orchestrator | 2025-08-29 17:50:14 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:14.891868 | orchestrator | 2025-08-29 17:50:14 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:14.893285 | orchestrator | 2025-08-29 17:50:14 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:14.894837 | orchestrator | 2025-08-29 17:50:14 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state STARTED 2025-08-29 17:50:14.895420 | orchestrator | 2025-08-29 17:50:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:17.937321 | orchestrator | 2025-08-29 17:50:17 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:17.940265 | orchestrator | 2025-08-29 17:50:17 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:17.942843 | orchestrator | 2025-08-29 17:50:17 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:17.945403 | orchestrator | 2025-08-29 17:50:17 | INFO  | Task 3a4414b2-937c-4ec6-b36c-1f0d2681dc2b is in state SUCCESS 2025-08-29 17:50:17.945751 | orchestrator | 2025-08-29 17:50:17.945768 | orchestrator | 2025-08-29 17:50:17.945773 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:50:17.945777 | orchestrator | 2025-08-29 17:50:17.945781 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:50:17.945785 | orchestrator | Friday 29 August 2025 17:49:23 +0000 (0:00:00.241) 0:00:00.241 ********* 2025-08-29 17:50:17.945789 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:50:17.945793 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:50:17.945797 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:50:17.945801 | orchestrator | 2025-08-29 17:50:17.945805 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:50:17.945809 | orchestrator | Friday 29 August 2025 17:49:23 +0000 (0:00:00.266) 0:00:00.508 ********* 2025-08-29 17:50:17.945812 | orchestrator | ok: [testbed-node-0] => (item=enable_octavia_True) 2025-08-29 17:50:17.945816 | orchestrator | ok: [testbed-node-1] => (item=enable_octavia_True) 2025-08-29 17:50:17.945820 | orchestrator | ok: [testbed-node-2] => (item=enable_octavia_True) 2025-08-29 17:50:17.945824 | orchestrator | 2025-08-29 17:50:17.945827 | orchestrator | PLAY [Apply role octavia] ****************************************************** 2025-08-29 17:50:17.945831 | orchestrator | 2025-08-29 17:50:17.945835 | orchestrator | TASK [octavia : include_tasks] ************************************************* 2025-08-29 17:50:17.945839 | orchestrator | Friday 29 August 2025 17:49:24 +0000 (0:00:00.349) 0:00:00.857 ********* 2025-08-29 17:50:17.945842 | orchestrator | included: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:50:17.945846 | orchestrator | 2025-08-29 17:50:17.945850 | orchestrator | TASK [service-ks-register : octavia | Creating services] *********************** 2025-08-29 17:50:17.945854 | orchestrator | Friday 29 August 2025 17:49:24 +0000 (0:00:00.507) 0:00:01.364 ********* 2025-08-29 17:50:17.945858 | orchestrator | changed: [testbed-node-0] => (item=octavia (load-balancer)) 2025-08-29 17:50:17.945862 | orchestrator | 2025-08-29 17:50:17.945866 | orchestrator | TASK [service-ks-register : octavia | Creating endpoints] ********************** 2025-08-29 17:50:17.945869 | orchestrator | Friday 29 August 2025 17:49:28 +0000 (0:00:03.573) 0:00:04.938 ********* 2025-08-29 17:50:17.945873 | orchestrator | changed: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal) 2025-08-29 17:50:17.945877 | orchestrator | changed: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public) 2025-08-29 17:50:17.945881 | orchestrator | 2025-08-29 17:50:17.945885 | orchestrator | TASK [service-ks-register : octavia | Creating projects] *********************** 2025-08-29 17:50:17.945888 | orchestrator | Friday 29 August 2025 17:49:35 +0000 (0:00:06.909) 0:00:11.847 ********* 2025-08-29 17:50:17.945904 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:50:17.945908 | orchestrator | 2025-08-29 17:50:17.945912 | orchestrator | TASK [service-ks-register : octavia | Creating users] ************************** 2025-08-29 17:50:17.945916 | orchestrator | Friday 29 August 2025 17:49:38 +0000 (0:00:03.524) 0:00:15.372 ********* 2025-08-29 17:50:17.945920 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:50:17.945926 | orchestrator | changed: [testbed-node-0] => (item=octavia -> service) 2025-08-29 17:50:17.945933 | orchestrator | changed: [testbed-node-0] => (item=octavia -> service) 2025-08-29 17:50:17.945939 | orchestrator | 2025-08-29 17:50:17.945946 | orchestrator | TASK [service-ks-register : octavia | Creating roles] ************************** 2025-08-29 17:50:17.945991 | orchestrator | Friday 29 August 2025 17:49:46 +0000 (0:00:08.417) 0:00:23.790 ********* 2025-08-29 17:50:17.945997 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:50:17.946000 | orchestrator | 2025-08-29 17:50:17.946004 | orchestrator | TASK [service-ks-register : octavia | Granting user roles] ********************* 2025-08-29 17:50:17.946008 | orchestrator | Friday 29 August 2025 17:49:50 +0000 (0:00:03.668) 0:00:27.458 ********* 2025-08-29 17:50:17.946012 | orchestrator | changed: [testbed-node-0] => (item=octavia -> service -> admin) 2025-08-29 17:50:17.946030 | orchestrator | ok: [testbed-node-0] => (item=octavia -> service -> admin) 2025-08-29 17:50:17.946034 | orchestrator | 2025-08-29 17:50:17.946038 | orchestrator | TASK [octavia : Adding octavia related roles] ********************************** 2025-08-29 17:50:17.946041 | orchestrator | Friday 29 August 2025 17:49:58 +0000 (0:00:07.671) 0:00:35.130 ********* 2025-08-29 17:50:17.946045 | orchestrator | changed: [testbed-node-0] => (item=load-balancer_observer) 2025-08-29 17:50:17.946049 | orchestrator | changed: [testbed-node-0] => (item=load-balancer_global_observer) 2025-08-29 17:50:17.946053 | orchestrator | changed: [testbed-node-0] => (item=load-balancer_member) 2025-08-29 17:50:17.946056 | orchestrator | changed: [testbed-node-0] => (item=load-balancer_admin) 2025-08-29 17:50:17.946060 | orchestrator | changed: [testbed-node-0] => (item=load-balancer_quota_admin) 2025-08-29 17:50:17.946064 | orchestrator | 2025-08-29 17:50:17.946068 | orchestrator | TASK [octavia : include_tasks] ************************************************* 2025-08-29 17:50:17.946072 | orchestrator | Friday 29 August 2025 17:50:13 +0000 (0:00:15.230) 0:00:50.360 ********* 2025-08-29 17:50:17.946075 | orchestrator | included: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:50:17.946079 | orchestrator | 2025-08-29 17:50:17.946083 | orchestrator | TASK [octavia : Create amphora flavor] ***************************************** 2025-08-29 17:50:17.946087 | orchestrator | Friday 29 August 2025 17:50:14 +0000 (0:00:00.482) 0:00:50.842 ********* 2025-08-29 17:50:17.946091 | orchestrator | An exception occurred during task execution. To see the full traceback, use -vvv. The error was: keystoneauth1.exceptions.catalog.EndpointNotFound: internal endpoint for compute service in RegionOne region not found 2025-08-29 17:50:17.946107 | orchestrator | fatal: [testbed-node-0]: FAILED! => {"action": "os_nova_flavor", "changed": false, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible-tmp-1756489815.5447247-6682-66008460672029/AnsiballZ_compute_flavor.py\", line 107, in \n _ansiballz_main()\n File \"/tmp/ansible-tmp-1756489815.5447247-6682-66008460672029/AnsiballZ_compute_flavor.py\", line 99, in _ansiballz_main\n invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)\n File \"/tmp/ansible-tmp-1756489815.5447247-6682-66008460672029/AnsiballZ_compute_flavor.py\", line 47, in invoke_module\n runpy.run_module(mod_name='ansible_collections.openstack.cloud.plugins.modules.compute_flavor', init_globals=dict(_module_fqn='ansible_collections.openstack.cloud.plugins.modules.compute_flavor', _modlib_path=modlib_path),\n File \"\", line 226, in run_module\n File \"\", line 98, in _run_module_code\n File \"\", line 88, in _run_code\n File \"/tmp/ansible_os_nova_flavor_payload_uhlan6c4/ansible_os_nova_flavor_payload.zip/ansible_collections/openstack/cloud/plugins/modules/compute_flavor.py\", line 367, in \n File \"/tmp/ansible_os_nova_flavor_payload_uhlan6c4/ansible_os_nova_flavor_payload.zip/ansible_collections/openstack/cloud/plugins/modules/compute_flavor.py\", line 363, in main\n File \"/tmp/ansible_os_nova_flavor_payload_uhlan6c4/ansible_os_nova_flavor_payload.zip/ansible_collections/openstack/cloud/plugins/module_utils/openstack.py\", line 417, in __call__\n File \"/tmp/ansible_os_nova_flavor_payload_uhlan6c4/ansible_os_nova_flavor_payload.zip/ansible_collections/openstack/cloud/plugins/modules/compute_flavor.py\", line 220, in run\n File \"/opt/ansible/lib/python3.11/site-packages/openstack/service_description.py\", line 88, in __get__\n proxy = self._make_proxy(instance)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/ansible/lib/python3.11/site-packages/openstack/service_description.py\", line 286, in _make_proxy\n found_version = temp_adapter.get_api_major_version()\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/ansible/lib/python3.11/site-packages/keystoneauth1/adapter.py\", line 352, in get_api_major_version\n return self.session.get_api_major_version(auth or self.auth, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/ansible/lib/python3.11/site-packages/keystoneauth1/session.py\", line 1289, in get_api_major_version\n return auth.get_api_major_version(self, **kwargs)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/ansible/lib/python3.11/site-packages/keystoneauth1/identity/base.py\", line 497, in get_api_major_version\n data = get_endpoint_data(discover_versions=discover_versions)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/ansible/lib/python3.11/site-packages/keystoneauth1/identity/base.py\", line 272, in get_endpoint_data\n endpoint_data = service_catalog.endpoint_data_for(\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/ansible/lib/python3.11/site-packages/keystoneauth1/access/service_catalog.py\", line 459, in endpoint_data_for\n raise exceptions.EndpointNotFound(msg)\nkeystoneauth1.exceptions.catalog.EndpointNotFound: internal endpoint for compute service in RegionOne region not found\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1} 2025-08-29 17:50:17.946139 | orchestrator | 2025-08-29 17:50:17.946144 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:50:17.946148 | orchestrator | testbed-node-0 : ok=11  changed=5  unreachable=0 failed=1  skipped=0 rescued=0 ignored=0 2025-08-29 17:50:17.946152 | orchestrator | testbed-node-1 : ok=4  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:50:17.946157 | orchestrator | testbed-node-2 : ok=4  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:50:17.946161 | orchestrator | 2025-08-29 17:50:17.946164 | orchestrator | 2025-08-29 17:50:17.946168 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:50:17.946172 | orchestrator | Friday 29 August 2025 17:50:17 +0000 (0:00:03.544) 0:00:54.387 ********* 2025-08-29 17:50:17.946178 | orchestrator | =============================================================================== 2025-08-29 17:50:17.946182 | orchestrator | octavia : Adding octavia related roles --------------------------------- 15.23s 2025-08-29 17:50:17.946186 | orchestrator | service-ks-register : octavia | Creating users -------------------------- 8.42s 2025-08-29 17:50:17.946190 | orchestrator | service-ks-register : octavia | Granting user roles --------------------- 7.67s 2025-08-29 17:50:17.946194 | orchestrator | service-ks-register : octavia | Creating endpoints ---------------------- 6.91s 2025-08-29 17:50:17.946198 | orchestrator | service-ks-register : octavia | Creating roles -------------------------- 3.67s 2025-08-29 17:50:17.946204 | orchestrator | service-ks-register : octavia | Creating services ----------------------- 3.57s 2025-08-29 17:50:17.946208 | orchestrator | octavia : Create amphora flavor ----------------------------------------- 3.55s 2025-08-29 17:50:17.946212 | orchestrator | service-ks-register : octavia | Creating projects ----------------------- 3.52s 2025-08-29 17:50:17.946216 | orchestrator | octavia : include_tasks ------------------------------------------------- 0.51s 2025-08-29 17:50:17.946219 | orchestrator | octavia : include_tasks ------------------------------------------------- 0.48s 2025-08-29 17:50:17.946223 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.35s 2025-08-29 17:50:17.946227 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.27s 2025-08-29 17:50:17.946231 | orchestrator | 2025-08-29 17:50:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:20.993797 | orchestrator | 2025-08-29 17:50:20 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:20.996830 | orchestrator | 2025-08-29 17:50:20 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:20.997846 | orchestrator | 2025-08-29 17:50:20 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:20.999293 | orchestrator | 2025-08-29 17:50:20 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:20.999424 | orchestrator | 2025-08-29 17:50:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:24.042568 | orchestrator | 2025-08-29 17:50:24 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:24.044492 | orchestrator | 2025-08-29 17:50:24 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state STARTED 2025-08-29 17:50:24.047014 | orchestrator | 2025-08-29 17:50:24 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:24.048612 | orchestrator | 2025-08-29 17:50:24 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:24.048629 | orchestrator | 2025-08-29 17:50:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:27.093059 | orchestrator | 2025-08-29 17:50:27 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:27.093513 | orchestrator | 2025-08-29 17:50:27 | INFO  | Task 61e4ba6d-4010-41e6-bd11-b42ce7f6f002 is in state SUCCESS 2025-08-29 17:50:27.095073 | orchestrator | 2025-08-29 17:50:27.095092 | orchestrator | 2025-08-29 17:50:27.095101 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:50:27.095108 | orchestrator | 2025-08-29 17:50:27.095115 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:50:27.095123 | orchestrator | Friday 29 August 2025 17:46:53 +0000 (0:00:00.258) 0:00:00.258 ********* 2025-08-29 17:50:27.095130 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:50:27.095137 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:50:27.095145 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:50:27.095152 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:50:27.095159 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:50:27.095166 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:50:27.095172 | orchestrator | 2025-08-29 17:50:27.095179 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:50:27.095187 | orchestrator | Friday 29 August 2025 17:46:54 +0000 (0:00:00.655) 0:00:00.914 ********* 2025-08-29 17:50:27.095193 | orchestrator | ok: [testbed-node-0] => (item=enable_cinder_True) 2025-08-29 17:50:27.095201 | orchestrator | ok: [testbed-node-1] => (item=enable_cinder_True) 2025-08-29 17:50:27.095207 | orchestrator | ok: [testbed-node-2] => (item=enable_cinder_True) 2025-08-29 17:50:27.095214 | orchestrator | ok: [testbed-node-3] => (item=enable_cinder_True) 2025-08-29 17:50:27.095221 | orchestrator | ok: [testbed-node-4] => (item=enable_cinder_True) 2025-08-29 17:50:27.095244 | orchestrator | ok: [testbed-node-5] => (item=enable_cinder_True) 2025-08-29 17:50:27.095251 | orchestrator | 2025-08-29 17:50:27.095258 | orchestrator | PLAY [Apply role cinder] ******************************************************* 2025-08-29 17:50:27.095265 | orchestrator | 2025-08-29 17:50:27.095272 | orchestrator | TASK [cinder : include_tasks] ************************************************** 2025-08-29 17:50:27.095279 | orchestrator | Friday 29 August 2025 17:46:54 +0000 (0:00:00.588) 0:00:01.502 ********* 2025-08-29 17:50:27.095286 | orchestrator | included: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:50:27.095294 | orchestrator | 2025-08-29 17:50:27.095301 | orchestrator | TASK [service-ks-register : cinder | Creating services] ************************ 2025-08-29 17:50:27.095307 | orchestrator | Friday 29 August 2025 17:46:55 +0000 (0:00:01.137) 0:00:02.640 ********* 2025-08-29 17:50:27.095315 | orchestrator | changed: [testbed-node-0] => (item=cinderv3 (volumev3)) 2025-08-29 17:50:27.095322 | orchestrator | 2025-08-29 17:50:27.095328 | orchestrator | TASK [service-ks-register : cinder | Creating endpoints] *********************** 2025-08-29 17:50:27.095335 | orchestrator | Friday 29 August 2025 17:46:58 +0000 (0:00:03.049) 0:00:05.690 ********* 2025-08-29 17:50:27.095342 | orchestrator | changed: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal) 2025-08-29 17:50:27.095349 | orchestrator | changed: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public) 2025-08-29 17:50:27.095356 | orchestrator | 2025-08-29 17:50:27.095363 | orchestrator | TASK [service-ks-register : cinder | Creating projects] ************************ 2025-08-29 17:50:27.095370 | orchestrator | Friday 29 August 2025 17:47:05 +0000 (0:00:06.497) 0:00:12.188 ********* 2025-08-29 17:50:27.095377 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:50:27.095384 | orchestrator | 2025-08-29 17:50:27.095391 | orchestrator | TASK [service-ks-register : cinder | Creating users] *************************** 2025-08-29 17:50:27.095398 | orchestrator | Friday 29 August 2025 17:47:08 +0000 (0:00:03.361) 0:00:15.549 ********* 2025-08-29 17:50:27.095405 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:50:27.095412 | orchestrator | changed: [testbed-node-0] => (item=cinder -> service) 2025-08-29 17:50:27.095418 | orchestrator | 2025-08-29 17:50:27.095425 | orchestrator | TASK [service-ks-register : cinder | Creating roles] *************************** 2025-08-29 17:50:27.095432 | orchestrator | Friday 29 August 2025 17:47:12 +0000 (0:00:04.161) 0:00:19.710 ********* 2025-08-29 17:50:27.095439 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:50:27.095457 | orchestrator | 2025-08-29 17:50:27.095464 | orchestrator | TASK [service-ks-register : cinder | Granting user roles] ********************** 2025-08-29 17:50:27.095470 | orchestrator | Friday 29 August 2025 17:47:16 +0000 (0:00:03.942) 0:00:23.653 ********* 2025-08-29 17:50:27.095476 | orchestrator | changed: [testbed-node-0] => (item=cinder -> service -> admin) 2025-08-29 17:50:27.095483 | orchestrator | changed: [testbed-node-0] => (item=cinder -> service -> service) 2025-08-29 17:50:27.095490 | orchestrator | 2025-08-29 17:50:27.095496 | orchestrator | TASK [cinder : Ensuring config directories exist] ****************************** 2025-08-29 17:50:27.095503 | orchestrator | Friday 29 August 2025 17:47:23 +0000 (0:00:07.138) 0:00:30.791 ********* 2025-08-29 17:50:27.095520 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.095534 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.095542 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.095549 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.095647 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.095937 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096012 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096024 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096032 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096039 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096047 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096072 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096084 | orchestrator | 2025-08-29 17:50:27.096092 | orchestrator | TASK [cinder : include_tasks] ************************************************** 2025-08-29 17:50:27.096099 | orchestrator | Friday 29 August 2025 17:47:27 +0000 (0:00:03.600) 0:00:34.392 ********* 2025-08-29 17:50:27.096106 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.096113 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.096120 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.096127 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.096134 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.096141 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.096148 | orchestrator | 2025-08-29 17:50:27.096155 | orchestrator | TASK [cinder : include_tasks] ************************************************** 2025-08-29 17:50:27.096162 | orchestrator | Friday 29 August 2025 17:47:28 +0000 (0:00:00.851) 0:00:35.243 ********* 2025-08-29 17:50:27.096168 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.096175 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.096182 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.096189 | orchestrator | included: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:50:27.096196 | orchestrator | 2025-08-29 17:50:27.096203 | orchestrator | TASK [cinder : Ensuring cinder service ceph config subdirs exists] ************* 2025-08-29 17:50:27.096210 | orchestrator | Friday 29 August 2025 17:47:29 +0000 (0:00:00.806) 0:00:36.050 ********* 2025-08-29 17:50:27.096217 | orchestrator | changed: [testbed-node-3] => (item=cinder-volume) 2025-08-29 17:50:27.096224 | orchestrator | changed: [testbed-node-5] => (item=cinder-volume) 2025-08-29 17:50:27.096231 | orchestrator | changed: [testbed-node-4] => (item=cinder-volume) 2025-08-29 17:50:27.096238 | orchestrator | changed: [testbed-node-5] => (item=cinder-backup) 2025-08-29 17:50:27.096245 | orchestrator | changed: [testbed-node-3] => (item=cinder-backup) 2025-08-29 17:50:27.096252 | orchestrator | changed: [testbed-node-4] => (item=cinder-backup) 2025-08-29 17:50:27.096259 | orchestrator | 2025-08-29 17:50:27.096266 | orchestrator | TASK [cinder : Copying over multiple ceph.conf for cinder services] ************ 2025-08-29 17:50:27.096273 | orchestrator | Friday 29 August 2025 17:47:31 +0000 (0:00:01.910) 0:00:37.960 ********* 2025-08-29 17:50:27.096281 | orchestrator | skipping: [testbed-node-3] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.13:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])  2025-08-29 17:50:27.096289 | orchestrator | skipping: [testbed-node-3] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])  2025-08-29 17:50:27.096316 | orchestrator | skipping: [testbed-node-4] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.14:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])  2025-08-29 17:50:27.096325 | orchestrator | skipping: [testbed-node-4] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])  2025-08-29 17:50:27.096332 | orchestrator | skipping: [testbed-node-5] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.15:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])  2025-08-29 17:50:27.096340 | orchestrator | skipping: [testbed-node-5] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])  2025-08-29 17:50:27.096347 | orchestrator | changed: [testbed-node-5] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) 2025-08-29 17:50:27.096374 | orchestrator | changed: [testbed-node-4] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) 2025-08-29 17:50:27.096382 | orchestrator | changed: [testbed-node-3] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) 2025-08-29 17:50:27.096390 | orchestrator | changed: [testbed-node-5] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) 2025-08-29 17:50:27.096397 | orchestrator | changed: [testbed-node-4] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) 2025-08-29 17:50:27.096409 | orchestrator | changed: [testbed-node-3] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) 2025-08-29 17:50:27.096416 | orchestrator | 2025-08-29 17:50:27.096423 | orchestrator | TASK [cinder : Copy over Ceph keyring files for cinder-volume] ***************** 2025-08-29 17:50:27.096430 | orchestrator | Friday 29 August 2025 17:47:35 +0000 (0:00:03.928) 0:00:41.889 ********* 2025-08-29 17:50:27.096437 | orchestrator | changed: [testbed-node-3] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:50:27.096468 | orchestrator | changed: [testbed-node-4] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:50:27.096476 | orchestrator | changed: [testbed-node-5] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}) 2025-08-29 17:50:27.096483 | orchestrator | 2025-08-29 17:50:27.096489 | orchestrator | TASK [cinder : Copy over Ceph keyring files for cinder-backup] ***************** 2025-08-29 17:50:27.096496 | orchestrator | Friday 29 August 2025 17:47:36 +0000 (0:00:01.957) 0:00:43.846 ********* 2025-08-29 17:50:27.096519 | orchestrator | changed: [testbed-node-5] => (item=ceph.client.cinder.keyring) 2025-08-29 17:50:27.096526 | orchestrator | changed: [testbed-node-3] => (item=ceph.client.cinder.keyring) 2025-08-29 17:50:27.096533 | orchestrator | changed: [testbed-node-4] => (item=ceph.client.cinder.keyring) 2025-08-29 17:50:27.096539 | orchestrator | changed: [testbed-node-5] => (item=ceph.client.cinder-backup.keyring) 2025-08-29 17:50:27.096546 | orchestrator | changed: [testbed-node-3] => (item=ceph.client.cinder-backup.keyring) 2025-08-29 17:50:27.096553 | orchestrator | changed: [testbed-node-4] => (item=ceph.client.cinder-backup.keyring) 2025-08-29 17:50:27.096559 | orchestrator | 2025-08-29 17:50:27.096566 | orchestrator | TASK [cinder : Ensuring config directory has correct owner and permission] ***** 2025-08-29 17:50:27.096572 | orchestrator | Friday 29 August 2025 17:47:39 +0000 (0:00:02.924) 0:00:46.771 ********* 2025-08-29 17:50:27.096579 | orchestrator | ok: [testbed-node-3] => (item=cinder-volume) 2025-08-29 17:50:27.096585 | orchestrator | ok: [testbed-node-5] => (item=cinder-volume) 2025-08-29 17:50:27.096592 | orchestrator | ok: [testbed-node-4] => (item=cinder-volume) 2025-08-29 17:50:27.096598 | orchestrator | ok: [testbed-node-3] => (item=cinder-backup) 2025-08-29 17:50:27.096605 | orchestrator | ok: [testbed-node-5] => (item=cinder-backup) 2025-08-29 17:50:27.096611 | orchestrator | ok: [testbed-node-4] => (item=cinder-backup) 2025-08-29 17:50:27.096618 | orchestrator | 2025-08-29 17:50:27.096624 | orchestrator | TASK [cinder : Check if policies shall be overwritten] ************************* 2025-08-29 17:50:27.096631 | orchestrator | Friday 29 August 2025 17:47:41 +0000 (0:00:01.165) 0:00:47.936 ********* 2025-08-29 17:50:27.096637 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.096644 | orchestrator | 2025-08-29 17:50:27.096651 | orchestrator | TASK [cinder : Set cinder policy file] ***************************************** 2025-08-29 17:50:27.096657 | orchestrator | Friday 29 August 2025 17:47:41 +0000 (0:00:00.182) 0:00:48.119 ********* 2025-08-29 17:50:27.096665 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.096671 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.096678 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.096690 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.096697 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.096704 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.096711 | orchestrator | 2025-08-29 17:50:27.096718 | orchestrator | TASK [cinder : include_tasks] ************************************************** 2025-08-29 17:50:27.096725 | orchestrator | Friday 29 August 2025 17:47:42 +0000 (0:00:00.859) 0:00:48.978 ********* 2025-08-29 17:50:27.096734 | orchestrator | included: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:50:27.096742 | orchestrator | 2025-08-29 17:50:27.096749 | orchestrator | TASK [service-cert-copy : cinder | Copying over extra CA certificates] ********* 2025-08-29 17:50:27.096757 | orchestrator | Friday 29 August 2025 17:47:44 +0000 (0:00:02.443) 0:00:51.422 ********* 2025-08-29 17:50:27.096765 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.096773 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.096796 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.096804 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096815 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096822 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096829 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096850 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096858 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096869 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096876 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096883 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.096890 | orchestrator | 2025-08-29 17:50:27.096897 | orchestrator | TASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] *** 2025-08-29 17:50:27.096903 | orchestrator | Friday 29 August 2025 17:47:48 +0000 (0:00:04.439) 0:00:55.862 ********* 2025-08-29 17:50:27.096913 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.096920 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.096931 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.096938 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.096945 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.096952 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.096959 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.096966 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.096972 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.096983 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.096993 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097000 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.097007 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097014 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097021 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.097028 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097039 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097049 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.097056 | orchestrator | 2025-08-29 17:50:27.097063 | orchestrator | TASK [service-cert-copy : cinder | Copying over backend internal TLS key] ****** 2025-08-29 17:50:27.097069 | orchestrator | Friday 29 August 2025 17:47:51 +0000 (0:00:02.073) 0:00:57.936 ********* 2025-08-29 17:50:27.097076 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.097083 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097090 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.097097 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097104 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.097114 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.097125 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097132 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.097138 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.097145 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097152 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097159 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.097166 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097176 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097191 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.097197 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097204 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097211 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.097218 | orchestrator | 2025-08-29 17:50:27.097224 | orchestrator | TASK [cinder : Copying over config.json files for services] ******************** 2025-08-29 17:50:27.097231 | orchestrator | Friday 29 August 2025 17:47:53 +0000 (0:00:02.161) 0:01:00.097 ********* 2025-08-29 17:50:27.097238 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097245 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097259 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097266 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097273 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097280 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097287 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097300 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097307 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097314 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097321 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097328 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097338 | orchestrator | 2025-08-29 17:50:27.097345 | orchestrator | TASK [cinder : Copying over cinder-wsgi.conf] ********************************** 2025-08-29 17:50:27.097352 | orchestrator | Friday 29 August 2025 17:47:56 +0000 (0:00:02.860) 0:01:02.958 ********* 2025-08-29 17:50:27.097358 | orchestrator | skipping: [testbed-node-3] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)  2025-08-29 17:50:27.097365 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.097371 | orchestrator | skipping: [testbed-node-4] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)  2025-08-29 17:50:27.097378 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.097385 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) 2025-08-29 17:50:27.097391 | orchestrator | skipping: [testbed-node-5] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)  2025-08-29 17:50:27.097398 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.097404 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) 2025-08-29 17:50:27.097414 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2) 2025-08-29 17:50:27.097420 | orchestrator | 2025-08-29 17:50:27.097427 | orchestrator | TASK [cinder : Copying over cinder.conf] *************************************** 2025-08-29 17:50:27.097433 | orchestrator | Friday 29 August 2025 17:47:58 +0000 (0:00:01.983) 0:01:04.942 ********* 2025-08-29 17:50:27.097440 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097459 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097466 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097477 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097488 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097496 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097503 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097510 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097522 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097529 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097539 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097546 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097553 | orchestrator | 2025-08-29 17:50:27.097560 | orchestrator | TASK [cinder : Generating 'hostnqn' file for cinder_volume] ******************** 2025-08-29 17:50:27.097567 | orchestrator | Friday 29 August 2025 17:48:07 +0000 (0:00:09.387) 0:01:14.330 ********* 2025-08-29 17:50:27.097574 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.097580 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.097587 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.097593 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:50:27.097600 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:50:27.097607 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:50:27.097613 | orchestrator | 2025-08-29 17:50:27.097620 | orchestrator | TASK [cinder : Copying over existing policy file] ****************************** 2025-08-29 17:50:27.097626 | orchestrator | Friday 29 August 2025 17:48:09 +0000 (0:00:02.209) 0:01:16.540 ********* 2025-08-29 17:50:27.097633 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.097644 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097651 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.097661 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.097668 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097675 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.097682 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})  2025-08-29 17:50:27.097689 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097700 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.097707 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097714 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097721 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.097731 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097738 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097745 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.097752 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097763 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})  2025-08-29 17:50:27.097770 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.097776 | orchestrator | 2025-08-29 17:50:27.097783 | orchestrator | TASK [cinder : Copying over nfs_shares files for cinder_volume] **************** 2025-08-29 17:50:27.097790 | orchestrator | Friday 29 August 2025 17:48:10 +0000 (0:00:00.896) 0:01:17.436 ********* 2025-08-29 17:50:27.097796 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.097803 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.097810 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.097816 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.097823 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.097830 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.097836 | orchestrator | 2025-08-29 17:50:27.097843 | orchestrator | TASK [cinder : Check cinder containers] **************************************** 2025-08-29 17:50:27.097850 | orchestrator | Friday 29 August 2025 17:48:11 +0000 (0:00:00.608) 0:01:18.044 ********* 2025-08-29 17:50:27.097859 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097867 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097881 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097888 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) 2025-08-29 17:50:27.097895 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097905 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', '', 'kolla_logs:/var/log/kolla/', '', '/opt/cinder-driver-dm-clone:/var/lib/kolla/venv/lib/python3/site-packages/cinder-driver-dm-clone'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097912 | orchestrator | changed: [testbed-node-3] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097923 | orchestrator | changed: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097930 | orchestrator | changed: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097937 | orchestrator | changed: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097944 | orchestrator | changed: [testbed-node-5] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097954 | orchestrator | changed: [testbed-node-4] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', 'iscsi_info:/etc/iscsi', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) 2025-08-29 17:50:27.097961 | orchestrator | 2025-08-29 17:50:27.097968 | orchestrator | TASK [cinder : include_tasks] ************************************************** 2025-08-29 17:50:27.097974 | orchestrator | Friday 29 August 2025 17:48:13 +0000 (0:00:02.239) 0:01:20.284 ********* 2025-08-29 17:50:27.097985 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.097991 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:50:27.097998 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:50:27.098004 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:50:27.098011 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:50:27.098036 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:50:27.098043 | orchestrator | 2025-08-29 17:50:27.098050 | orchestrator | TASK [cinder : Creating Cinder database] *************************************** 2025-08-29 17:50:27.098057 | orchestrator | Friday 29 August 2025 17:48:14 +0000 (0:00:00.733) 0:01:21.018 ********* 2025-08-29 17:50:27.098064 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:50:27.098070 | orchestrator | 2025-08-29 17:50:27.098077 | orchestrator | TASK [cinder : Creating Cinder database user and setting permissions] ********** 2025-08-29 17:50:27.098084 | orchestrator | Friday 29 August 2025 17:48:16 +0000 (0:00:02.341) 0:01:23.359 ********* 2025-08-29 17:50:27.098091 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:50:27.098097 | orchestrator | 2025-08-29 17:50:27.098103 | orchestrator | TASK [cinder : Running Cinder bootstrap container] ***************************** 2025-08-29 17:50:27.098109 | orchestrator | Friday 29 August 2025 17:48:18 +0000 (0:00:02.412) 0:01:25.772 ********* 2025-08-29 17:50:27.098116 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:50:27.098122 | orchestrator | 2025-08-29 17:50:27.098128 | orchestrator | TASK [cinder : Flush handlers] ************************************************* 2025-08-29 17:50:27.098135 | orchestrator | Friday 29 August 2025 17:48:41 +0000 (0:00:23.023) 0:01:48.796 ********* 2025-08-29 17:50:27.098141 | orchestrator | 2025-08-29 17:50:27.098147 | orchestrator | TASK [cinder : Flush handlers] ************************************************* 2025-08-29 17:50:27.098154 | orchestrator | Friday 29 August 2025 17:48:41 +0000 (0:00:00.069) 0:01:48.865 ********* 2025-08-29 17:50:27.098161 | orchestrator | 2025-08-29 17:50:27.098167 | orchestrator | TASK [cinder : Flush handlers] ************************************************* 2025-08-29 17:50:27.098173 | orchestrator | Friday 29 August 2025 17:48:42 +0000 (0:00:00.065) 0:01:48.930 ********* 2025-08-29 17:50:27.098180 | orchestrator | 2025-08-29 17:50:27.098186 | orchestrator | TASK [cinder : Flush handlers] ************************************************* 2025-08-29 17:50:27.098193 | orchestrator | Friday 29 August 2025 17:48:42 +0000 (0:00:00.061) 0:01:48.992 ********* 2025-08-29 17:50:27.098199 | orchestrator | 2025-08-29 17:50:27.098205 | orchestrator | TASK [cinder : Flush handlers] ************************************************* 2025-08-29 17:50:27.098212 | orchestrator | Friday 29 August 2025 17:48:42 +0000 (0:00:00.065) 0:01:49.057 ********* 2025-08-29 17:50:27.098218 | orchestrator | 2025-08-29 17:50:27.098224 | orchestrator | TASK [cinder : Flush handlers] ************************************************* 2025-08-29 17:50:27.098231 | orchestrator | Friday 29 August 2025 17:48:42 +0000 (0:00:00.064) 0:01:49.121 ********* 2025-08-29 17:50:27.098238 | orchestrator | 2025-08-29 17:50:27.098244 | orchestrator | RUNNING HANDLER [cinder : Restart cinder-api container] ************************ 2025-08-29 17:50:27.098250 | orchestrator | Friday 29 August 2025 17:48:42 +0000 (0:00:00.065) 0:01:49.186 ********* 2025-08-29 17:50:27.098257 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:50:27.098263 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:50:27.098269 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:50:27.098275 | orchestrator | 2025-08-29 17:50:27.098282 | orchestrator | RUNNING HANDLER [cinder : Restart cinder-scheduler container] ****************** 2025-08-29 17:50:27.098288 | orchestrator | Friday 29 August 2025 17:49:30 +0000 (0:00:48.007) 0:02:37.194 ********* 2025-08-29 17:50:27.098295 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:50:27.098301 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:50:27.098308 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:50:27.098315 | orchestrator | 2025-08-29 17:50:27.098322 | orchestrator | RUNNING HANDLER [cinder : Restart cinder-volume container] ********************* 2025-08-29 17:50:27.098328 | orchestrator | Friday 29 August 2025 17:49:35 +0000 (0:00:05.236) 0:02:42.430 ********* 2025-08-29 17:50:27.098334 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:50:27.098345 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:50:27.098351 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:50:27.098358 | orchestrator | 2025-08-29 17:50:27.098365 | orchestrator | RUNNING HANDLER [cinder : Restart cinder-backup container] ********************* 2025-08-29 17:50:27.098371 | orchestrator | Friday 29 August 2025 17:50:14 +0000 (0:00:38.658) 0:03:21.089 ********* 2025-08-29 17:50:27.098378 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:50:27.098384 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:50:27.098390 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:50:27.098396 | orchestrator | 2025-08-29 17:50:27.098403 | orchestrator | RUNNING HANDLER [cinder : Wait for cinder services to update service versions] *** 2025-08-29 17:50:27.098409 | orchestrator | Friday 29 August 2025 17:50:24 +0000 (0:00:10.484) 0:03:31.573 ********* 2025-08-29 17:50:27.098416 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:50:27.098423 | orchestrator | 2025-08-29 17:50:27.098429 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:50:27.098439 | orchestrator | testbed-node-0 : ok=21  changed=15  unreachable=0 failed=0 skipped=11  rescued=0 ignored=0 2025-08-29 17:50:27.098459 | orchestrator | testbed-node-1 : ok=12  changed=8  unreachable=0 failed=0 skipped=9  rescued=0 ignored=0 2025-08-29 17:50:27.098466 | orchestrator | testbed-node-2 : ok=12  changed=8  unreachable=0 failed=0 skipped=9  rescued=0 ignored=0 2025-08-29 17:50:27.098473 | orchestrator | testbed-node-3 : ok=18  changed=12  unreachable=0 failed=0 skipped=8  rescued=0 ignored=0 2025-08-29 17:50:27.098480 | orchestrator | testbed-node-4 : ok=18  changed=12  unreachable=0 failed=0 skipped=8  rescued=0 ignored=0 2025-08-29 17:50:27.098487 | orchestrator | testbed-node-5 : ok=18  changed=12  unreachable=0 failed=0 skipped=8  rescued=0 ignored=0 2025-08-29 17:50:27.098494 | orchestrator | 2025-08-29 17:50:27.098500 | orchestrator | 2025-08-29 17:50:27.098507 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:50:27.098513 | orchestrator | Friday 29 August 2025 17:50:25 +0000 (0:00:00.676) 0:03:32.250 ********* 2025-08-29 17:50:27.098519 | orchestrator | =============================================================================== 2025-08-29 17:50:27.098525 | orchestrator | cinder : Restart cinder-api container ---------------------------------- 48.01s 2025-08-29 17:50:27.098531 | orchestrator | cinder : Restart cinder-volume container ------------------------------- 38.66s 2025-08-29 17:50:27.098538 | orchestrator | cinder : Running Cinder bootstrap container ---------------------------- 23.02s 2025-08-29 17:50:27.098544 | orchestrator | cinder : Restart cinder-backup container ------------------------------- 10.48s 2025-08-29 17:50:27.098551 | orchestrator | cinder : Copying over cinder.conf --------------------------------------- 9.39s 2025-08-29 17:50:27.098557 | orchestrator | service-ks-register : cinder | Granting user roles ---------------------- 7.14s 2025-08-29 17:50:27.098563 | orchestrator | service-ks-register : cinder | Creating endpoints ----------------------- 6.50s 2025-08-29 17:50:27.098570 | orchestrator | cinder : Restart cinder-scheduler container ----------------------------- 5.24s 2025-08-29 17:50:27.098576 | orchestrator | service-cert-copy : cinder | Copying over extra CA certificates --------- 4.44s 2025-08-29 17:50:27.098582 | orchestrator | service-ks-register : cinder | Creating users --------------------------- 4.16s 2025-08-29 17:50:27.098588 | orchestrator | service-ks-register : cinder | Creating roles --------------------------- 3.94s 2025-08-29 17:50:27.098594 | orchestrator | cinder : Copying over multiple ceph.conf for cinder services ------------ 3.93s 2025-08-29 17:50:27.098600 | orchestrator | cinder : Ensuring config directories exist ------------------------------ 3.60s 2025-08-29 17:50:27.098606 | orchestrator | service-ks-register : cinder | Creating projects ------------------------ 3.36s 2025-08-29 17:50:27.098618 | orchestrator | service-ks-register : cinder | Creating services ------------------------ 3.05s 2025-08-29 17:50:27.098624 | orchestrator | cinder : Copy over Ceph keyring files for cinder-backup ----------------- 2.92s 2025-08-29 17:50:27.098631 | orchestrator | cinder : Copying over config.json files for services -------------------- 2.86s 2025-08-29 17:50:27.098637 | orchestrator | cinder : include_tasks -------------------------------------------------- 2.44s 2025-08-29 17:50:27.098643 | orchestrator | cinder : Creating Cinder database user and setting permissions ---------- 2.41s 2025-08-29 17:50:27.098648 | orchestrator | cinder : Creating Cinder database --------------------------------------- 2.34s 2025-08-29 17:50:27.098654 | orchestrator | 2025-08-29 17:50:27 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:27.098661 | orchestrator | 2025-08-29 17:50:27 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:27.098668 | orchestrator | 2025-08-29 17:50:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:30.131302 | orchestrator | 2025-08-29 17:50:30 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:30.131971 | orchestrator | 2025-08-29 17:50:30 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:30.133183 | orchestrator | 2025-08-29 17:50:30 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:30.133213 | orchestrator | 2025-08-29 17:50:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:33.174629 | orchestrator | 2025-08-29 17:50:33 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:33.175839 | orchestrator | 2025-08-29 17:50:33 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:33.177233 | orchestrator | 2025-08-29 17:50:33 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:33.177249 | orchestrator | 2025-08-29 17:50:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:36.216971 | orchestrator | 2025-08-29 17:50:36 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:36.219340 | orchestrator | 2025-08-29 17:50:36 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:36.221273 | orchestrator | 2025-08-29 17:50:36 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:36.221632 | orchestrator | 2025-08-29 17:50:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:39.267064 | orchestrator | 2025-08-29 17:50:39 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:39.268171 | orchestrator | 2025-08-29 17:50:39 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:39.270754 | orchestrator | 2025-08-29 17:50:39 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:39.271315 | orchestrator | 2025-08-29 17:50:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:42.302283 | orchestrator | 2025-08-29 17:50:42 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:42.304096 | orchestrator | 2025-08-29 17:50:42 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:42.305031 | orchestrator | 2025-08-29 17:50:42 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:42.305069 | orchestrator | 2025-08-29 17:50:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:45.364399 | orchestrator | 2025-08-29 17:50:45 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:45.365340 | orchestrator | 2025-08-29 17:50:45 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:45.366548 | orchestrator | 2025-08-29 17:50:45 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:45.366679 | orchestrator | 2025-08-29 17:50:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:48.411892 | orchestrator | 2025-08-29 17:50:48 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:48.411948 | orchestrator | 2025-08-29 17:50:48 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:48.412114 | orchestrator | 2025-08-29 17:50:48 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:48.412363 | orchestrator | 2025-08-29 17:50:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:51.458281 | orchestrator | 2025-08-29 17:50:51 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:51.460612 | orchestrator | 2025-08-29 17:50:51 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:51.462259 | orchestrator | 2025-08-29 17:50:51 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:51.462477 | orchestrator | 2025-08-29 17:50:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:54.499803 | orchestrator | 2025-08-29 17:50:54 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:54.499878 | orchestrator | 2025-08-29 17:50:54 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:54.500969 | orchestrator | 2025-08-29 17:50:54 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:54.501046 | orchestrator | 2025-08-29 17:50:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:50:57.542248 | orchestrator | 2025-08-29 17:50:57 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:50:57.542357 | orchestrator | 2025-08-29 17:50:57 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:50:57.542869 | orchestrator | 2025-08-29 17:50:57 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:50:57.543569 | orchestrator | 2025-08-29 17:50:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:00.577058 | orchestrator | 2025-08-29 17:51:00 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:00.577724 | orchestrator | 2025-08-29 17:51:00 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:00.578761 | orchestrator | 2025-08-29 17:51:00 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:00.578794 | orchestrator | 2025-08-29 17:51:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:03.617610 | orchestrator | 2025-08-29 17:51:03 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:03.618639 | orchestrator | 2025-08-29 17:51:03 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:03.619873 | orchestrator | 2025-08-29 17:51:03 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:03.619897 | orchestrator | 2025-08-29 17:51:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:06.669110 | orchestrator | 2025-08-29 17:51:06 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:06.673371 | orchestrator | 2025-08-29 17:51:06 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:06.675397 | orchestrator | 2025-08-29 17:51:06 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:06.675727 | orchestrator | 2025-08-29 17:51:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:09.717591 | orchestrator | 2025-08-29 17:51:09 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:09.719776 | orchestrator | 2025-08-29 17:51:09 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:09.724861 | orchestrator | 2025-08-29 17:51:09 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:09.724914 | orchestrator | 2025-08-29 17:51:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:12.779413 | orchestrator | 2025-08-29 17:51:12 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:12.781302 | orchestrator | 2025-08-29 17:51:12 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:12.782224 | orchestrator | 2025-08-29 17:51:12 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:12.782252 | orchestrator | 2025-08-29 17:51:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:15.827908 | orchestrator | 2025-08-29 17:51:15 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:15.829693 | orchestrator | 2025-08-29 17:51:15 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:15.831714 | orchestrator | 2025-08-29 17:51:15 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:15.831754 | orchestrator | 2025-08-29 17:51:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:18.871172 | orchestrator | 2025-08-29 17:51:18 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:18.874566 | orchestrator | 2025-08-29 17:51:18 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:18.876874 | orchestrator | 2025-08-29 17:51:18 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:18.876907 | orchestrator | 2025-08-29 17:51:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:21.922813 | orchestrator | 2025-08-29 17:51:21 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:21.924283 | orchestrator | 2025-08-29 17:51:21 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:21.925643 | orchestrator | 2025-08-29 17:51:21 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:21.925836 | orchestrator | 2025-08-29 17:51:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:24.967985 | orchestrator | 2025-08-29 17:51:24 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:24.969898 | orchestrator | 2025-08-29 17:51:24 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:24.972848 | orchestrator | 2025-08-29 17:51:24 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:24.972881 | orchestrator | 2025-08-29 17:51:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:28.018849 | orchestrator | 2025-08-29 17:51:28 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:28.020170 | orchestrator | 2025-08-29 17:51:28 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:28.021723 | orchestrator | 2025-08-29 17:51:28 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:28.021761 | orchestrator | 2025-08-29 17:51:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:31.076917 | orchestrator | 2025-08-29 17:51:31 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:31.078183 | orchestrator | 2025-08-29 17:51:31 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:31.080491 | orchestrator | 2025-08-29 17:51:31 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:31.080805 | orchestrator | 2025-08-29 17:51:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:34.125181 | orchestrator | 2025-08-29 17:51:34 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:34.126205 | orchestrator | 2025-08-29 17:51:34 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:34.127817 | orchestrator | 2025-08-29 17:51:34 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:34.127853 | orchestrator | 2025-08-29 17:51:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:37.170669 | orchestrator | 2025-08-29 17:51:37 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:37.172259 | orchestrator | 2025-08-29 17:51:37 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:37.175680 | orchestrator | 2025-08-29 17:51:37 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:37.175870 | orchestrator | 2025-08-29 17:51:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:40.236581 | orchestrator | 2025-08-29 17:51:40 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state STARTED 2025-08-29 17:51:40.237207 | orchestrator | 2025-08-29 17:51:40 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:40.237756 | orchestrator | 2025-08-29 17:51:40 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:40.237788 | orchestrator | 2025-08-29 17:51:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:43.284849 | orchestrator | 2025-08-29 17:51:43 | INFO  | Task aecb195b-977d-4d65-b9e3-214e22db188a is in state SUCCESS 2025-08-29 17:51:43.285146 | orchestrator | 2025-08-29 17:51:43 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:43.287488 | orchestrator | 2025-08-29 17:51:43 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:43.288462 | orchestrator | 2025-08-29 17:51:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:46.340957 | orchestrator | 2025-08-29 17:51:46 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:46.343522 | orchestrator | 2025-08-29 17:51:46 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:46.343570 | orchestrator | 2025-08-29 17:51:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:49.390522 | orchestrator | 2025-08-29 17:51:49 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:49.391162 | orchestrator | 2025-08-29 17:51:49 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:49.391544 | orchestrator | 2025-08-29 17:51:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:52.437490 | orchestrator | 2025-08-29 17:51:52 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:52.438155 | orchestrator | 2025-08-29 17:51:52 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:52.438196 | orchestrator | 2025-08-29 17:51:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:55.473299 | orchestrator | 2025-08-29 17:51:55 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:55.474340 | orchestrator | 2025-08-29 17:51:55 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:55.474378 | orchestrator | 2025-08-29 17:51:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:51:58.514166 | orchestrator | 2025-08-29 17:51:58 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:51:58.514954 | orchestrator | 2025-08-29 17:51:58 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:51:58.515005 | orchestrator | 2025-08-29 17:51:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:01.554208 | orchestrator | 2025-08-29 17:52:01 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:01.554310 | orchestrator | 2025-08-29 17:52:01 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:01.554325 | orchestrator | 2025-08-29 17:52:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:04.594922 | orchestrator | 2025-08-29 17:52:04 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:04.595992 | orchestrator | 2025-08-29 17:52:04 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:04.596053 | orchestrator | 2025-08-29 17:52:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:07.636904 | orchestrator | 2025-08-29 17:52:07 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:07.636986 | orchestrator | 2025-08-29 17:52:07 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:07.636996 | orchestrator | 2025-08-29 17:52:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:10.676187 | orchestrator | 2025-08-29 17:52:10 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:10.678814 | orchestrator | 2025-08-29 17:52:10 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:10.678849 | orchestrator | 2025-08-29 17:52:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:13.724303 | orchestrator | 2025-08-29 17:52:13 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:13.725283 | orchestrator | 2025-08-29 17:52:13 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:13.725327 | orchestrator | 2025-08-29 17:52:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:16.769836 | orchestrator | 2025-08-29 17:52:16 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:16.771516 | orchestrator | 2025-08-29 17:52:16 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:16.771569 | orchestrator | 2025-08-29 17:52:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:19.808310 | orchestrator | 2025-08-29 17:52:19 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:19.809985 | orchestrator | 2025-08-29 17:52:19 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:19.810128 | orchestrator | 2025-08-29 17:52:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:22.852021 | orchestrator | 2025-08-29 17:52:22 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:22.852141 | orchestrator | 2025-08-29 17:52:22 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:22.852159 | orchestrator | 2025-08-29 17:52:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:25.892497 | orchestrator | 2025-08-29 17:52:25 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:25.894481 | orchestrator | 2025-08-29 17:52:25 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:25.894530 | orchestrator | 2025-08-29 17:52:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:28.928957 | orchestrator | 2025-08-29 17:52:28 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:28.929054 | orchestrator | 2025-08-29 17:52:28 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:28.929071 | orchestrator | 2025-08-29 17:52:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:31.965458 | orchestrator | 2025-08-29 17:52:31 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:31.967548 | orchestrator | 2025-08-29 17:52:31 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:31.967845 | orchestrator | 2025-08-29 17:52:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:35.022181 | orchestrator | 2025-08-29 17:52:35 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:35.022319 | orchestrator | 2025-08-29 17:52:35 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:35.022331 | orchestrator | 2025-08-29 17:52:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:38.070855 | orchestrator | 2025-08-29 17:52:38 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:38.072473 | orchestrator | 2025-08-29 17:52:38 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state STARTED 2025-08-29 17:52:38.072558 | orchestrator | 2025-08-29 17:52:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:41.102740 | orchestrator | 2025-08-29 17:52:41 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:41.105312 | orchestrator | 2025-08-29 17:52:41.105343 | orchestrator | 2025-08-29 17:52:41.105351 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:52:41.105359 | orchestrator | 2025-08-29 17:52:41.105366 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:52:41.105374 | orchestrator | Friday 29 August 2025 17:48:37 +0000 (0:00:00.171) 0:00:00.171 ********* 2025-08-29 17:52:41.105381 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:52:41.105389 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:52:41.105395 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:52:41.105418 | orchestrator | 2025-08-29 17:52:41.105425 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:52:41.105431 | orchestrator | Friday 29 August 2025 17:48:37 +0000 (0:00:00.285) 0:00:00.456 ********* 2025-08-29 17:52:41.105438 | orchestrator | ok: [testbed-node-0] => (item=enable_nova_True) 2025-08-29 17:52:41.105446 | orchestrator | ok: [testbed-node-1] => (item=enable_nova_True) 2025-08-29 17:52:41.105453 | orchestrator | ok: [testbed-node-2] => (item=enable_nova_True) 2025-08-29 17:52:41.105459 | orchestrator | 2025-08-29 17:52:41.105465 | orchestrator | PLAY [Wait for the Nova service] *********************************************** 2025-08-29 17:52:41.105471 | orchestrator | 2025-08-29 17:52:41.105478 | orchestrator | TASK [Waiting for Nova public port to be UP] *********************************** 2025-08-29 17:52:41.105508 | orchestrator | Friday 29 August 2025 17:48:38 +0000 (0:00:00.616) 0:00:01.073 ********* 2025-08-29 17:52:41.105514 | orchestrator | 2025-08-29 17:52:41.105521 | orchestrator | STILL ALIVE [task 'Waiting for Nova public port to be UP' is running] ********** 2025-08-29 17:52:41.105527 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:52:41.105533 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:52:41.105540 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:52:41.105546 | orchestrator | 2025-08-29 17:52:41.105552 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:52:41.105560 | orchestrator | testbed-node-0 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:52:41.105568 | orchestrator | testbed-node-1 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:52:41.105575 | orchestrator | testbed-node-2 : ok=3  changed=0 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:52:41.105581 | orchestrator | 2025-08-29 17:52:41.105587 | orchestrator | 2025-08-29 17:52:41.105593 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:52:41.105600 | orchestrator | Friday 29 August 2025 17:51:40 +0000 (0:03:01.818) 0:03:02.891 ********* 2025-08-29 17:52:41.105606 | orchestrator | =============================================================================== 2025-08-29 17:52:41.105613 | orchestrator | Waiting for Nova public port to be UP --------------------------------- 181.82s 2025-08-29 17:52:41.105619 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.62s 2025-08-29 17:52:41.105626 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.29s 2025-08-29 17:52:41.105632 | orchestrator | 2025-08-29 17:52:41.105638 | orchestrator | 2025-08-29 17:52:41.105644 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:52:41.105650 | orchestrator | 2025-08-29 17:52:41.105656 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:52:41.105662 | orchestrator | Friday 29 August 2025 17:50:21 +0000 (0:00:00.234) 0:00:00.234 ********* 2025-08-29 17:52:41.105668 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:52:41.105675 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:52:41.105747 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:52:41.105758 | orchestrator | 2025-08-29 17:52:41.105767 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:52:41.105777 | orchestrator | Friday 29 August 2025 17:50:21 +0000 (0:00:00.258) 0:00:00.492 ********* 2025-08-29 17:52:41.105786 | orchestrator | ok: [testbed-node-0] => (item=enable_grafana_True) 2025-08-29 17:52:41.105795 | orchestrator | ok: [testbed-node-1] => (item=enable_grafana_True) 2025-08-29 17:52:41.105803 | orchestrator | ok: [testbed-node-2] => (item=enable_grafana_True) 2025-08-29 17:52:41.105811 | orchestrator | 2025-08-29 17:52:41.105819 | orchestrator | PLAY [Apply role grafana] ****************************************************** 2025-08-29 17:52:41.105828 | orchestrator | 2025-08-29 17:52:41.105836 | orchestrator | TASK [grafana : include_tasks] ************************************************* 2025-08-29 17:52:41.105845 | orchestrator | Friday 29 August 2025 17:50:22 +0000 (0:00:00.343) 0:00:00.835 ********* 2025-08-29 17:52:41.105854 | orchestrator | included: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:52:41.105865 | orchestrator | 2025-08-29 17:52:41.105874 | orchestrator | TASK [grafana : Ensuring config directories exist] ***************************** 2025-08-29 17:52:41.105884 | orchestrator | Friday 29 August 2025 17:50:22 +0000 (0:00:00.449) 0:00:01.284 ********* 2025-08-29 17:52:41.105898 | orchestrator | changed: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.105937 | orchestrator | changed: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.105946 | orchestrator | changed: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106327 | orchestrator | 2025-08-29 17:52:41.106344 | orchestrator | TASK [grafana : Check if extra configuration file exists] ********************** 2025-08-29 17:52:41.106351 | orchestrator | Friday 29 August 2025 17:50:23 +0000 (0:00:00.674) 0:00:01.959 ********* 2025-08-29 17:52:41.106357 | orchestrator | [WARNING]: Skipped '/operations/prometheus/grafana' path due to this access 2025-08-29 17:52:41.106364 | orchestrator | issue: '/operations/prometheus/grafana' is not a directory 2025-08-29 17:52:41.106371 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:52:41.106377 | orchestrator | 2025-08-29 17:52:41.106383 | orchestrator | TASK [grafana : include_tasks] ************************************************* 2025-08-29 17:52:41.106389 | orchestrator | Friday 29 August 2025 17:50:23 +0000 (0:00:00.780) 0:00:02.740 ********* 2025-08-29 17:52:41.106470 | orchestrator | included: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:52:41.106655 | orchestrator | 2025-08-29 17:52:41.106666 | orchestrator | TASK [service-cert-copy : grafana | Copying over extra CA certificates] ******** 2025-08-29 17:52:41.106672 | orchestrator | Friday 29 August 2025 17:50:24 +0000 (0:00:00.727) 0:00:03.467 ********* 2025-08-29 17:52:41.106680 | orchestrator | changed: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106688 | orchestrator | changed: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106715 | orchestrator | changed: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106722 | orchestrator | 2025-08-29 17:52:41.106729 | orchestrator | TASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] *** 2025-08-29 17:52:41.106735 | orchestrator | Friday 29 August 2025 17:50:25 +0000 (0:00:01.323) 0:00:04.791 ********* 2025-08-29 17:52:41.106742 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:52:41.106749 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:52:41.106756 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:52:41.106763 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:52:41.106770 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:52:41.106776 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:52:41.106783 | orchestrator | 2025-08-29 17:52:41.106789 | orchestrator | TASK [service-cert-copy : grafana | Copying over backend internal TLS key] ***** 2025-08-29 17:52:41.106795 | orchestrator | Friday 29 August 2025 17:50:26 +0000 (0:00:00.349) 0:00:05.141 ********* 2025-08-29 17:52:41.106802 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:52:41.106814 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:52:41.106821 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:52:41.106827 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:52:41.106839 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})  2025-08-29 17:52:41.106847 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:52:41.106853 | orchestrator | 2025-08-29 17:52:41.106860 | orchestrator | TASK [grafana : Copying over config.json files] ******************************** 2025-08-29 17:52:41.106866 | orchestrator | Friday 29 August 2025 17:50:26 +0000 (0:00:00.620) 0:00:05.761 ********* 2025-08-29 17:52:41.106873 | orchestrator | changed: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106880 | orchestrator | changed: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106886 | orchestrator | changed: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106898 | orchestrator | 2025-08-29 17:52:41.106904 | orchestrator | TASK [grafana : Copying over grafana.ini] ************************************** 2025-08-29 17:52:41.106911 | orchestrator | Friday 29 August 2025 17:50:28 +0000 (0:00:01.153) 0:00:06.914 ********* 2025-08-29 17:52:41.106918 | orchestrator | changed: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106930 | orchestrator | changed: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106938 | orchestrator | changed: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.106944 | orchestrator | 2025-08-29 17:52:41.106951 | orchestrator | TASK [grafana : Copying over extra configuration file] ************************* 2025-08-29 17:52:41.106958 | orchestrator | Friday 29 August 2025 17:50:29 +0000 (0:00:01.229) 0:00:08.144 ********* 2025-08-29 17:52:41.106964 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:52:41.106971 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:52:41.106977 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:52:41.106984 | orchestrator | 2025-08-29 17:52:41.106990 | orchestrator | TASK [grafana : Configuring Prometheus as data source for Grafana] ************* 2025-08-29 17:52:41.106997 | orchestrator | Friday 29 August 2025 17:50:29 +0000 (0:00:00.364) 0:00:08.509 ********* 2025-08-29 17:52:41.107004 | orchestrator | changed: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2) 2025-08-29 17:52:41.107010 | orchestrator | changed: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2) 2025-08-29 17:52:41.107017 | orchestrator | changed: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2) 2025-08-29 17:52:41.107023 | orchestrator | 2025-08-29 17:52:41.107030 | orchestrator | TASK [grafana : Configuring dashboards provisioning] *************************** 2025-08-29 17:52:41.107036 | orchestrator | Friday 29 August 2025 17:50:30 +0000 (0:00:01.049) 0:00:09.559 ********* 2025-08-29 17:52:41.107047 | orchestrator | changed: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml) 2025-08-29 17:52:41.107054 | orchestrator | changed: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml) 2025-08-29 17:52:41.107060 | orchestrator | changed: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml) 2025-08-29 17:52:41.107067 | orchestrator | 2025-08-29 17:52:41.107073 | orchestrator | TASK [grafana : Find custom grafana dashboards] ******************************** 2025-08-29 17:52:41.107080 | orchestrator | Friday 29 August 2025 17:50:31 +0000 (0:00:01.090) 0:00:10.650 ********* 2025-08-29 17:52:41.107086 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:52:41.107093 | orchestrator | 2025-08-29 17:52:41.107099 | orchestrator | TASK [grafana : Find templated grafana dashboards] ***************************** 2025-08-29 17:52:41.107106 | orchestrator | Friday 29 August 2025 17:50:32 +0000 (0:00:00.675) 0:00:11.325 ********* 2025-08-29 17:52:41.107112 | orchestrator | [WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access 2025-08-29 17:52:41.107119 | orchestrator | issue: '/etc/kolla/grafana/dashboards' is not a directory 2025-08-29 17:52:41.107125 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:52:41.107132 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:52:41.107139 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:52:41.107145 | orchestrator | 2025-08-29 17:52:41.107152 | orchestrator | TASK [grafana : Prune templated Grafana dashboards] **************************** 2025-08-29 17:52:41.107158 | orchestrator | Friday 29 August 2025 17:50:33 +0000 (0:00:00.715) 0:00:12.041 ********* 2025-08-29 17:52:41.107165 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:52:41.107171 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:52:41.107178 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:52:41.107184 | orchestrator | 2025-08-29 17:52:41.107191 | orchestrator | TASK [grafana : Copying over custom dashboards] ******************************** 2025-08-29 17:52:41.107197 | orchestrator | Friday 29 August 2025 17:50:33 +0000 (0:00:00.419) 0:00:12.460 ********* 2025-08-29 17:52:41.107204 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1093910, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.547321, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107216 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1093910, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.547321, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107224 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 117836, 'inode': 1093910, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.547321, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107235 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1093955, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.559531, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107244 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1093955, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.559531, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107250 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25686, 'inode': 1093955, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.559531, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107257 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1093921, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5510788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107269 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1093921, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5510788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107276 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1093921, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5510788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107299 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1093957, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.562306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107307 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1093957, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.562306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107315 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 167897, 'inode': 1093957, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.562306, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107322 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1093935, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5549133, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107334 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1093935, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5549133, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107343 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26655, 'inode': 1093935, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5549133, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107350 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1093946, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5583537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107365 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1093946, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5583537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107373 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 39556, 'inode': 1093946, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5583537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107381 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1093908, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5461779, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107389 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1093908, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5461779, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107419 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1093908, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5461779, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107427 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1093914, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5487304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107443 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1093914, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5487304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107450 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1093914, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5487304, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107458 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1093924, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5510788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107465 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1093924, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5510788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107476 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1093924, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5510788, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107483 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1093938, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.556111, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107495 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1093938, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.556111, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107510 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19609, 'inode': 1093938, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.556111, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107518 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1093951, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5592587, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107525 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1093951, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5592587, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107533 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12997, 'inode': 1093951, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5592587, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107544 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1093917, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5502024, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107556 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1093917, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5502024, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107567 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1093917, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5502024, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107574 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1093943, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5577352, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107580 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1093943, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5577352, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107587 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19695, 'inode': 1093943, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5577352, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107597 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1093936, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5553598, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107608 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1093936, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5553598, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107617 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1093931, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5532525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107624 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38432, 'inode': 1093936, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5553598, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107631 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1093931, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5532525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107637 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1093928, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5529263, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107647 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 62676, 'inode': 1093931, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5532525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107658 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1093928, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5529263, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107665 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1093939, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5569687, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107675 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27218, 'inode': 1093928, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5529263, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107682 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1093939, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5569687, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107688 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1093925, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.551623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107694 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49139, 'inode': 1093939, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5569687, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107710 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1093925, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.551623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107717 | orchestrator | changed: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1093950, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5588686, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107727 | orchestrator | changed: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1093950, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5588686, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107734 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 44791, 'inode': 1093925, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.551623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107741 | orchestrator | changed: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1094064, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.594698, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107747 | orchestrator | changed: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16156, 'inode': 1093950, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5588686, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107761 | orchestrator | changed: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1094064, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.594698, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107768 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1093988, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.572289, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107778 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1093988, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.572289, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107784 | orchestrator | changed: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1094064, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.594698, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107791 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1093977, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5652058, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107797 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1093977, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5652058, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107811 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1093988, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.572289, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107818 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1094014, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5767994, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107825 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1094014, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5767994, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107835 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1093977, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5652058, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107841 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1093968, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5627925, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107850 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1093968, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5627925, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107860 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15725, 'inode': 1094014, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5767994, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107871 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1094036, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5862732, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107877 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1093968, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5627925, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107887 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1094036, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5862732, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107894 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1094016, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5844736, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107900 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1094016, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5844736, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107913 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1094040, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5875182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107923 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1094036, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5862732, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107930 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1094040, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5875182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107940 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1094059, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5936184, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107947 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1094016, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5844736, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107953 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1094059, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5936184, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.107964 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1094034, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5853202, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108008 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22317, 'inode': 1094040, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5875182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108016 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1094012, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5757644, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108026 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1094034, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5853202, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108032 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1094059, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5936184, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108039 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1093982, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5683086, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108049 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1094012, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5757644, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108061 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21109, 'inode': 1094034, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5853202, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108067 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1093982, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5683086, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108077 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1094001, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5753725, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108084 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1094012, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5757644, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108090 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1094001, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5753725, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108101 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1093980, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5659454, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108111 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1093982, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5683086, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108118 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1093980, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5659454, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108124 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1094013, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.576509, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108134 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1094001, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5753725, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108140 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1094051, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5929232, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108151 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1094013, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.576509, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108157 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1093980, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5659454, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108167 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1094045, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5887885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108174 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1094051, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5929232, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108184 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16098, 'inode': 1094013, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.576509, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108190 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1093972, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5641463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108201 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1094051, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5929232, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108207 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1094045, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5887885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108218 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1093974, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5646741, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108225 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1093972, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5641463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108235 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1094045, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5887885, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108242 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1094031, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5853202, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108252 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1093972, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5641463, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108259 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1093974, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5646741, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108269 | orchestrator | changed: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1094042, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5875182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108276 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1093974, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5646741, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108283 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1094031, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5853202, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108290 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1094031, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5853202, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108300 | orchestrator | changed: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1094042, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5875182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108307 | orchestrator | changed: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21898, 'inode': 1094042, 'dev': 103, 'nlink': 1, 'atime': 1752315970.0, 'mtime': 1752315970.0, 'ctime': 1756486870.5875182, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}}) 2025-08-29 17:52:41.108313 | orchestrator | 2025-08-29 17:52:41.108319 | orchestrator | TASK [grafana : Check grafana containers] ************************************** 2025-08-29 17:52:41.108325 | orchestrator | Friday 29 August 2025 17:51:10 +0000 (0:00:37.085) 0:00:49.546 ********* 2025-08-29 17:52:41.108335 | orchestrator | changed: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.108373 | orchestrator | changed: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.108385 | orchestrator | changed: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/grafana:12.0.2.20250711', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) 2025-08-29 17:52:41.108395 | orchestrator | 2025-08-29 17:52:41.108447 | orchestrator | TASK [grafana : Creating grafana database] ************************************* 2025-08-29 17:52:41.108454 | orchestrator | Friday 29 August 2025 17:51:11 +0000 (0:00:01.159) 0:00:50.706 ********* 2025-08-29 17:52:41.108460 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:52:41.108467 | orchestrator | 2025-08-29 17:52:41.108473 | orchestrator | TASK [grafana : Creating grafana database user and setting permissions] ******** 2025-08-29 17:52:41.108479 | orchestrator | Friday 29 August 2025 17:51:14 +0000 (0:00:02.160) 0:00:52.867 ********* 2025-08-29 17:52:41.108485 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:52:41.108491 | orchestrator | 2025-08-29 17:52:41.108497 | orchestrator | TASK [grafana : Flush handlers] ************************************************ 2025-08-29 17:52:41.108503 | orchestrator | Friday 29 August 2025 17:51:15 +0000 (0:00:01.921) 0:00:54.788 ********* 2025-08-29 17:52:41.108510 | orchestrator | 2025-08-29 17:52:41.108516 | orchestrator | TASK [grafana : Flush handlers] ************************************************ 2025-08-29 17:52:41.108522 | orchestrator | Friday 29 August 2025 17:51:16 +0000 (0:00:00.226) 0:00:55.014 ********* 2025-08-29 17:52:41.108528 | orchestrator | 2025-08-29 17:52:41.108534 | orchestrator | TASK [grafana : Flush handlers] ************************************************ 2025-08-29 17:52:41.108540 | orchestrator | Friday 29 August 2025 17:51:16 +0000 (0:00:00.063) 0:00:55.077 ********* 2025-08-29 17:52:41.108546 | orchestrator | 2025-08-29 17:52:41.108552 | orchestrator | RUNNING HANDLER [grafana : Restart first grafana container] ******************** 2025-08-29 17:52:41.108559 | orchestrator | Friday 29 August 2025 17:51:16 +0000 (0:00:00.067) 0:00:55.145 ********* 2025-08-29 17:52:41.108565 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:52:41.108571 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:52:41.108577 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:52:41.108583 | orchestrator | 2025-08-29 17:52:41.108589 | orchestrator | RUNNING HANDLER [grafana : Waiting for grafana to start on first node] ********* 2025-08-29 17:52:41.108595 | orchestrator | Friday 29 August 2025 17:51:23 +0000 (0:00:06.740) 0:01:01.886 ********* 2025-08-29 17:52:41.108602 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:52:41.108608 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:52:41.108614 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left). 2025-08-29 17:52:41.108620 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (11 retries left). 2025-08-29 17:52:41.108626 | orchestrator | FAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (10 retries left). 2025-08-29 17:52:41.108632 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:52:41.108638 | orchestrator | 2025-08-29 17:52:41.108644 | orchestrator | RUNNING HANDLER [grafana : Restart remaining grafana containers] *************** 2025-08-29 17:52:41.108651 | orchestrator | Friday 29 August 2025 17:52:01 +0000 (0:00:38.134) 0:01:40.020 ********* 2025-08-29 17:52:41.108657 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:52:41.108663 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:52:41.108669 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:52:41.108675 | orchestrator | 2025-08-29 17:52:41.108681 | orchestrator | TASK [grafana : Wait for grafana application ready] **************************** 2025-08-29 17:52:41.108687 | orchestrator | Friday 29 August 2025 17:52:33 +0000 (0:00:32.604) 0:02:12.625 ********* 2025-08-29 17:52:41.108693 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:52:41.108699 | orchestrator | 2025-08-29 17:52:41.108706 | orchestrator | TASK [grafana : Remove old grafana docker volume] ****************************** 2025-08-29 17:52:41.108715 | orchestrator | Friday 29 August 2025 17:52:36 +0000 (0:00:02.283) 0:02:14.909 ********* 2025-08-29 17:52:41.108722 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:52:41.108728 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:52:41.108734 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:52:41.108745 | orchestrator | 2025-08-29 17:52:41.108751 | orchestrator | TASK [grafana : Enable grafana datasources] ************************************ 2025-08-29 17:52:41.108757 | orchestrator | Friday 29 August 2025 17:52:36 +0000 (0:00:00.712) 0:02:15.622 ********* 2025-08-29 17:52:41.108764 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}})  2025-08-29 17:52:41.108772 | orchestrator | changed: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}}) 2025-08-29 17:52:41.108778 | orchestrator | 2025-08-29 17:52:41.108785 | orchestrator | TASK [grafana : Disable Getting Started panel] ********************************* 2025-08-29 17:52:41.108791 | orchestrator | Friday 29 August 2025 17:52:39 +0000 (0:00:02.708) 0:02:18.330 ********* 2025-08-29 17:52:41.108797 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:52:41.108803 | orchestrator | 2025-08-29 17:52:41.108809 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:52:41.108819 | orchestrator | testbed-node-0 : ok=21  changed=12  unreachable=0 failed=0 skipped=7  rescued=0 ignored=0 2025-08-29 17:52:41.108828 | orchestrator | testbed-node-1 : ok=14  changed=9  unreachable=0 failed=0 skipped=7  rescued=0 ignored=0 2025-08-29 17:52:41.108834 | orchestrator | testbed-node-2 : ok=14  changed=9  unreachable=0 failed=0 skipped=7  rescued=0 ignored=0 2025-08-29 17:52:41.108840 | orchestrator | 2025-08-29 17:52:41.108846 | orchestrator | 2025-08-29 17:52:41.108852 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:52:41.108859 | orchestrator | Friday 29 August 2025 17:52:39 +0000 (0:00:00.442) 0:02:18.772 ********* 2025-08-29 17:52:41.108865 | orchestrator | =============================================================================== 2025-08-29 17:52:41.108871 | orchestrator | grafana : Waiting for grafana to start on first node ------------------- 38.13s 2025-08-29 17:52:41.108877 | orchestrator | grafana : Copying over custom dashboards ------------------------------- 37.09s 2025-08-29 17:52:41.108883 | orchestrator | grafana : Restart remaining grafana containers ------------------------- 32.60s 2025-08-29 17:52:41.108889 | orchestrator | grafana : Restart first grafana container ------------------------------- 6.74s 2025-08-29 17:52:41.108895 | orchestrator | grafana : Enable grafana datasources ------------------------------------ 2.71s 2025-08-29 17:52:41.108901 | orchestrator | grafana : Wait for grafana application ready ---------------------------- 2.28s 2025-08-29 17:52:41.108908 | orchestrator | grafana : Creating grafana database ------------------------------------- 2.16s 2025-08-29 17:52:41.108914 | orchestrator | grafana : Creating grafana database user and setting permissions -------- 1.92s 2025-08-29 17:52:41.108920 | orchestrator | service-cert-copy : grafana | Copying over extra CA certificates -------- 1.32s 2025-08-29 17:52:41.108926 | orchestrator | grafana : Copying over grafana.ini -------------------------------------- 1.23s 2025-08-29 17:52:41.108932 | orchestrator | grafana : Check grafana containers -------------------------------------- 1.16s 2025-08-29 17:52:41.108938 | orchestrator | grafana : Copying over config.json files -------------------------------- 1.15s 2025-08-29 17:52:41.108944 | orchestrator | grafana : Configuring dashboards provisioning --------------------------- 1.09s 2025-08-29 17:52:41.108950 | orchestrator | grafana : Configuring Prometheus as data source for Grafana ------------- 1.05s 2025-08-29 17:52:41.108956 | orchestrator | grafana : Check if extra configuration file exists ---------------------- 0.78s 2025-08-29 17:52:41.108962 | orchestrator | grafana : include_tasks ------------------------------------------------- 0.73s 2025-08-29 17:52:41.108972 | orchestrator | grafana : Find templated grafana dashboards ----------------------------- 0.72s 2025-08-29 17:52:41.108979 | orchestrator | grafana : Remove old grafana docker volume ------------------------------ 0.71s 2025-08-29 17:52:41.108985 | orchestrator | grafana : Find custom grafana dashboards -------------------------------- 0.68s 2025-08-29 17:52:41.108991 | orchestrator | grafana : Ensuring config directories exist ----------------------------- 0.67s 2025-08-29 17:52:41.108997 | orchestrator | 2025-08-29 17:52:41 | INFO  | Task 2db7a092-41af-4611-bd9c-2e4e92a9f77f is in state SUCCESS 2025-08-29 17:52:41.109003 | orchestrator | 2025-08-29 17:52:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:44.153207 | orchestrator | 2025-08-29 17:52:44 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:44.153313 | orchestrator | 2025-08-29 17:52:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:47.182712 | orchestrator | 2025-08-29 17:52:47 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:47.182789 | orchestrator | 2025-08-29 17:52:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:50.213555 | orchestrator | 2025-08-29 17:52:50 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:50.213606 | orchestrator | 2025-08-29 17:52:50 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:53.254673 | orchestrator | 2025-08-29 17:52:53 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:53.254762 | orchestrator | 2025-08-29 17:52:53 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:56.296729 | orchestrator | 2025-08-29 17:52:56 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:56.296818 | orchestrator | 2025-08-29 17:52:56 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:52:59.334147 | orchestrator | 2025-08-29 17:52:59 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:52:59.334248 | orchestrator | 2025-08-29 17:52:59 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:02.373532 | orchestrator | 2025-08-29 17:53:02 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:02.373641 | orchestrator | 2025-08-29 17:53:02 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:05.412235 | orchestrator | 2025-08-29 17:53:05 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:05.412359 | orchestrator | 2025-08-29 17:53:05 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:08.456960 | orchestrator | 2025-08-29 17:53:08 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:08.457064 | orchestrator | 2025-08-29 17:53:08 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:11.492935 | orchestrator | 2025-08-29 17:53:11 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:11.493033 | orchestrator | 2025-08-29 17:53:11 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:14.533709 | orchestrator | 2025-08-29 17:53:14 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:14.533806 | orchestrator | 2025-08-29 17:53:14 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:17.580457 | orchestrator | 2025-08-29 17:53:17 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:17.580616 | orchestrator | 2025-08-29 17:53:17 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:20.622925 | orchestrator | 2025-08-29 17:53:20 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:20.623964 | orchestrator | 2025-08-29 17:53:20 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:23.664269 | orchestrator | 2025-08-29 17:53:23 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:23.664364 | orchestrator | 2025-08-29 17:53:23 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:26.706302 | orchestrator | 2025-08-29 17:53:26 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:26.706455 | orchestrator | 2025-08-29 17:53:26 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:29.741907 | orchestrator | 2025-08-29 17:53:29 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:29.742006 | orchestrator | 2025-08-29 17:53:29 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:32.786598 | orchestrator | 2025-08-29 17:53:32 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:32.786721 | orchestrator | 2025-08-29 17:53:32 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:35.822930 | orchestrator | 2025-08-29 17:53:35 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:35.823049 | orchestrator | 2025-08-29 17:53:35 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:38.868240 | orchestrator | 2025-08-29 17:53:38 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:38.868439 | orchestrator | 2025-08-29 17:53:38 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:41.906159 | orchestrator | 2025-08-29 17:53:41 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:41.906261 | orchestrator | 2025-08-29 17:53:41 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:44.948015 | orchestrator | 2025-08-29 17:53:44 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:44.948135 | orchestrator | 2025-08-29 17:53:44 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:47.988272 | orchestrator | 2025-08-29 17:53:47 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:47.988398 | orchestrator | 2025-08-29 17:53:47 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:51.036067 | orchestrator | 2025-08-29 17:53:51 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:51.036172 | orchestrator | 2025-08-29 17:53:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:54.073722 | orchestrator | 2025-08-29 17:53:54 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:54.073802 | orchestrator | 2025-08-29 17:53:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:53:57.115276 | orchestrator | 2025-08-29 17:53:57 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:53:57.115450 | orchestrator | 2025-08-29 17:53:57 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:00.156169 | orchestrator | 2025-08-29 17:54:00 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:00.156268 | orchestrator | 2025-08-29 17:54:00 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:03.204610 | orchestrator | 2025-08-29 17:54:03 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:03.204779 | orchestrator | 2025-08-29 17:54:03 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:06.250448 | orchestrator | 2025-08-29 17:54:06 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:06.250633 | orchestrator | 2025-08-29 17:54:06 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:09.294240 | orchestrator | 2025-08-29 17:54:09 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:09.294345 | orchestrator | 2025-08-29 17:54:09 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:12.332450 | orchestrator | 2025-08-29 17:54:12 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:12.333229 | orchestrator | 2025-08-29 17:54:12 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:15.385961 | orchestrator | 2025-08-29 17:54:15 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:15.386092 | orchestrator | 2025-08-29 17:54:15 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:18.430610 | orchestrator | 2025-08-29 17:54:18 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:18.430714 | orchestrator | 2025-08-29 17:54:18 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:21.479213 | orchestrator | 2025-08-29 17:54:21 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:21.479417 | orchestrator | 2025-08-29 17:54:21 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:24.524557 | orchestrator | 2025-08-29 17:54:24 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:24.524691 | orchestrator | 2025-08-29 17:54:24 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:27.566125 | orchestrator | 2025-08-29 17:54:27 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:27.566283 | orchestrator | 2025-08-29 17:54:27 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:30.613541 | orchestrator | 2025-08-29 17:54:30 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:30.613627 | orchestrator | 2025-08-29 17:54:30 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:33.664606 | orchestrator | 2025-08-29 17:54:33 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:33.664736 | orchestrator | 2025-08-29 17:54:33 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:36.704491 | orchestrator | 2025-08-29 17:54:36 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:36.704594 | orchestrator | 2025-08-29 17:54:36 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:39.749661 | orchestrator | 2025-08-29 17:54:39 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:39.749720 | orchestrator | 2025-08-29 17:54:39 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:42.784900 | orchestrator | 2025-08-29 17:54:42 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:42.784984 | orchestrator | 2025-08-29 17:54:42 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:45.827731 | orchestrator | 2025-08-29 17:54:45 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:45.827808 | orchestrator | 2025-08-29 17:54:45 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:48.872128 | orchestrator | 2025-08-29 17:54:48 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:48.872229 | orchestrator | 2025-08-29 17:54:48 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:51.922535 | orchestrator | 2025-08-29 17:54:51 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:51.922660 | orchestrator | 2025-08-29 17:54:51 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:54.966450 | orchestrator | 2025-08-29 17:54:54 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:54.966577 | orchestrator | 2025-08-29 17:54:54 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:54:58.011839 | orchestrator | 2025-08-29 17:54:58 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:54:58.011939 | orchestrator | 2025-08-29 17:54:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:01.047085 | orchestrator | 2025-08-29 17:55:01 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:01.047205 | orchestrator | 2025-08-29 17:55:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:04.087113 | orchestrator | 2025-08-29 17:55:04 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:04.087222 | orchestrator | 2025-08-29 17:55:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:07.126421 | orchestrator | 2025-08-29 17:55:07 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:07.126493 | orchestrator | 2025-08-29 17:55:07 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:10.173188 | orchestrator | 2025-08-29 17:55:10 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:10.173273 | orchestrator | 2025-08-29 17:55:10 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:13.219721 | orchestrator | 2025-08-29 17:55:13 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:13.219817 | orchestrator | 2025-08-29 17:55:13 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:16.265838 | orchestrator | 2025-08-29 17:55:16 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:16.265953 | orchestrator | 2025-08-29 17:55:16 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:19.308728 | orchestrator | 2025-08-29 17:55:19 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:19.308834 | orchestrator | 2025-08-29 17:55:19 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:22.353589 | orchestrator | 2025-08-29 17:55:22 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:22.353690 | orchestrator | 2025-08-29 17:55:22 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:25.385531 | orchestrator | 2025-08-29 17:55:25 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:25.385610 | orchestrator | 2025-08-29 17:55:25 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:28.413406 | orchestrator | 2025-08-29 17:55:28 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:28.413492 | orchestrator | 2025-08-29 17:55:28 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:31.442722 | orchestrator | 2025-08-29 17:55:31 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:31.442828 | orchestrator | 2025-08-29 17:55:31 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:34.488194 | orchestrator | 2025-08-29 17:55:34 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:34.488261 | orchestrator | 2025-08-29 17:55:34 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:37.532399 | orchestrator | 2025-08-29 17:55:37 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:37.532513 | orchestrator | 2025-08-29 17:55:37 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:40.577436 | orchestrator | 2025-08-29 17:55:40 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:40.577536 | orchestrator | 2025-08-29 17:55:40 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:43.625079 | orchestrator | 2025-08-29 17:55:43 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:43.625185 | orchestrator | 2025-08-29 17:55:43 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:46.674928 | orchestrator | 2025-08-29 17:55:46 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:46.675055 | orchestrator | 2025-08-29 17:55:46 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:49.718681 | orchestrator | 2025-08-29 17:55:49 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:49.718777 | orchestrator | 2025-08-29 17:55:49 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:52.763442 | orchestrator | 2025-08-29 17:55:52 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:52.763548 | orchestrator | 2025-08-29 17:55:52 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:55.811598 | orchestrator | 2025-08-29 17:55:55 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:55.811698 | orchestrator | 2025-08-29 17:55:55 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:55:58.853367 | orchestrator | 2025-08-29 17:55:58 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:55:58.853493 | orchestrator | 2025-08-29 17:55:58 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:56:01.902480 | orchestrator | 2025-08-29 17:56:01 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:56:01.902576 | orchestrator | 2025-08-29 17:56:01 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:56:04.951200 | orchestrator | 2025-08-29 17:56:04 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state STARTED 2025-08-29 17:56:04.951304 | orchestrator | 2025-08-29 17:56:04 | INFO  | Wait 1 second(s) until the next check 2025-08-29 17:56:08.000788 | orchestrator | 2025-08-29 17:56:07 | INFO  | Task 4badeaf3-5c81-483d-a5ab-c4bd2969eace is in state SUCCESS 2025-08-29 17:56:08.002767 | orchestrator | 2025-08-29 17:56:08.002833 | orchestrator | 2025-08-29 17:56:08.002848 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 17:56:08.002860 | orchestrator | 2025-08-29 17:56:08.002872 | orchestrator | TASK [Group hosts based on OpenStack release] ********************************** 2025-08-29 17:56:08.002892 | orchestrator | Friday 29 August 2025 17:47:30 +0000 (0:00:00.391) 0:00:00.391 ********* 2025-08-29 17:56:08.002911 | orchestrator | changed: [testbed-manager] 2025-08-29 17:56:08.002934 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.002955 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.002972 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.002990 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.003009 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.003027 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.003046 | orchestrator | 2025-08-29 17:56:08.003065 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 17:56:08.003084 | orchestrator | Friday 29 August 2025 17:47:31 +0000 (0:00:00.743) 0:00:01.134 ********* 2025-08-29 17:56:08.003104 | orchestrator | changed: [testbed-manager] 2025-08-29 17:56:08.003122 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.003179 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.003199 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.003290 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.003304 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.003315 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.003326 | orchestrator | 2025-08-29 17:56:08.003375 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 17:56:08.003389 | orchestrator | Friday 29 August 2025 17:47:32 +0000 (0:00:01.076) 0:00:02.210 ********* 2025-08-29 17:56:08.003418 | orchestrator | changed: [testbed-manager] => (item=enable_nova_True) 2025-08-29 17:56:08.003431 | orchestrator | changed: [testbed-node-0] => (item=enable_nova_True) 2025-08-29 17:56:08.003445 | orchestrator | changed: [testbed-node-1] => (item=enable_nova_True) 2025-08-29 17:56:08.003465 | orchestrator | changed: [testbed-node-2] => (item=enable_nova_True) 2025-08-29 17:56:08.003484 | orchestrator | changed: [testbed-node-3] => (item=enable_nova_True) 2025-08-29 17:56:08.003504 | orchestrator | changed: [testbed-node-4] => (item=enable_nova_True) 2025-08-29 17:56:08.003524 | orchestrator | changed: [testbed-node-5] => (item=enable_nova_True) 2025-08-29 17:56:08.003542 | orchestrator | 2025-08-29 17:56:08.003682 | orchestrator | PLAY [Bootstrap nova API databases] ******************************************** 2025-08-29 17:56:08.003704 | orchestrator | 2025-08-29 17:56:08.003723 | orchestrator | TASK [Bootstrap deploy] ******************************************************** 2025-08-29 17:56:08.003741 | orchestrator | Friday 29 August 2025 17:47:33 +0000 (0:00:01.341) 0:00:03.552 ********* 2025-08-29 17:56:08.003761 | orchestrator | included: nova for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:56:08.003781 | orchestrator | 2025-08-29 17:56:08.003796 | orchestrator | TASK [nova : Creating Nova databases] ****************************************** 2025-08-29 17:56:08.003808 | orchestrator | Friday 29 August 2025 17:47:34 +0000 (0:00:00.535) 0:00:04.088 ********* 2025-08-29 17:56:08.003820 | orchestrator | changed: [testbed-node-0] => (item=nova_cell0) 2025-08-29 17:56:08.003832 | orchestrator | changed: [testbed-node-0] => (item=nova_api) 2025-08-29 17:56:08.003843 | orchestrator | 2025-08-29 17:56:08.003854 | orchestrator | TASK [nova : Creating Nova databases user and setting permissions] ************* 2025-08-29 17:56:08.003865 | orchestrator | Friday 29 August 2025 17:47:38 +0000 (0:00:04.571) 0:00:08.659 ********* 2025-08-29 17:56:08.003876 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:56:08.003887 | orchestrator | changed: [testbed-node-0] => (item=None) 2025-08-29 17:56:08.003897 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.003911 | orchestrator | 2025-08-29 17:56:08.003929 | orchestrator | TASK [nova : Ensuring config directories exist] ******************************** 2025-08-29 17:56:08.003948 | orchestrator | Friday 29 August 2025 17:47:43 +0000 (0:00:04.653) 0:00:13.312 ********* 2025-08-29 17:56:08.003966 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.003985 | orchestrator | 2025-08-29 17:56:08.004005 | orchestrator | TASK [nova : Copying over config.json files for nova-api-bootstrap] ************ 2025-08-29 17:56:08.004016 | orchestrator | Friday 29 August 2025 17:47:44 +0000 (0:00:01.337) 0:00:14.650 ********* 2025-08-29 17:56:08.004027 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.004038 | orchestrator | 2025-08-29 17:56:08.004049 | orchestrator | TASK [nova : Copying over nova.conf for nova-api-bootstrap] ******************** 2025-08-29 17:56:08.004060 | orchestrator | Friday 29 August 2025 17:47:47 +0000 (0:00:02.386) 0:00:17.036 ********* 2025-08-29 17:56:08.004071 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.004086 | orchestrator | 2025-08-29 17:56:08.004104 | orchestrator | TASK [nova : include_tasks] **************************************************** 2025-08-29 17:56:08.004123 | orchestrator | Friday 29 August 2025 17:47:51 +0000 (0:00:04.043) 0:00:21.080 ********* 2025-08-29 17:56:08.004142 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.004161 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.004179 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.004197 | orchestrator | 2025-08-29 17:56:08.004232 | orchestrator | TASK [nova : Running Nova API bootstrap container] ***************************** 2025-08-29 17:56:08.004259 | orchestrator | Friday 29 August 2025 17:47:51 +0000 (0:00:00.558) 0:00:21.639 ********* 2025-08-29 17:56:08.004278 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.004297 | orchestrator | 2025-08-29 17:56:08.004315 | orchestrator | TASK [nova : Create cell0 mappings] ******************************************** 2025-08-29 17:56:08.004362 | orchestrator | Friday 29 August 2025 17:48:26 +0000 (0:00:34.452) 0:00:56.091 ********* 2025-08-29 17:56:08.004382 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.004401 | orchestrator | 2025-08-29 17:56:08.004420 | orchestrator | TASK [nova-cell : Get a list of existing cells] ******************************** 2025-08-29 17:56:08.004432 | orchestrator | Friday 29 August 2025 17:48:43 +0000 (0:00:17.183) 0:01:13.275 ********* 2025-08-29 17:56:08.004443 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.004454 | orchestrator | 2025-08-29 17:56:08.004465 | orchestrator | TASK [nova-cell : Extract current cell settings from list] ********************* 2025-08-29 17:56:08.004476 | orchestrator | Friday 29 August 2025 17:48:54 +0000 (0:00:11.309) 0:01:24.584 ********* 2025-08-29 17:56:08.004505 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.004517 | orchestrator | 2025-08-29 17:56:08.004530 | orchestrator | TASK [nova : Update cell0 mappings] ******************************************** 2025-08-29 17:56:08.004549 | orchestrator | Friday 29 August 2025 17:48:55 +0000 (0:00:01.199) 0:01:25.784 ********* 2025-08-29 17:56:08.004567 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.004586 | orchestrator | 2025-08-29 17:56:08.004605 | orchestrator | TASK [nova : include_tasks] **************************************************** 2025-08-29 17:56:08.004624 | orchestrator | Friday 29 August 2025 17:48:56 +0000 (0:00:00.534) 0:01:26.319 ********* 2025-08-29 17:56:08.004644 | orchestrator | included: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:56:08.004662 | orchestrator | 2025-08-29 17:56:08.004680 | orchestrator | TASK [nova : Running Nova API bootstrap container] ***************************** 2025-08-29 17:56:08.004691 | orchestrator | Friday 29 August 2025 17:48:56 +0000 (0:00:00.509) 0:01:26.828 ********* 2025-08-29 17:56:08.004702 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.004719 | orchestrator | 2025-08-29 17:56:08.004737 | orchestrator | TASK [Bootstrap upgrade] ******************************************************* 2025-08-29 17:56:08.004756 | orchestrator | Friday 29 August 2025 17:49:14 +0000 (0:00:17.787) 0:01:44.616 ********* 2025-08-29 17:56:08.004774 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.004789 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.004800 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.004810 | orchestrator | 2025-08-29 17:56:08.004821 | orchestrator | PLAY [Bootstrap nova cell databases] ******************************************* 2025-08-29 17:56:08.004831 | orchestrator | 2025-08-29 17:56:08.004842 | orchestrator | TASK [Bootstrap deploy] ******************************************************** 2025-08-29 17:56:08.004853 | orchestrator | Friday 29 August 2025 17:49:14 +0000 (0:00:00.268) 0:01:44.884 ********* 2025-08-29 17:56:08.004864 | orchestrator | included: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:56:08.004881 | orchestrator | 2025-08-29 17:56:08.004898 | orchestrator | TASK [nova-cell : Creating Nova cell database] ********************************* 2025-08-29 17:56:08.004916 | orchestrator | Friday 29 August 2025 17:49:15 +0000 (0:00:00.541) 0:01:45.426 ********* 2025-08-29 17:56:08.004935 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.004954 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.004973 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.004991 | orchestrator | 2025-08-29 17:56:08.005009 | orchestrator | TASK [nova-cell : Creating Nova cell database user and setting permissions] **** 2025-08-29 17:56:08.005028 | orchestrator | Friday 29 August 2025 17:49:17 +0000 (0:00:02.079) 0:01:47.505 ********* 2025-08-29 17:56:08.005047 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005066 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005078 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.005098 | orchestrator | 2025-08-29 17:56:08.005110 | orchestrator | TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ****************** 2025-08-29 17:56:08.005120 | orchestrator | Friday 29 August 2025 17:49:19 +0000 (0:00:02.115) 0:01:49.621 ********* 2025-08-29 17:56:08.005131 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.005142 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005153 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005164 | orchestrator | 2025-08-29 17:56:08.005174 | orchestrator | TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] ******************* 2025-08-29 17:56:08.005192 | orchestrator | Friday 29 August 2025 17:49:19 +0000 (0:00:00.308) 0:01:49.929 ********* 2025-08-29 17:56:08.005211 | orchestrator | skipping: [testbed-node-1] => (item=None)  2025-08-29 17:56:08.005230 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005249 | orchestrator | skipping: [testbed-node-2] => (item=None)  2025-08-29 17:56:08.005267 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005283 | orchestrator | ok: [testbed-node-0] => (item=None) 2025-08-29 17:56:08.005294 | orchestrator | ok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}] 2025-08-29 17:56:08.005304 | orchestrator | 2025-08-29 17:56:08.005315 | orchestrator | TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ****************** 2025-08-29 17:56:08.005326 | orchestrator | Friday 29 August 2025 17:49:28 +0000 (0:00:08.514) 0:01:58.444 ********* 2025-08-29 17:56:08.005370 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.005389 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005407 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005426 | orchestrator | 2025-08-29 17:56:08.005444 | orchestrator | TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] ******************* 2025-08-29 17:56:08.005459 | orchestrator | Friday 29 August 2025 17:49:28 +0000 (0:00:00.300) 0:01:58.745 ********* 2025-08-29 17:56:08.005470 | orchestrator | skipping: [testbed-node-0] => (item=None)  2025-08-29 17:56:08.005481 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.005491 | orchestrator | skipping: [testbed-node-1] => (item=None)  2025-08-29 17:56:08.005502 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005513 | orchestrator | skipping: [testbed-node-2] => (item=None)  2025-08-29 17:56:08.005524 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005534 | orchestrator | 2025-08-29 17:56:08.005545 | orchestrator | TASK [nova-cell : Ensuring config directories exist] *************************** 2025-08-29 17:56:08.005556 | orchestrator | Friday 29 August 2025 17:49:29 +0000 (0:00:00.567) 0:01:59.312 ********* 2025-08-29 17:56:08.005574 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005585 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005596 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.005606 | orchestrator | 2025-08-29 17:56:08.005621 | orchestrator | TASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ****** 2025-08-29 17:56:08.005640 | orchestrator | Friday 29 August 2025 17:49:29 +0000 (0:00:00.447) 0:01:59.759 ********* 2025-08-29 17:56:08.005659 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005679 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005698 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.005716 | orchestrator | 2025-08-29 17:56:08.005731 | orchestrator | TASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] ************** 2025-08-29 17:56:08.005742 | orchestrator | Friday 29 August 2025 17:49:30 +0000 (0:00:01.006) 0:02:00.766 ********* 2025-08-29 17:56:08.005753 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005764 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005784 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.005795 | orchestrator | 2025-08-29 17:56:08.005806 | orchestrator | TASK [nova-cell : Running Nova cell bootstrap container] *********************** 2025-08-29 17:56:08.005816 | orchestrator | Friday 29 August 2025 17:49:32 +0000 (0:00:02.115) 0:02:02.881 ********* 2025-08-29 17:56:08.005827 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005838 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005848 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.005868 | orchestrator | 2025-08-29 17:56:08.005879 | orchestrator | TASK [nova-cell : Get a list of existing cells] ******************************** 2025-08-29 17:56:08.005890 | orchestrator | Friday 29 August 2025 17:49:55 +0000 (0:00:22.545) 0:02:25.427 ********* 2025-08-29 17:56:08.005901 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005911 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005922 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.005933 | orchestrator | 2025-08-29 17:56:08.005943 | orchestrator | TASK [nova-cell : Extract current cell settings from list] ********************* 2025-08-29 17:56:08.005954 | orchestrator | Friday 29 August 2025 17:50:05 +0000 (0:00:10.432) 0:02:35.860 ********* 2025-08-29 17:56:08.005965 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.005975 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.005986 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.005997 | orchestrator | 2025-08-29 17:56:08.006008 | orchestrator | TASK [nova-cell : Create cell] ************************************************* 2025-08-29 17:56:08.006069 | orchestrator | Friday 29 August 2025 17:50:06 +0000 (0:00:00.811) 0:02:36.671 ********* 2025-08-29 17:56:08.006084 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.006095 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.006106 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.006116 | orchestrator | 2025-08-29 17:56:08.006127 | orchestrator | TASK [nova-cell : Update cell] ************************************************* 2025-08-29 17:56:08.006138 | orchestrator | Friday 29 August 2025 17:50:19 +0000 (0:00:12.610) 0:02:49.282 ********* 2025-08-29 17:56:08.006149 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.006160 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.006171 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.006182 | orchestrator | 2025-08-29 17:56:08.006193 | orchestrator | TASK [Bootstrap upgrade] ******************************************************* 2025-08-29 17:56:08.006204 | orchestrator | Friday 29 August 2025 17:50:20 +0000 (0:00:01.236) 0:02:50.518 ********* 2025-08-29 17:56:08.006215 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.006225 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.006236 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.006247 | orchestrator | 2025-08-29 17:56:08.006258 | orchestrator | PLAY [Apply role nova] ********************************************************* 2025-08-29 17:56:08.006268 | orchestrator | 2025-08-29 17:56:08.006279 | orchestrator | TASK [nova : include_tasks] **************************************************** 2025-08-29 17:56:08.006290 | orchestrator | Friday 29 August 2025 17:50:20 +0000 (0:00:00.282) 0:02:50.800 ********* 2025-08-29 17:56:08.006306 | orchestrator | included: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:56:08.006326 | orchestrator | 2025-08-29 17:56:08.006376 | orchestrator | TASK [service-ks-register : nova | Creating services] ************************** 2025-08-29 17:56:08.006395 | orchestrator | Friday 29 August 2025 17:50:21 +0000 (0:00:00.527) 0:02:51.328 ********* 2025-08-29 17:56:08.006413 | orchestrator | skipping: [testbed-node-0] => (item=nova_legacy (compute_legacy))  2025-08-29 17:56:08.006429 | orchestrator | changed: [testbed-node-0] => (item=nova (compute)) 2025-08-29 17:56:08.006441 | orchestrator | 2025-08-29 17:56:08.006452 | orchestrator | TASK [service-ks-register : nova | Creating endpoints] ************************* 2025-08-29 17:56:08.006462 | orchestrator | Friday 29 August 2025 17:50:24 +0000 (0:00:03.338) 0:02:54.666 ********* 2025-08-29 17:56:08.006473 | orchestrator | skipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal)  2025-08-29 17:56:08.006487 | orchestrator | skipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public)  2025-08-29 17:56:08.006498 | orchestrator | changed: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal) 2025-08-29 17:56:08.006509 | orchestrator | changed: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public) 2025-08-29 17:56:08.006520 | orchestrator | 2025-08-29 17:56:08.006539 | orchestrator | TASK [service-ks-register : nova | Creating projects] ************************** 2025-08-29 17:56:08.006550 | orchestrator | Friday 29 August 2025 17:50:30 +0000 (0:00:05.729) 0:03:00.395 ********* 2025-08-29 17:56:08.006561 | orchestrator | ok: [testbed-node-0] => (item=service) 2025-08-29 17:56:08.006572 | orchestrator | 2025-08-29 17:56:08.006582 | orchestrator | TASK [service-ks-register : nova | Creating users] ***************************** 2025-08-29 17:56:08.006593 | orchestrator | Friday 29 August 2025 17:50:33 +0000 (0:00:02.798) 0:03:03.193 ********* 2025-08-29 17:56:08.006604 | orchestrator | [WARNING]: Module did not set no_log for update_password 2025-08-29 17:56:08.006621 | orchestrator | changed: [testbed-node-0] => (item=nova -> service) 2025-08-29 17:56:08.006632 | orchestrator | 2025-08-29 17:56:08.006643 | orchestrator | TASK [service-ks-register : nova | Creating roles] ***************************** 2025-08-29 17:56:08.006659 | orchestrator | Friday 29 August 2025 17:50:36 +0000 (0:00:03.414) 0:03:06.608 ********* 2025-08-29 17:56:08.006678 | orchestrator | ok: [testbed-node-0] => (item=admin) 2025-08-29 17:56:08.006698 | orchestrator | 2025-08-29 17:56:08.006714 | orchestrator | TASK [service-ks-register : nova | Granting user roles] ************************ 2025-08-29 17:56:08.006725 | orchestrator | Friday 29 August 2025 17:50:39 +0000 (0:00:02.923) 0:03:09.531 ********* 2025-08-29 17:56:08.006736 | orchestrator | changed: [testbed-node-0] => (item=nova -> service -> admin) 2025-08-29 17:56:08.006746 | orchestrator | changed: [testbed-node-0] => (item=nova -> service -> service) 2025-08-29 17:56:08.006758 | orchestrator | 2025-08-29 17:56:08.006776 | orchestrator | TASK [nova : Ensuring config directories exist] ******************************** 2025-08-29 17:56:08.006803 | orchestrator | Friday 29 August 2025 17:50:46 +0000 (0:00:06.575) 0:03:16.107 ********* 2025-08-29 17:56:08.006829 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.006852 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.006890 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.006923 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.006944 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.006961 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.006978 | orchestrator | 2025-08-29 17:56:08.006995 | orchestrator | TASK [nova : Check if policies shall be overwritten] *************************** 2025-08-29 17:56:08.007013 | orchestrator | Friday 29 August 2025 17:50:47 +0000 (0:00:01.273) 0:03:17.381 ********* 2025-08-29 17:56:08.007030 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.007049 | orchestrator | 2025-08-29 17:56:08.007068 | orchestrator | TASK [nova : Set nova policy file] ********************************************* 2025-08-29 17:56:08.007087 | orchestrator | Friday 29 August 2025 17:50:47 +0000 (0:00:00.125) 0:03:17.506 ********* 2025-08-29 17:56:08.007101 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.007112 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.007133 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.007144 | orchestrator | 2025-08-29 17:56:08.007155 | orchestrator | TASK [nova : Check for vendordata file] **************************************** 2025-08-29 17:56:08.007165 | orchestrator | Friday 29 August 2025 17:50:48 +0000 (0:00:00.538) 0:03:18.044 ********* 2025-08-29 17:56:08.007176 | orchestrator | ok: [testbed-node-0 -> localhost] 2025-08-29 17:56:08.007187 | orchestrator | 2025-08-29 17:56:08.007198 | orchestrator | TASK [nova : Set vendordata file path] ***************************************** 2025-08-29 17:56:08.007209 | orchestrator | Friday 29 August 2025 17:50:48 +0000 (0:00:00.705) 0:03:18.750 ********* 2025-08-29 17:56:08.007275 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.007287 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.007385 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.007401 | orchestrator | 2025-08-29 17:56:08.007413 | orchestrator | TASK [nova : include_tasks] **************************************************** 2025-08-29 17:56:08.007424 | orchestrator | Friday 29 August 2025 17:50:49 +0000 (0:00:00.320) 0:03:19.070 ********* 2025-08-29 17:56:08.007434 | orchestrator | included: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:56:08.007445 | orchestrator | 2025-08-29 17:56:08.007456 | orchestrator | TASK [service-cert-copy : nova | Copying over extra CA certificates] *********** 2025-08-29 17:56:08.007466 | orchestrator | Friday 29 August 2025 17:50:49 +0000 (0:00:00.520) 0:03:19.591 ********* 2025-08-29 17:56:08.007499 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.007514 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.007527 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.007548 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.007573 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.007595 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.007608 | orchestrator | 2025-08-29 17:56:08.007619 | orchestrator | TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] *** 2025-08-29 17:56:08.007630 | orchestrator | Friday 29 August 2025 17:50:52 +0000 (0:00:02.581) 0:03:22.172 ********* 2025-08-29 17:56:08.007798 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.007821 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.007834 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.007852 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.007873 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.007885 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.007897 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.007916 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.007928 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.007939 | orchestrator | 2025-08-29 17:56:08.007950 | orchestrator | TASK [service-cert-copy : nova | Copying over backend internal TLS key] ******** 2025-08-29 17:56:08.007961 | orchestrator | Friday 29 August 2025 17:50:52 +0000 (0:00:00.577) 0:03:22.749 ********* 2025-08-29 17:56:08.007973 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.007992 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.008004 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.008025 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.008046 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.008058 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.008070 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.008087 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.008099 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.008110 | orchestrator | 2025-08-29 17:56:08.008121 | orchestrator | TASK [nova : Copying over config.json files for services] ********************** 2025-08-29 17:56:08.008132 | orchestrator | Friday 29 August 2025 17:50:53 +0000 (0:00:00.928) 0:03:23.678 ********* 2025-08-29 17:56:08.008153 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008174 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008192 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008212 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008225 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008243 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008255 | orchestrator | 2025-08-29 17:56:08.008266 | orchestrator | TASK [nova : Copying over nova.conf] ******************************************* 2025-08-29 17:56:08.008277 | orchestrator | Friday 29 August 2025 17:50:56 +0000 (0:00:02.731) 0:03:26.409 ********* 2025-08-29 17:56:08.008289 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008306 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008366 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008431 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008446 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008458 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008469 | orchestrator | 2025-08-29 17:56:08.008481 | orchestrator | TASK [nova : Copying over existing policy file] ******************************** 2025-08-29 17:56:08.008533 | orchestrator | Friday 29 August 2025 17:51:02 +0000 (0:00:05.714) 0:03:32.124 ********* 2025-08-29 17:56:08.008560 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.008582 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.008594 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.008606 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.008618 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.008629 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.008675 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})  2025-08-29 17:56:08.008713 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.008725 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.008737 | orchestrator | 2025-08-29 17:56:08.008748 | orchestrator | TASK [nova : Copying over nova-api-wsgi.conf] ********************************** 2025-08-29 17:56:08.008759 | orchestrator | Friday 29 August 2025 17:51:02 +0000 (0:00:00.587) 0:03:32.712 ********* 2025-08-29 17:56:08.008771 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.008782 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.008793 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.008804 | orchestrator | 2025-08-29 17:56:08.008815 | orchestrator | TASK [nova : Copying over vendordata file] ************************************* 2025-08-29 17:56:08.008826 | orchestrator | Friday 29 August 2025 17:51:04 +0000 (0:00:01.748) 0:03:34.460 ********* 2025-08-29 17:56:08.008837 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.008848 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.008859 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.008870 | orchestrator | 2025-08-29 17:56:08.008881 | orchestrator | TASK [nova : Check nova containers] ******************************************** 2025-08-29 17:56:08.008892 | orchestrator | Friday 29 August 2025 17:51:05 +0000 (0:00:00.532) 0:03:34.992 ********* 2025-08-29 17:56:08.008904 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008922 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008953 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/release/nova-api:30.0.1.20250711', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) 2025-08-29 17:56:08.008967 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008980 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.008992 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.009003 | orchestrator | 2025-08-29 17:56:08.009014 | orchestrator | TASK [nova : Flush handlers] *************************************************** 2025-08-29 17:56:08.009026 | orchestrator | Friday 29 August 2025 17:51:06 +0000 (0:00:01.881) 0:03:36.874 ********* 2025-08-29 17:56:08.009037 | orchestrator | 2025-08-29 17:56:08.009048 | orchestrator | TASK [nova : Flush handlers] *************************************************** 2025-08-29 17:56:08.009058 | orchestrator | Friday 29 August 2025 17:51:07 +0000 (0:00:00.167) 0:03:37.041 ********* 2025-08-29 17:56:08.009069 | orchestrator | 2025-08-29 17:56:08.009080 | orchestrator | TASK [nova : Flush handlers] *************************************************** 2025-08-29 17:56:08.009097 | orchestrator | Friday 29 August 2025 17:51:07 +0000 (0:00:00.130) 0:03:37.172 ********* 2025-08-29 17:56:08.009108 | orchestrator | 2025-08-29 17:56:08.009119 | orchestrator | RUNNING HANDLER [nova : Restart nova-scheduler container] ********************** 2025-08-29 17:56:08.009130 | orchestrator | Friday 29 August 2025 17:51:07 +0000 (0:00:00.133) 0:03:37.305 ********* 2025-08-29 17:56:08.009140 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.009151 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.009162 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.009173 | orchestrator | 2025-08-29 17:56:08.009184 | orchestrator | RUNNING HANDLER [nova : Restart nova-api container] **************************** 2025-08-29 17:56:08.009200 | orchestrator | Friday 29 August 2025 17:51:31 +0000 (0:00:23.754) 0:04:01.060 ********* 2025-08-29 17:56:08.009211 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.009222 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.009233 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.009244 | orchestrator | 2025-08-29 17:56:08.009255 | orchestrator | PLAY [Apply role nova-cell] **************************************************** 2025-08-29 17:56:08.009266 | orchestrator | 2025-08-29 17:56:08.009277 | orchestrator | TASK [nova-cell : include_tasks] *********************************************** 2025-08-29 17:56:08.009288 | orchestrator | Friday 29 August 2025 17:51:39 +0000 (0:00:08.416) 0:04:09.476 ********* 2025-08-29 17:56:08.009299 | orchestrator | included: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:56:08.009312 | orchestrator | 2025-08-29 17:56:08.009349 | orchestrator | TASK [nova-cell : include_tasks] *********************************************** 2025-08-29 17:56:08.009363 | orchestrator | Friday 29 August 2025 17:51:40 +0000 (0:00:01.221) 0:04:10.698 ********* 2025-08-29 17:56:08.009374 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.009385 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.009396 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.009407 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.009418 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.009429 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.009440 | orchestrator | 2025-08-29 17:56:08.009451 | orchestrator | TASK [Load and persist br_netfilter module] ************************************ 2025-08-29 17:56:08.009462 | orchestrator | Friday 29 August 2025 17:51:41 +0000 (0:00:00.738) 0:04:11.436 ********* 2025-08-29 17:56:08.009473 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.009484 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.009495 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.009506 | orchestrator | included: module-load for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:56:08.009517 | orchestrator | 2025-08-29 17:56:08.009528 | orchestrator | TASK [module-load : Load modules] ********************************************** 2025-08-29 17:56:08.009539 | orchestrator | Friday 29 August 2025 17:51:42 +0000 (0:00:00.862) 0:04:12.299 ********* 2025-08-29 17:56:08.009551 | orchestrator | ok: [testbed-node-3] => (item=br_netfilter) 2025-08-29 17:56:08.009562 | orchestrator | ok: [testbed-node-4] => (item=br_netfilter) 2025-08-29 17:56:08.009573 | orchestrator | ok: [testbed-node-5] => (item=br_netfilter) 2025-08-29 17:56:08.009584 | orchestrator | 2025-08-29 17:56:08.009595 | orchestrator | TASK [module-load : Persist modules via modules-load.d] ************************ 2025-08-29 17:56:08.009606 | orchestrator | Friday 29 August 2025 17:51:43 +0000 (0:00:00.887) 0:04:13.186 ********* 2025-08-29 17:56:08.009617 | orchestrator | changed: [testbed-node-3] => (item=br_netfilter) 2025-08-29 17:56:08.009628 | orchestrator | changed: [testbed-node-4] => (item=br_netfilter) 2025-08-29 17:56:08.009639 | orchestrator | changed: [testbed-node-5] => (item=br_netfilter) 2025-08-29 17:56:08.009650 | orchestrator | 2025-08-29 17:56:08.009661 | orchestrator | TASK [module-load : Drop module persistence] *********************************** 2025-08-29 17:56:08.009672 | orchestrator | Friday 29 August 2025 17:51:44 +0000 (0:00:01.202) 0:04:14.388 ********* 2025-08-29 17:56:08.009691 | orchestrator | skipping: [testbed-node-3] => (item=br_netfilter)  2025-08-29 17:56:08.009702 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.009713 | orchestrator | skipping: [testbed-node-4] => (item=br_netfilter)  2025-08-29 17:56:08.009724 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.009735 | orchestrator | skipping: [testbed-node-5] => (item=br_netfilter)  2025-08-29 17:56:08.009746 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.009756 | orchestrator | 2025-08-29 17:56:08.009768 | orchestrator | TASK [nova-cell : Enable bridge-nf-call sysctl variables] ********************** 2025-08-29 17:56:08.009779 | orchestrator | Friday 29 August 2025 17:51:44 +0000 (0:00:00.546) 0:04:14.935 ********* 2025-08-29 17:56:08.009791 | orchestrator | changed: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) 2025-08-29 17:56:08.009801 | orchestrator | changed: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) 2025-08-29 17:56:08.009812 | orchestrator | skipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:56:08.009824 | orchestrator | skipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:56:08.009835 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.009846 | orchestrator | skipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:56:08.009857 | orchestrator | skipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:56:08.009868 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.009879 | orchestrator | changed: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) 2025-08-29 17:56:08.009890 | orchestrator | changed: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) 2025-08-29 17:56:08.009901 | orchestrator | changed: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) 2025-08-29 17:56:08.009912 | orchestrator | skipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables)  2025-08-29 17:56:08.009923 | orchestrator | skipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables)  2025-08-29 17:56:08.009934 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.009945 | orchestrator | changed: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) 2025-08-29 17:56:08.009956 | orchestrator | 2025-08-29 17:56:08.009967 | orchestrator | TASK [nova-cell : Install udev kolla kvm rules] ******************************** 2025-08-29 17:56:08.009978 | orchestrator | Friday 29 August 2025 17:51:46 +0000 (0:00:01.334) 0:04:16.270 ********* 2025-08-29 17:56:08.009989 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.010000 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.010011 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.010082 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.010093 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.010110 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.010122 | orchestrator | 2025-08-29 17:56:08.010133 | orchestrator | TASK [nova-cell : Mask qemu-kvm service] *************************************** 2025-08-29 17:56:08.010144 | orchestrator | Friday 29 August 2025 17:51:47 +0000 (0:00:01.076) 0:04:17.346 ********* 2025-08-29 17:56:08.010155 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.010166 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.010177 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.010188 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.010198 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.010209 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.010220 | orchestrator | 2025-08-29 17:56:08.010231 | orchestrator | TASK [nova-cell : Ensuring config directories exist] *************************** 2025-08-29 17:56:08.010242 | orchestrator | Friday 29 August 2025 17:51:49 +0000 (0:00:01.952) 0:04:19.298 ********* 2025-08-29 17:56:08.011093 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011223 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011239 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011252 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011280 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011313 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011377 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011391 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011402 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011415 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011432 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011455 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011480 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011493 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011505 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011517 | orchestrator | 2025-08-29 17:56:08.011530 | orchestrator | TASK [nova-cell : include_tasks] *********************************************** 2025-08-29 17:56:08.011543 | orchestrator | Friday 29 August 2025 17:51:51 +0000 (0:00:02.360) 0:04:21.658 ********* 2025-08-29 17:56:08.011555 | orchestrator | included: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 17:56:08.011568 | orchestrator | 2025-08-29 17:56:08.011580 | orchestrator | TASK [service-cert-copy : nova | Copying over extra CA certificates] *********** 2025-08-29 17:56:08.011591 | orchestrator | Friday 29 August 2025 17:51:52 +0000 (0:00:01.265) 0:04:22.924 ********* 2025-08-29 17:56:08.011610 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011641 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011657 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011671 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011685 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011699 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011719 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011748 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011763 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011776 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011790 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011803 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011817 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011852 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011866 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.011880 | orchestrator | 2025-08-29 17:56:08.011893 | orchestrator | TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] *** 2025-08-29 17:56:08.011909 | orchestrator | Friday 29 August 2025 17:51:56 +0000 (0:00:03.941) 0:04:26.866 ********* 2025-08-29 17:56:08.011931 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.011953 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.011973 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012005 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.012049 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.012069 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.012082 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012094 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.012107 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.012128 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.012151 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012166 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.012187 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.012200 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012213 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.012225 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.012238 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012250 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.012262 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.012288 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012302 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.012314 | orchestrator | 2025-08-29 17:56:08.012326 | orchestrator | TASK [service-cert-copy : nova | Copying over backend internal TLS key] ******** 2025-08-29 17:56:08.012361 | orchestrator | Friday 29 August 2025 17:51:58 +0000 (0:00:01.440) 0:04:28.306 ********* 2025-08-29 17:56:08.012381 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.012395 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.012408 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012420 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.012432 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.012458 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.012478 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012490 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.012502 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.012514 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.012526 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012547 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.012559 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.012576 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012588 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.012607 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.012620 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012632 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.012644 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.012657 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.012675 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.012687 | orchestrator | 2025-08-29 17:56:08.012699 | orchestrator | TASK [nova-cell : include_tasks] *********************************************** 2025-08-29 17:56:08.012711 | orchestrator | Friday 29 August 2025 17:52:00 +0000 (0:00:02.083) 0:04:30.390 ********* 2025-08-29 17:56:08.012722 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.012734 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.012745 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.012756 | orchestrator | included: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5 2025-08-29 17:56:08.012768 | orchestrator | 2025-08-29 17:56:08.012779 | orchestrator | TASK [nova-cell : Check nova keyring file] ************************************* 2025-08-29 17:56:08.012791 | orchestrator | Friday 29 August 2025 17:52:01 +0000 (0:00:01.070) 0:04:31.460 ********* 2025-08-29 17:56:08.012803 | orchestrator | ok: [testbed-node-3 -> localhost] 2025-08-29 17:56:08.012815 | orchestrator | ok: [testbed-node-4 -> localhost] 2025-08-29 17:56:08.012827 | orchestrator | ok: [testbed-node-5 -> localhost] 2025-08-29 17:56:08.012838 | orchestrator | 2025-08-29 17:56:08.012849 | orchestrator | TASK [nova-cell : Check cinder keyring file] *********************************** 2025-08-29 17:56:08.012861 | orchestrator | Friday 29 August 2025 17:52:02 +0000 (0:00:00.928) 0:04:32.389 ********* 2025-08-29 17:56:08.012873 | orchestrator | ok: [testbed-node-3 -> localhost] 2025-08-29 17:56:08.012883 | orchestrator | ok: [testbed-node-5 -> localhost] 2025-08-29 17:56:08.012895 | orchestrator | ok: [testbed-node-4 -> localhost] 2025-08-29 17:56:08.012906 | orchestrator | 2025-08-29 17:56:08.012917 | orchestrator | TASK [nova-cell : Extract nova key from file] ********************************** 2025-08-29 17:56:08.012933 | orchestrator | Friday 29 August 2025 17:52:03 +0000 (0:00:01.243) 0:04:33.632 ********* 2025-08-29 17:56:08.012945 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:56:08.012957 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:56:08.012968 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:56:08.012979 | orchestrator | 2025-08-29 17:56:08.012991 | orchestrator | TASK [nova-cell : Extract cinder key from file] ******************************** 2025-08-29 17:56:08.013002 | orchestrator | Friday 29 August 2025 17:52:04 +0000 (0:00:00.608) 0:04:34.241 ********* 2025-08-29 17:56:08.013013 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:56:08.013024 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:56:08.013036 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:56:08.013047 | orchestrator | 2025-08-29 17:56:08.013058 | orchestrator | TASK [nova-cell : Copy over ceph nova keyring file] **************************** 2025-08-29 17:56:08.013069 | orchestrator | Friday 29 August 2025 17:52:04 +0000 (0:00:00.562) 0:04:34.803 ********* 2025-08-29 17:56:08.013081 | orchestrator | changed: [testbed-node-3] => (item=nova-compute) 2025-08-29 17:56:08.013100 | orchestrator | changed: [testbed-node-4] => (item=nova-compute) 2025-08-29 17:56:08.013112 | orchestrator | changed: [testbed-node-5] => (item=nova-compute) 2025-08-29 17:56:08.013123 | orchestrator | 2025-08-29 17:56:08.013134 | orchestrator | TASK [nova-cell : Copy over ceph cinder keyring file] ************************** 2025-08-29 17:56:08.013146 | orchestrator | Friday 29 August 2025 17:52:06 +0000 (0:00:01.260) 0:04:36.063 ********* 2025-08-29 17:56:08.013157 | orchestrator | changed: [testbed-node-3] => (item=nova-compute) 2025-08-29 17:56:08.013168 | orchestrator | changed: [testbed-node-4] => (item=nova-compute) 2025-08-29 17:56:08.013179 | orchestrator | changed: [testbed-node-5] => (item=nova-compute) 2025-08-29 17:56:08.013191 | orchestrator | 2025-08-29 17:56:08.013202 | orchestrator | TASK [nova-cell : Copy over ceph.conf] ***************************************** 2025-08-29 17:56:08.013222 | orchestrator | Friday 29 August 2025 17:52:07 +0000 (0:00:01.356) 0:04:37.420 ********* 2025-08-29 17:56:08.013234 | orchestrator | changed: [testbed-node-4] => (item=nova-compute) 2025-08-29 17:56:08.013245 | orchestrator | changed: [testbed-node-5] => (item=nova-compute) 2025-08-29 17:56:08.013256 | orchestrator | changed: [testbed-node-3] => (item=nova-compute) 2025-08-29 17:56:08.013268 | orchestrator | changed: [testbed-node-5] => (item=nova-libvirt) 2025-08-29 17:56:08.013279 | orchestrator | changed: [testbed-node-3] => (item=nova-libvirt) 2025-08-29 17:56:08.013290 | orchestrator | changed: [testbed-node-4] => (item=nova-libvirt) 2025-08-29 17:56:08.013301 | orchestrator | 2025-08-29 17:56:08.013313 | orchestrator | TASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************ 2025-08-29 17:56:08.013324 | orchestrator | Friday 29 August 2025 17:52:11 +0000 (0:00:04.248) 0:04:41.669 ********* 2025-08-29 17:56:08.013365 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.013377 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.013389 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.013401 | orchestrator | 2025-08-29 17:56:08.013412 | orchestrator | TASK [nova-cell : Copy over ceph.conf (host libvirt)] ************************** 2025-08-29 17:56:08.013423 | orchestrator | Friday 29 August 2025 17:52:12 +0000 (0:00:00.318) 0:04:41.987 ********* 2025-08-29 17:56:08.013434 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.013446 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.013457 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.013468 | orchestrator | 2025-08-29 17:56:08.013480 | orchestrator | TASK [nova-cell : Ensuring libvirt secrets directory exists] ******************* 2025-08-29 17:56:08.013491 | orchestrator | Friday 29 August 2025 17:52:12 +0000 (0:00:00.300) 0:04:42.288 ********* 2025-08-29 17:56:08.013502 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.013513 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.013524 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.013535 | orchestrator | 2025-08-29 17:56:08.013546 | orchestrator | TASK [nova-cell : Pushing nova secret xml for libvirt] ************************* 2025-08-29 17:56:08.013558 | orchestrator | Friday 29 August 2025 17:52:13 +0000 (0:00:01.641) 0:04:43.929 ********* 2025-08-29 17:56:08.013570 | orchestrator | changed: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True}) 2025-08-29 17:56:08.013581 | orchestrator | changed: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True}) 2025-08-29 17:56:08.013593 | orchestrator | changed: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True}) 2025-08-29 17:56:08.013604 | orchestrator | changed: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'}) 2025-08-29 17:56:08.013615 | orchestrator | changed: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'}) 2025-08-29 17:56:08.013626 | orchestrator | changed: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'}) 2025-08-29 17:56:08.013637 | orchestrator | 2025-08-29 17:56:08.013649 | orchestrator | TASK [nova-cell : Pushing secrets key for libvirt] ***************************** 2025-08-29 17:56:08.013660 | orchestrator | Friday 29 August 2025 17:52:17 +0000 (0:00:03.385) 0:04:47.314 ********* 2025-08-29 17:56:08.013672 | orchestrator | changed: [testbed-node-3] => (item=None) 2025-08-29 17:56:08.013684 | orchestrator | changed: [testbed-node-4] => (item=None) 2025-08-29 17:56:08.013695 | orchestrator | changed: [testbed-node-5] => (item=None) 2025-08-29 17:56:08.013706 | orchestrator | changed: [testbed-node-3] => (item=None) 2025-08-29 17:56:08.013717 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.013728 | orchestrator | changed: [testbed-node-4] => (item=None) 2025-08-29 17:56:08.013739 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.013760 | orchestrator | changed: [testbed-node-5] => (item=None) 2025-08-29 17:56:08.013772 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.013784 | orchestrator | 2025-08-29 17:56:08.013801 | orchestrator | TASK [nova-cell : Check if policies shall be overwritten] ********************** 2025-08-29 17:56:08.013813 | orchestrator | Friday 29 August 2025 17:52:20 +0000 (0:00:03.038) 0:04:50.353 ********* 2025-08-29 17:56:08.013825 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.013837 | orchestrator | 2025-08-29 17:56:08.013849 | orchestrator | TASK [nova-cell : Set nova policy file] **************************************** 2025-08-29 17:56:08.013860 | orchestrator | Friday 29 August 2025 17:52:20 +0000 (0:00:00.125) 0:04:50.479 ********* 2025-08-29 17:56:08.013872 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.013883 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.013894 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.013906 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.013917 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.013928 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.013939 | orchestrator | 2025-08-29 17:56:08.013951 | orchestrator | TASK [nova-cell : Check for vendordata file] *********************************** 2025-08-29 17:56:08.013971 | orchestrator | Friday 29 August 2025 17:52:21 +0000 (0:00:00.629) 0:04:51.109 ********* 2025-08-29 17:56:08.013982 | orchestrator | ok: [testbed-node-3 -> localhost] 2025-08-29 17:56:08.013993 | orchestrator | 2025-08-29 17:56:08.014004 | orchestrator | TASK [nova-cell : Set vendordata file path] ************************************ 2025-08-29 17:56:08.014163 | orchestrator | Friday 29 August 2025 17:52:21 +0000 (0:00:00.594) 0:04:51.703 ********* 2025-08-29 17:56:08.014182 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.014195 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.014206 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.014217 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.014227 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.014238 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.014250 | orchestrator | 2025-08-29 17:56:08.014261 | orchestrator | TASK [nova-cell : Copying over config.json files for services] ***************** 2025-08-29 17:56:08.014273 | orchestrator | Friday 29 August 2025 17:52:22 +0000 (0:00:00.498) 0:04:52.202 ********* 2025-08-29 17:56:08.014285 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014298 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014322 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014364 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014387 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014400 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014412 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014424 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014444 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014457 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014478 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014498 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014510 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014524 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014545 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014558 | orchestrator | 2025-08-29 17:56:08.014570 | orchestrator | TASK [nova-cell : Copying over nova.conf] ************************************** 2025-08-29 17:56:08.014581 | orchestrator | Friday 29 August 2025 17:52:26 +0000 (0:00:04.133) 0:04:56.335 ********* 2025-08-29 17:56:08.014598 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.014617 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.014629 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.014640 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.014660 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.014676 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.014695 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014708 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014720 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014738 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014750 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014767 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014787 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014799 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014811 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.014822 | orchestrator | 2025-08-29 17:56:08.014833 | orchestrator | TASK [nova-cell : Copying over Nova compute provider config] ******************* 2025-08-29 17:56:08.014851 | orchestrator | Friday 29 August 2025 17:52:32 +0000 (0:00:06.375) 0:05:02.711 ********* 2025-08-29 17:56:08.014863 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.014874 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.014885 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.014896 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.014906 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.014917 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.014928 | orchestrator | 2025-08-29 17:56:08.014939 | orchestrator | TASK [nova-cell : Copying over libvirt configuration] ************************** 2025-08-29 17:56:08.014950 | orchestrator | Friday 29 August 2025 17:52:34 +0000 (0:00:01.770) 0:05:04.482 ********* 2025-08-29 17:56:08.014962 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})  2025-08-29 17:56:08.014973 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})  2025-08-29 17:56:08.014984 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})  2025-08-29 17:56:08.014995 | orchestrator | changed: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) 2025-08-29 17:56:08.015006 | orchestrator | changed: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) 2025-08-29 17:56:08.015017 | orchestrator | changed: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) 2025-08-29 17:56:08.015029 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})  2025-08-29 17:56:08.015040 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.015051 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})  2025-08-29 17:56:08.015062 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.015073 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})  2025-08-29 17:56:08.015084 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.015096 | orchestrator | changed: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) 2025-08-29 17:56:08.015107 | orchestrator | changed: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) 2025-08-29 17:56:08.015118 | orchestrator | changed: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) 2025-08-29 17:56:08.015129 | orchestrator | 2025-08-29 17:56:08.015140 | orchestrator | TASK [nova-cell : Copying over libvirt TLS keys] ******************************* 2025-08-29 17:56:08.015151 | orchestrator | Friday 29 August 2025 17:52:38 +0000 (0:00:03.746) 0:05:08.229 ********* 2025-08-29 17:56:08.015161 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.015172 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.015189 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.015200 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.015211 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.015222 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.015232 | orchestrator | 2025-08-29 17:56:08.015243 | orchestrator | TASK [nova-cell : Copying over libvirt SASL configuration] ********************* 2025-08-29 17:56:08.015254 | orchestrator | Friday 29 August 2025 17:52:39 +0000 (0:00:00.824) 0:05:09.053 ********* 2025-08-29 17:56:08.015266 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})  2025-08-29 17:56:08.015277 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})  2025-08-29 17:56:08.015295 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})  2025-08-29 17:56:08.015307 | orchestrator | changed: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) 2025-08-29 17:56:08.015318 | orchestrator | changed: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) 2025-08-29 17:56:08.015357 | orchestrator | changed: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) 2025-08-29 17:56:08.015369 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})  2025-08-29 17:56:08.015380 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})  2025-08-29 17:56:08.015391 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})  2025-08-29 17:56:08.015402 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})  2025-08-29 17:56:08.015413 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.015424 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})  2025-08-29 17:56:08.015435 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.015446 | orchestrator | changed: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) 2025-08-29 17:56:08.015457 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})  2025-08-29 17:56:08.015468 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.015479 | orchestrator | changed: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) 2025-08-29 17:56:08.015490 | orchestrator | changed: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) 2025-08-29 17:56:08.015500 | orchestrator | changed: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) 2025-08-29 17:56:08.015511 | orchestrator | changed: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) 2025-08-29 17:56:08.015522 | orchestrator | changed: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) 2025-08-29 17:56:08.015533 | orchestrator | 2025-08-29 17:56:08.015544 | orchestrator | TASK [nova-cell : Copying files for nova-ssh] ********************************** 2025-08-29 17:56:08.015556 | orchestrator | Friday 29 August 2025 17:52:44 +0000 (0:00:05.609) 0:05:14.663 ********* 2025-08-29 17:56:08.015567 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})  2025-08-29 17:56:08.015578 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})  2025-08-29 17:56:08.015589 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})  2025-08-29 17:56:08.015601 | orchestrator | changed: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) 2025-08-29 17:56:08.015612 | orchestrator | changed: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) 2025-08-29 17:56:08.015623 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})  2025-08-29 17:56:08.015634 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})  2025-08-29 17:56:08.015646 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})  2025-08-29 17:56:08.015656 | orchestrator | changed: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) 2025-08-29 17:56:08.015667 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})  2025-08-29 17:56:08.015678 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})  2025-08-29 17:56:08.015689 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})  2025-08-29 17:56:08.015700 | orchestrator | skipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})  2025-08-29 17:56:08.015711 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.015728 | orchestrator | skipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})  2025-08-29 17:56:08.015745 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.015756 | orchestrator | skipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})  2025-08-29 17:56:08.015767 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.015778 | orchestrator | changed: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) 2025-08-29 17:56:08.015789 | orchestrator | changed: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) 2025-08-29 17:56:08.015800 | orchestrator | changed: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) 2025-08-29 17:56:08.015811 | orchestrator | changed: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) 2025-08-29 17:56:08.015822 | orchestrator | changed: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) 2025-08-29 17:56:08.015840 | orchestrator | changed: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) 2025-08-29 17:56:08.015852 | orchestrator | changed: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) 2025-08-29 17:56:08.015863 | orchestrator | changed: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) 2025-08-29 17:56:08.015874 | orchestrator | changed: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) 2025-08-29 17:56:08.015885 | orchestrator | 2025-08-29 17:56:08.015896 | orchestrator | TASK [nova-cell : Copying VMware vCenter CA file] ****************************** 2025-08-29 17:56:08.015907 | orchestrator | Friday 29 August 2025 17:52:51 +0000 (0:00:07.029) 0:05:21.692 ********* 2025-08-29 17:56:08.015918 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.015929 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.015939 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.015950 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.015961 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.015972 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.015983 | orchestrator | 2025-08-29 17:56:08.015994 | orchestrator | TASK [nova-cell : Copying 'release' file for nova_compute] ********************* 2025-08-29 17:56:08.016005 | orchestrator | Friday 29 August 2025 17:52:52 +0000 (0:00:00.517) 0:05:22.210 ********* 2025-08-29 17:56:08.016016 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.016026 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.016037 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.016048 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.016058 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.016069 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.016080 | orchestrator | 2025-08-29 17:56:08.016091 | orchestrator | TASK [nova-cell : Generating 'hostnqn' file for nova_compute] ****************** 2025-08-29 17:56:08.016102 | orchestrator | Friday 29 August 2025 17:52:52 +0000 (0:00:00.674) 0:05:22.885 ********* 2025-08-29 17:56:08.016112 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.016124 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.016134 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.016145 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.016156 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.016167 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.016178 | orchestrator | 2025-08-29 17:56:08.016189 | orchestrator | TASK [nova-cell : Copying over existing policy file] *************************** 2025-08-29 17:56:08.016200 | orchestrator | Friday 29 August 2025 17:52:54 +0000 (0:00:01.735) 0:05:24.621 ********* 2025-08-29 17:56:08.016212 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.016231 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.016248 | orchestrator | skipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.016261 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.016281 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.016293 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.016304 | orchestrator | skipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.016322 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.016351 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})  2025-08-29 17:56:08.016371 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})  2025-08-29 17:56:08.016392 | orchestrator | skipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.016404 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.016416 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.016427 | orchestrator | skipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.016445 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.016457 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.016468 | orchestrator | skipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.016479 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.016495 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})  2025-08-29 17:56:08.016513 | orchestrator | skipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})  2025-08-29 17:56:08.016525 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.016536 | orchestrator | 2025-08-29 17:56:08.016547 | orchestrator | TASK [nova-cell : Copying over vendordata file to containers] ****************** 2025-08-29 17:56:08.016558 | orchestrator | Friday 29 August 2025 17:52:56 +0000 (0:00:01.365) 0:05:25.986 ********* 2025-08-29 17:56:08.016569 | orchestrator | skipping: [testbed-node-3] => (item=nova-compute)  2025-08-29 17:56:08.016580 | orchestrator | skipping: [testbed-node-3] => (item=nova-compute-ironic)  2025-08-29 17:56:08.016591 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.016602 | orchestrator | skipping: [testbed-node-4] => (item=nova-compute)  2025-08-29 17:56:08.016613 | orchestrator | skipping: [testbed-node-4] => (item=nova-compute-ironic)  2025-08-29 17:56:08.016624 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.016634 | orchestrator | skipping: [testbed-node-5] => (item=nova-compute)  2025-08-29 17:56:08.016645 | orchestrator | skipping: [testbed-node-5] => (item=nova-compute-ironic)  2025-08-29 17:56:08.016656 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.016667 | orchestrator | skipping: [testbed-node-0] => (item=nova-compute)  2025-08-29 17:56:08.016678 | orchestrator | skipping: [testbed-node-0] => (item=nova-compute-ironic)  2025-08-29 17:56:08.016689 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.016700 | orchestrator | skipping: [testbed-node-1] => (item=nova-compute)  2025-08-29 17:56:08.016718 | orchestrator | skipping: [testbed-node-1] => (item=nova-compute-ironic)  2025-08-29 17:56:08.016729 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.016739 | orchestrator | skipping: [testbed-node-2] => (item=nova-compute)  2025-08-29 17:56:08.016750 | orchestrator | skipping: [testbed-node-2] => (item=nova-compute-ironic)  2025-08-29 17:56:08.016761 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.016771 | orchestrator | 2025-08-29 17:56:08.016782 | orchestrator | TASK [nova-cell : Check nova-cell containers] ********************************** 2025-08-29 17:56:08.016794 | orchestrator | Friday 29 August 2025 17:52:56 +0000 (0:00:00.574) 0:05:26.561 ********* 2025-08-29 17:56:08.016805 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016817 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016840 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016852 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016864 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016882 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016893 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016905 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016921 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016939 | orchestrator | changed: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016951 | orchestrator | changed: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016969 | orchestrator | changed: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016981 | orchestrator | changed: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.016993 | orchestrator | changed: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.017009 | orchestrator | changed: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', 'iscsi_info:/etc/iscsi', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) 2025-08-29 17:56:08.017021 | orchestrator | 2025-08-29 17:56:08.017032 | orchestrator | TASK [nova-cell : include_tasks] *********************************************** 2025-08-29 17:56:08.017043 | orchestrator | Friday 29 August 2025 17:52:59 +0000 (0:00:02.950) 0:05:29.511 ********* 2025-08-29 17:56:08.017054 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.017065 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.017076 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.017092 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.017104 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.017114 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.017133 | orchestrator | 2025-08-29 17:56:08.017144 | orchestrator | TASK [nova-cell : Flush handlers] ********************************************** 2025-08-29 17:56:08.017155 | orchestrator | Friday 29 August 2025 17:53:00 +0000 (0:00:00.547) 0:05:30.059 ********* 2025-08-29 17:56:08.017166 | orchestrator | 2025-08-29 17:56:08.017177 | orchestrator | TASK [nova-cell : Flush handlers] ********************************************** 2025-08-29 17:56:08.017188 | orchestrator | Friday 29 August 2025 17:53:00 +0000 (0:00:00.155) 0:05:30.215 ********* 2025-08-29 17:56:08.017199 | orchestrator | 2025-08-29 17:56:08.017210 | orchestrator | TASK [nova-cell : Flush handlers] ********************************************** 2025-08-29 17:56:08.017220 | orchestrator | Friday 29 August 2025 17:53:00 +0000 (0:00:00.313) 0:05:30.528 ********* 2025-08-29 17:56:08.017231 | orchestrator | 2025-08-29 17:56:08.017243 | orchestrator | TASK [nova-cell : Flush handlers] ********************************************** 2025-08-29 17:56:08.017254 | orchestrator | Friday 29 August 2025 17:53:00 +0000 (0:00:00.135) 0:05:30.664 ********* 2025-08-29 17:56:08.017265 | orchestrator | 2025-08-29 17:56:08.017275 | orchestrator | TASK [nova-cell : Flush handlers] ********************************************** 2025-08-29 17:56:08.017286 | orchestrator | Friday 29 August 2025 17:53:00 +0000 (0:00:00.132) 0:05:30.796 ********* 2025-08-29 17:56:08.017297 | orchestrator | 2025-08-29 17:56:08.017308 | orchestrator | TASK [nova-cell : Flush handlers] ********************************************** 2025-08-29 17:56:08.017319 | orchestrator | Friday 29 August 2025 17:53:00 +0000 (0:00:00.126) 0:05:30.923 ********* 2025-08-29 17:56:08.017386 | orchestrator | 2025-08-29 17:56:08.017400 | orchestrator | RUNNING HANDLER [nova-cell : Restart nova-conductor container] ***************** 2025-08-29 17:56:08.017411 | orchestrator | Friday 29 August 2025 17:53:01 +0000 (0:00:00.125) 0:05:31.048 ********* 2025-08-29 17:56:08.017425 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.017442 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.017460 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.017489 | orchestrator | 2025-08-29 17:56:08.017510 | orchestrator | RUNNING HANDLER [nova-cell : Restart nova-novncproxy container] **************** 2025-08-29 17:56:08.017528 | orchestrator | Friday 29 August 2025 17:53:08 +0000 (0:00:07.470) 0:05:38.519 ********* 2025-08-29 17:56:08.017547 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.017565 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.017584 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.017603 | orchestrator | 2025-08-29 17:56:08.017622 | orchestrator | RUNNING HANDLER [nova-cell : Restart nova-ssh container] *********************** 2025-08-29 17:56:08.017639 | orchestrator | Friday 29 August 2025 17:53:22 +0000 (0:00:13.886) 0:05:52.405 ********* 2025-08-29 17:56:08.017656 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.017667 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.017678 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.017689 | orchestrator | 2025-08-29 17:56:08.017698 | orchestrator | RUNNING HANDLER [nova-cell : Restart nova-libvirt container] ******************* 2025-08-29 17:56:08.017708 | orchestrator | Friday 29 August 2025 17:53:48 +0000 (0:00:26.516) 0:06:18.922 ********* 2025-08-29 17:56:08.017718 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.017727 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.017737 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.017746 | orchestrator | 2025-08-29 17:56:08.017756 | orchestrator | RUNNING HANDLER [nova-cell : Checking libvirt container is ready] ************** 2025-08-29 17:56:08.017766 | orchestrator | Friday 29 August 2025 17:54:34 +0000 (0:00:45.821) 0:07:04.743 ********* 2025-08-29 17:56:08.017775 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.017785 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.017794 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.017804 | orchestrator | 2025-08-29 17:56:08.017814 | orchestrator | RUNNING HANDLER [nova-cell : Create libvirt SASL user] ************************* 2025-08-29 17:56:08.017824 | orchestrator | Friday 29 August 2025 17:54:35 +0000 (0:00:00.735) 0:07:05.478 ********* 2025-08-29 17:56:08.017833 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.017843 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.017863 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.017872 | orchestrator | 2025-08-29 17:56:08.017882 | orchestrator | RUNNING HANDLER [nova-cell : Restart nova-compute container] ******************* 2025-08-29 17:56:08.017891 | orchestrator | Friday 29 August 2025 17:54:36 +0000 (0:00:01.005) 0:07:06.484 ********* 2025-08-29 17:56:08.017901 | orchestrator | changed: [testbed-node-5] 2025-08-29 17:56:08.017911 | orchestrator | changed: [testbed-node-3] 2025-08-29 17:56:08.017920 | orchestrator | changed: [testbed-node-4] 2025-08-29 17:56:08.017930 | orchestrator | 2025-08-29 17:56:08.017939 | orchestrator | RUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] *** 2025-08-29 17:56:08.017949 | orchestrator | Friday 29 August 2025 17:54:58 +0000 (0:00:22.271) 0:07:28.756 ********* 2025-08-29 17:56:08.017958 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.017968 | orchestrator | 2025-08-29 17:56:08.017977 | orchestrator | TASK [nova-cell : Waiting for nova-compute services to register themselves] **** 2025-08-29 17:56:08.017987 | orchestrator | Friday 29 August 2025 17:54:58 +0000 (0:00:00.135) 0:07:28.891 ********* 2025-08-29 17:56:08.017997 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.018007 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.018058 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.018069 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.018078 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.018088 | orchestrator | FAILED - RETRYING: [testbed-node-4 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left). 2025-08-29 17:56:08.018099 | orchestrator | ok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:56:08.018109 | orchestrator | 2025-08-29 17:56:08.018118 | orchestrator | TASK [nova-cell : Fail if nova-compute service failed to register] ************* 2025-08-29 17:56:08.018128 | orchestrator | Friday 29 August 2025 17:55:21 +0000 (0:00:22.321) 0:07:51.213 ********* 2025-08-29 17:56:08.018138 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.018147 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.018157 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.018167 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.018187 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.018197 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.018207 | orchestrator | 2025-08-29 17:56:08.018216 | orchestrator | TASK [nova-cell : Include discover_computes.yml] ******************************* 2025-08-29 17:56:08.018226 | orchestrator | Friday 29 August 2025 17:55:29 +0000 (0:00:08.243) 0:07:59.456 ********* 2025-08-29 17:56:08.018236 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.018246 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.018255 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.018265 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.018274 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.018284 | orchestrator | included: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-4 2025-08-29 17:56:08.018294 | orchestrator | 2025-08-29 17:56:08.018303 | orchestrator | TASK [nova-cell : Get a list of existing cells] ******************************** 2025-08-29 17:56:08.018313 | orchestrator | Friday 29 August 2025 17:55:32 +0000 (0:00:03.330) 0:08:02.787 ********* 2025-08-29 17:56:08.018322 | orchestrator | ok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:56:08.018354 | orchestrator | 2025-08-29 17:56:08.018365 | orchestrator | TASK [nova-cell : Extract current cell settings from list] ********************* 2025-08-29 17:56:08.018374 | orchestrator | Friday 29 August 2025 17:55:44 +0000 (0:00:11.945) 0:08:14.732 ********* 2025-08-29 17:56:08.018384 | orchestrator | ok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:56:08.018393 | orchestrator | 2025-08-29 17:56:08.018403 | orchestrator | TASK [nova-cell : Fail if cell settings not found] ***************************** 2025-08-29 17:56:08.018412 | orchestrator | Friday 29 August 2025 17:55:46 +0000 (0:00:01.343) 0:08:16.075 ********* 2025-08-29 17:56:08.018422 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.018440 | orchestrator | 2025-08-29 17:56:08.018449 | orchestrator | TASK [nova-cell : Discover nova hosts] ***************************************** 2025-08-29 17:56:08.018459 | orchestrator | Friday 29 August 2025 17:55:47 +0000 (0:00:01.361) 0:08:17.437 ********* 2025-08-29 17:56:08.018469 | orchestrator | ok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] 2025-08-29 17:56:08.018478 | orchestrator | 2025-08-29 17:56:08.018488 | orchestrator | TASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************ 2025-08-29 17:56:08.018497 | orchestrator | Friday 29 August 2025 17:55:58 +0000 (0:00:10.782) 0:08:28.220 ********* 2025-08-29 17:56:08.018507 | orchestrator | ok: [testbed-node-3] 2025-08-29 17:56:08.018517 | orchestrator | ok: [testbed-node-5] 2025-08-29 17:56:08.018527 | orchestrator | ok: [testbed-node-4] 2025-08-29 17:56:08.018536 | orchestrator | ok: [testbed-node-0] 2025-08-29 17:56:08.018546 | orchestrator | ok: [testbed-node-1] 2025-08-29 17:56:08.018555 | orchestrator | ok: [testbed-node-2] 2025-08-29 17:56:08.018565 | orchestrator | 2025-08-29 17:56:08.018575 | orchestrator | PLAY [Refresh nova scheduler cell cache] *************************************** 2025-08-29 17:56:08.018584 | orchestrator | 2025-08-29 17:56:08.018594 | orchestrator | TASK [nova : Refresh cell cache in nova scheduler] ***************************** 2025-08-29 17:56:08.018603 | orchestrator | Friday 29 August 2025 17:56:00 +0000 (0:00:01.758) 0:08:29.978 ********* 2025-08-29 17:56:08.018613 | orchestrator | changed: [testbed-node-0] 2025-08-29 17:56:08.018623 | orchestrator | changed: [testbed-node-1] 2025-08-29 17:56:08.018632 | orchestrator | changed: [testbed-node-2] 2025-08-29 17:56:08.018642 | orchestrator | 2025-08-29 17:56:08.018652 | orchestrator | PLAY [Reload global Nova super conductor services] ***************************** 2025-08-29 17:56:08.018661 | orchestrator | 2025-08-29 17:56:08.018671 | orchestrator | TASK [nova : Reload nova super conductor services to remove RPC version pin] *** 2025-08-29 17:56:08.018680 | orchestrator | Friday 29 August 2025 17:56:00 +0000 (0:00:00.929) 0:08:30.907 ********* 2025-08-29 17:56:08.018690 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.018700 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.018709 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.018719 | orchestrator | 2025-08-29 17:56:08.018729 | orchestrator | PLAY [Reload Nova cell services] *********************************************** 2025-08-29 17:56:08.018738 | orchestrator | 2025-08-29 17:56:08.018748 | orchestrator | TASK [nova-cell : Reload nova cell services to remove RPC version cap] ********* 2025-08-29 17:56:08.018757 | orchestrator | Friday 29 August 2025 17:56:01 +0000 (0:00:00.723) 0:08:31.631 ********* 2025-08-29 17:56:08.018767 | orchestrator | skipping: [testbed-node-3] => (item=nova-conductor)  2025-08-29 17:56:08.018777 | orchestrator | skipping: [testbed-node-3] => (item=nova-compute)  2025-08-29 17:56:08.018786 | orchestrator | skipping: [testbed-node-3] => (item=nova-compute-ironic)  2025-08-29 17:56:08.018796 | orchestrator | skipping: [testbed-node-3] => (item=nova-novncproxy)  2025-08-29 17:56:08.018805 | orchestrator | skipping: [testbed-node-3] => (item=nova-serialproxy)  2025-08-29 17:56:08.018815 | orchestrator | skipping: [testbed-node-3] => (item=nova-spicehtml5proxy)  2025-08-29 17:56:08.018825 | orchestrator | skipping: [testbed-node-3] 2025-08-29 17:56:08.018834 | orchestrator | skipping: [testbed-node-4] => (item=nova-conductor)  2025-08-29 17:56:08.018844 | orchestrator | skipping: [testbed-node-4] => (item=nova-compute)  2025-08-29 17:56:08.018854 | orchestrator | skipping: [testbed-node-4] => (item=nova-compute-ironic)  2025-08-29 17:56:08.018863 | orchestrator | skipping: [testbed-node-4] => (item=nova-novncproxy)  2025-08-29 17:56:08.018878 | orchestrator | skipping: [testbed-node-4] => (item=nova-serialproxy)  2025-08-29 17:56:08.018888 | orchestrator | skipping: [testbed-node-4] => (item=nova-spicehtml5proxy)  2025-08-29 17:56:08.018897 | orchestrator | skipping: [testbed-node-4] 2025-08-29 17:56:08.018907 | orchestrator | skipping: [testbed-node-5] => (item=nova-conductor)  2025-08-29 17:56:08.018916 | orchestrator | skipping: [testbed-node-5] => (item=nova-compute)  2025-08-29 17:56:08.018926 | orchestrator | skipping: [testbed-node-5] => (item=nova-compute-ironic)  2025-08-29 17:56:08.018943 | orchestrator | skipping: [testbed-node-5] => (item=nova-novncproxy)  2025-08-29 17:56:08.018953 | orchestrator | skipping: [testbed-node-5] => (item=nova-serialproxy)  2025-08-29 17:56:08.018963 | orchestrator | skipping: [testbed-node-5] => (item=nova-spicehtml5proxy)  2025-08-29 17:56:08.018972 | orchestrator | skipping: [testbed-node-5] 2025-08-29 17:56:08.018982 | orchestrator | skipping: [testbed-node-0] => (item=nova-conductor)  2025-08-29 17:56:08.018999 | orchestrator | skipping: [testbed-node-0] => (item=nova-compute)  2025-08-29 17:56:08.019009 | orchestrator | skipping: [testbed-node-0] => (item=nova-compute-ironic)  2025-08-29 17:56:08.019018 | orchestrator | skipping: [testbed-node-0] => (item=nova-novncproxy)  2025-08-29 17:56:08.019028 | orchestrator | skipping: [testbed-node-0] => (item=nova-serialproxy)  2025-08-29 17:56:08.019037 | orchestrator | skipping: [testbed-node-0] => (item=nova-spicehtml5proxy)  2025-08-29 17:56:08.019047 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.019057 | orchestrator | skipping: [testbed-node-1] => (item=nova-conductor)  2025-08-29 17:56:08.019066 | orchestrator | skipping: [testbed-node-1] => (item=nova-compute)  2025-08-29 17:56:08.019076 | orchestrator | skipping: [testbed-node-1] => (item=nova-compute-ironic)  2025-08-29 17:56:08.019086 | orchestrator | skipping: [testbed-node-1] => (item=nova-novncproxy)  2025-08-29 17:56:08.019095 | orchestrator | skipping: [testbed-node-1] => (item=nova-serialproxy)  2025-08-29 17:56:08.019105 | orchestrator | skipping: [testbed-node-1] => (item=nova-spicehtml5proxy)  2025-08-29 17:56:08.019115 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.019124 | orchestrator | skipping: [testbed-node-2] => (item=nova-conductor)  2025-08-29 17:56:08.019134 | orchestrator | skipping: [testbed-node-2] => (item=nova-compute)  2025-08-29 17:56:08.019144 | orchestrator | skipping: [testbed-node-2] => (item=nova-compute-ironic)  2025-08-29 17:56:08.019154 | orchestrator | skipping: [testbed-node-2] => (item=nova-novncproxy)  2025-08-29 17:56:08.019163 | orchestrator | skipping: [testbed-node-2] => (item=nova-serialproxy)  2025-08-29 17:56:08.019172 | orchestrator | skipping: [testbed-node-2] => (item=nova-spicehtml5proxy)  2025-08-29 17:56:08.019182 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.019192 | orchestrator | 2025-08-29 17:56:08.019201 | orchestrator | PLAY [Reload global Nova API services] ***************************************** 2025-08-29 17:56:08.019211 | orchestrator | 2025-08-29 17:56:08.019221 | orchestrator | TASK [nova : Reload nova API services to remove RPC version pin] *************** 2025-08-29 17:56:08.019230 | orchestrator | Friday 29 August 2025 17:56:03 +0000 (0:00:01.334) 0:08:32.965 ********* 2025-08-29 17:56:08.019240 | orchestrator | skipping: [testbed-node-0] => (item=nova-scheduler)  2025-08-29 17:56:08.019250 | orchestrator | skipping: [testbed-node-0] => (item=nova-api)  2025-08-29 17:56:08.019259 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.019269 | orchestrator | skipping: [testbed-node-1] => (item=nova-scheduler)  2025-08-29 17:56:08.019278 | orchestrator | skipping: [testbed-node-1] => (item=nova-api)  2025-08-29 17:56:08.019288 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.019298 | orchestrator | skipping: [testbed-node-2] => (item=nova-scheduler)  2025-08-29 17:56:08.019307 | orchestrator | skipping: [testbed-node-2] => (item=nova-api)  2025-08-29 17:56:08.019317 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.019326 | orchestrator | 2025-08-29 17:56:08.019351 | orchestrator | PLAY [Run Nova API online data migrations] ************************************* 2025-08-29 17:56:08.019360 | orchestrator | 2025-08-29 17:56:08.019370 | orchestrator | TASK [nova : Run Nova API online database migrations] ************************** 2025-08-29 17:56:08.019380 | orchestrator | Friday 29 August 2025 17:56:03 +0000 (0:00:00.591) 0:08:33.557 ********* 2025-08-29 17:56:08.019389 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.019399 | orchestrator | 2025-08-29 17:56:08.019408 | orchestrator | PLAY [Run Nova cell online data migrations] ************************************ 2025-08-29 17:56:08.019418 | orchestrator | 2025-08-29 17:56:08.019428 | orchestrator | TASK [nova-cell : Run Nova cell online database migrations] ******************** 2025-08-29 17:56:08.019448 | orchestrator | Friday 29 August 2025 17:56:04 +0000 (0:00:00.845) 0:08:34.402 ********* 2025-08-29 17:56:08.019458 | orchestrator | skipping: [testbed-node-0] 2025-08-29 17:56:08.019467 | orchestrator | skipping: [testbed-node-1] 2025-08-29 17:56:08.019477 | orchestrator | skipping: [testbed-node-2] 2025-08-29 17:56:08.019486 | orchestrator | 2025-08-29 17:56:08.019496 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:56:08.019505 | orchestrator | testbed-manager : ok=3  changed=3  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:56:08.019516 | orchestrator | testbed-node-0 : ok=54  changed=35  unreachable=0 failed=0 skipped=44  rescued=0 ignored=0 2025-08-29 17:56:08.019526 | orchestrator | testbed-node-1 : ok=27  changed=19  unreachable=0 failed=0 skipped=51  rescued=0 ignored=0 2025-08-29 17:56:08.019536 | orchestrator | testbed-node-2 : ok=27  changed=19  unreachable=0 failed=0 skipped=51  rescued=0 ignored=0 2025-08-29 17:56:08.019550 | orchestrator | testbed-node-3 : ok=38  changed=27  unreachable=0 failed=0 skipped=21  rescued=0 ignored=0 2025-08-29 17:56:08.019561 | orchestrator | testbed-node-4 : ok=42  changed=27  unreachable=0 failed=0 skipped=18  rescued=0 ignored=0 2025-08-29 17:56:08.019570 | orchestrator | testbed-node-5 : ok=37  changed=27  unreachable=0 failed=0 skipped=19  rescued=0 ignored=0 2025-08-29 17:56:08.019580 | orchestrator | 2025-08-29 17:56:08.019590 | orchestrator | 2025-08-29 17:56:08.019599 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:56:08.019609 | orchestrator | Friday 29 August 2025 17:56:04 +0000 (0:00:00.425) 0:08:34.828 ********* 2025-08-29 17:56:08.019619 | orchestrator | =============================================================================== 2025-08-29 17:56:08.019635 | orchestrator | nova-cell : Restart nova-libvirt container ----------------------------- 45.82s 2025-08-29 17:56:08.019645 | orchestrator | nova : Running Nova API bootstrap container ---------------------------- 34.45s 2025-08-29 17:56:08.019655 | orchestrator | nova-cell : Restart nova-ssh container --------------------------------- 26.52s 2025-08-29 17:56:08.019664 | orchestrator | nova : Restart nova-scheduler container -------------------------------- 23.75s 2025-08-29 17:56:08.019674 | orchestrator | nova-cell : Running Nova cell bootstrap container ---------------------- 22.55s 2025-08-29 17:56:08.019683 | orchestrator | nova-cell : Waiting for nova-compute services to register themselves --- 22.32s 2025-08-29 17:56:08.019693 | orchestrator | nova-cell : Restart nova-compute container ----------------------------- 22.27s 2025-08-29 17:56:08.019703 | orchestrator | nova : Running Nova API bootstrap container ---------------------------- 17.79s 2025-08-29 17:56:08.019712 | orchestrator | nova : Create cell0 mappings ------------------------------------------- 17.18s 2025-08-29 17:56:08.019722 | orchestrator | nova-cell : Restart nova-novncproxy container -------------------------- 13.89s 2025-08-29 17:56:08.019731 | orchestrator | nova-cell : Create cell ------------------------------------------------ 12.61s 2025-08-29 17:56:08.019741 | orchestrator | nova-cell : Get a list of existing cells ------------------------------- 11.95s 2025-08-29 17:56:08.019750 | orchestrator | nova-cell : Get a list of existing cells ------------------------------- 11.31s 2025-08-29 17:56:08.019760 | orchestrator | nova-cell : Discover nova hosts ---------------------------------------- 10.78s 2025-08-29 17:56:08.019770 | orchestrator | nova-cell : Get a list of existing cells ------------------------------- 10.43s 2025-08-29 17:56:08.019779 | orchestrator | service-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 8.51s 2025-08-29 17:56:08.019789 | orchestrator | nova : Restart nova-api container --------------------------------------- 8.42s 2025-08-29 17:56:08.019799 | orchestrator | nova-cell : Fail if nova-compute service failed to register ------------- 8.24s 2025-08-29 17:56:08.019814 | orchestrator | nova-cell : Restart nova-conductor container ---------------------------- 7.47s 2025-08-29 17:56:08.019824 | orchestrator | nova-cell : Copying files for nova-ssh ---------------------------------- 7.03s 2025-08-29 17:56:08.019833 | orchestrator | 2025-08-29 17:56:08 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:11.052020 | orchestrator | 2025-08-29 17:56:11 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:14.095125 | orchestrator | 2025-08-29 17:56:14 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:17.140292 | orchestrator | 2025-08-29 17:56:17 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:20.184986 | orchestrator | 2025-08-29 17:56:20 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:23.232819 | orchestrator | 2025-08-29 17:56:23 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:26.282881 | orchestrator | 2025-08-29 17:56:26 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:29.328706 | orchestrator | 2025-08-29 17:56:29 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:32.374866 | orchestrator | 2025-08-29 17:56:32 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:35.418211 | orchestrator | 2025-08-29 17:56:35 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:38.463477 | orchestrator | 2025-08-29 17:56:38 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:41.511278 | orchestrator | 2025-08-29 17:56:41 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:44.558541 | orchestrator | 2025-08-29 17:56:44 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:47.597078 | orchestrator | 2025-08-29 17:56:47 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:50.644651 | orchestrator | 2025-08-29 17:56:50 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:53.690092 | orchestrator | 2025-08-29 17:56:53 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:56.734759 | orchestrator | 2025-08-29 17:56:56 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:56:59.786860 | orchestrator | 2025-08-29 17:56:59 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:57:02.831969 | orchestrator | 2025-08-29 17:57:02 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:57:05.873045 | orchestrator | 2025-08-29 17:57:05 | INFO  | Wait 1 second(s) until refresh of running tasks 2025-08-29 17:57:08.918978 | orchestrator | 2025-08-29 17:57:09.228029 | orchestrator | 2025-08-29 17:57:09.231214 | orchestrator | --> DEPLOY IN A NUTSHELL -- END -- Fri Aug 29 17:57:09 UTC 2025 2025-08-29 17:57:09.231392 | orchestrator | 2025-08-29 17:57:09.747602 | orchestrator | ok: Runtime: 0:34:32.968264 2025-08-29 17:57:09.988127 | 2025-08-29 17:57:09.988280 | TASK [Bootstrap services] 2025-08-29 17:57:10.720225 | orchestrator | 2025-08-29 17:57:10.720538 | orchestrator | # BOOTSTRAP 2025-08-29 17:57:10.720580 | orchestrator | 2025-08-29 17:57:10.720602 | orchestrator | + set -e 2025-08-29 17:57:10.720625 | orchestrator | + echo 2025-08-29 17:57:10.720647 | orchestrator | + echo '# BOOTSTRAP' 2025-08-29 17:57:10.720676 | orchestrator | + echo 2025-08-29 17:57:10.720737 | orchestrator | + sh -c /opt/configuration/scripts/bootstrap-services.sh 2025-08-29 17:57:10.731011 | orchestrator | + set -e 2025-08-29 17:57:10.731091 | orchestrator | + sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh 2025-08-29 17:57:15.430135 | orchestrator | 2025-08-29 17:57:15 | INFO  | It takes a moment until task c1ce51bb-642d-4ffb-950c-02ccc0141eab (flavor-manager) has been started and output is visible here. 2025-08-29 17:57:24.016257 | orchestrator | 2025-08-29 17:57:19 | INFO  | Flavor SCS-1V-4 created 2025-08-29 17:57:24.016442 | orchestrator | 2025-08-29 17:57:19 | INFO  | Flavor SCS-2V-8 created 2025-08-29 17:57:24.016465 | orchestrator | 2025-08-29 17:57:20 | INFO  | Flavor SCS-4V-16 created 2025-08-29 17:57:24.016481 | orchestrator | 2025-08-29 17:57:20 | INFO  | Flavor SCS-8V-32 created 2025-08-29 17:57:24.016503 | orchestrator | 2025-08-29 17:57:20 | INFO  | Flavor SCS-1V-2 created 2025-08-29 17:57:24.016522 | orchestrator | 2025-08-29 17:57:20 | INFO  | Flavor SCS-2V-4 created 2025-08-29 17:57:24.016539 | orchestrator | 2025-08-29 17:57:20 | INFO  | Flavor SCS-4V-8 created 2025-08-29 17:57:24.016552 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-8V-16 created 2025-08-29 17:57:24.016581 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-16V-32 created 2025-08-29 17:57:24.016602 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-1V-8 created 2025-08-29 17:57:24.016622 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-2V-16 created 2025-08-29 17:57:24.016633 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-4V-32 created 2025-08-29 17:57:24.016645 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-1L-1 created 2025-08-29 17:57:24.016662 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-2V-4-20s created 2025-08-29 17:57:24.016681 | orchestrator | 2025-08-29 17:57:21 | INFO  | Flavor SCS-4V-16-100s created 2025-08-29 17:57:24.016701 | orchestrator | 2025-08-29 17:57:22 | INFO  | Flavor SCS-1V-4-10 created 2025-08-29 17:57:24.016716 | orchestrator | 2025-08-29 17:57:22 | INFO  | Flavor SCS-2V-8-20 created 2025-08-29 17:57:24.016727 | orchestrator | 2025-08-29 17:57:22 | INFO  | Flavor SCS-4V-16-50 created 2025-08-29 17:57:24.016740 | orchestrator | 2025-08-29 17:57:22 | INFO  | Flavor SCS-8V-32-100 created 2025-08-29 17:57:24.016760 | orchestrator | 2025-08-29 17:57:22 | INFO  | Flavor SCS-1V-2-5 created 2025-08-29 17:57:24.016779 | orchestrator | 2025-08-29 17:57:22 | INFO  | Flavor SCS-2V-4-10 created 2025-08-29 17:57:24.016798 | orchestrator | 2025-08-29 17:57:22 | INFO  | Flavor SCS-4V-8-20 created 2025-08-29 17:57:24.016811 | orchestrator | 2025-08-29 17:57:23 | INFO  | Flavor SCS-8V-16-50 created 2025-08-29 17:57:24.016822 | orchestrator | 2025-08-29 17:57:23 | INFO  | Flavor SCS-16V-32-100 created 2025-08-29 17:57:24.016833 | orchestrator | 2025-08-29 17:57:23 | INFO  | Flavor SCS-1V-8-20 created 2025-08-29 17:57:24.016844 | orchestrator | 2025-08-29 17:57:23 | INFO  | Flavor SCS-2V-16-50 created 2025-08-29 17:57:24.016855 | orchestrator | 2025-08-29 17:57:23 | INFO  | Flavor SCS-4V-32-100 created 2025-08-29 17:57:24.016866 | orchestrator | 2025-08-29 17:57:23 | INFO  | Flavor SCS-1L-1-5 created 2025-08-29 17:57:26.072801 | orchestrator | 2025-08-29 17:57:26 | INFO  | Trying to run play bootstrap-basic in environment openstack 2025-08-29 17:57:36.182704 | orchestrator | 2025-08-29 17:57:36 | INFO  | Task 10784141-7258-4e99-bae4-5249b674050d (bootstrap-basic) was prepared for execution. 2025-08-29 17:57:36.182821 | orchestrator | 2025-08-29 17:57:36 | INFO  | It takes a moment until task 10784141-7258-4e99-bae4-5249b674050d (bootstrap-basic) has been started and output is visible here. 2025-08-29 17:58:40.912119 | orchestrator | 2025-08-29 17:58:40.912341 | orchestrator | PLAY [Bootstrap basic OpenStack services] ************************************** 2025-08-29 17:58:40.912364 | orchestrator | 2025-08-29 17:58:40.912376 | orchestrator | TASK [Gathering Facts] ********************************************************* 2025-08-29 17:58:40.912387 | orchestrator | Friday 29 August 2025 17:57:39 +0000 (0:00:00.069) 0:00:00.069 ********* 2025-08-29 17:58:40.912399 | orchestrator | ok: [localhost] 2025-08-29 17:58:40.912411 | orchestrator | 2025-08-29 17:58:40.912422 | orchestrator | TASK [Get volume type LUKS] **************************************************** 2025-08-29 17:58:40.912436 | orchestrator | Friday 29 August 2025 17:57:41 +0000 (0:00:01.910) 0:00:01.980 ********* 2025-08-29 17:58:40.912447 | orchestrator | ok: [localhost] 2025-08-29 17:58:40.912458 | orchestrator | 2025-08-29 17:58:40.912470 | orchestrator | TASK [Create volume type LUKS] ************************************************* 2025-08-29 17:58:40.912481 | orchestrator | Friday 29 August 2025 17:57:50 +0000 (0:00:08.517) 0:00:10.498 ********* 2025-08-29 17:58:40.912492 | orchestrator | changed: [localhost] 2025-08-29 17:58:40.912503 | orchestrator | 2025-08-29 17:58:40.912515 | orchestrator | TASK [Get volume type local] *************************************************** 2025-08-29 17:58:40.912526 | orchestrator | Friday 29 August 2025 17:57:58 +0000 (0:00:07.945) 0:00:18.444 ********* 2025-08-29 17:58:40.912537 | orchestrator | ok: [localhost] 2025-08-29 17:58:40.912548 | orchestrator | 2025-08-29 17:58:40.912560 | orchestrator | TASK [Create volume type local] ************************************************ 2025-08-29 17:58:40.912571 | orchestrator | Friday 29 August 2025 17:58:06 +0000 (0:00:07.780) 0:00:26.224 ********* 2025-08-29 17:58:40.912582 | orchestrator | changed: [localhost] 2025-08-29 17:58:40.912598 | orchestrator | 2025-08-29 17:58:40.912609 | orchestrator | TASK [Create public network] *************************************************** 2025-08-29 17:58:40.912620 | orchestrator | Friday 29 August 2025 17:58:13 +0000 (0:00:07.054) 0:00:33.278 ********* 2025-08-29 17:58:40.912631 | orchestrator | changed: [localhost] 2025-08-29 17:58:40.912642 | orchestrator | 2025-08-29 17:58:40.912655 | orchestrator | TASK [Set public network to default] ******************************************* 2025-08-29 17:58:40.912667 | orchestrator | Friday 29 August 2025 17:58:20 +0000 (0:00:07.330) 0:00:40.609 ********* 2025-08-29 17:58:40.912679 | orchestrator | changed: [localhost] 2025-08-29 17:58:40.912692 | orchestrator | 2025-08-29 17:58:40.912716 | orchestrator | TASK [Create public subnet] **************************************************** 2025-08-29 17:58:40.912728 | orchestrator | Friday 29 August 2025 17:58:27 +0000 (0:00:07.358) 0:00:47.968 ********* 2025-08-29 17:58:40.912741 | orchestrator | changed: [localhost] 2025-08-29 17:58:40.912753 | orchestrator | 2025-08-29 17:58:40.912765 | orchestrator | TASK [Create default IPv4 subnet pool] ***************************************** 2025-08-29 17:58:40.912777 | orchestrator | Friday 29 August 2025 17:58:32 +0000 (0:00:04.880) 0:00:52.848 ********* 2025-08-29 17:58:40.912789 | orchestrator | changed: [localhost] 2025-08-29 17:58:40.912801 | orchestrator | 2025-08-29 17:58:40.912813 | orchestrator | TASK [Create manager role] ***************************************************** 2025-08-29 17:58:40.912825 | orchestrator | Friday 29 August 2025 17:58:37 +0000 (0:00:04.331) 0:00:57.180 ********* 2025-08-29 17:58:40.912837 | orchestrator | ok: [localhost] 2025-08-29 17:58:40.912849 | orchestrator | 2025-08-29 17:58:40.912862 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 17:58:40.912874 | orchestrator | localhost : ok=10  changed=6  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 17:58:40.912887 | orchestrator | 2025-08-29 17:58:40.912900 | orchestrator | 2025-08-29 17:58:40.912912 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 17:58:40.912924 | orchestrator | Friday 29 August 2025 17:58:40 +0000 (0:00:03.596) 0:01:00.776 ********* 2025-08-29 17:58:40.912960 | orchestrator | =============================================================================== 2025-08-29 17:58:40.912973 | orchestrator | Get volume type LUKS ---------------------------------------------------- 8.52s 2025-08-29 17:58:40.912986 | orchestrator | Create volume type LUKS ------------------------------------------------- 7.95s 2025-08-29 17:58:40.912998 | orchestrator | Get volume type local --------------------------------------------------- 7.78s 2025-08-29 17:58:40.913010 | orchestrator | Set public network to default ------------------------------------------- 7.36s 2025-08-29 17:58:40.913021 | orchestrator | Create public network --------------------------------------------------- 7.33s 2025-08-29 17:58:40.913032 | orchestrator | Create volume type local ------------------------------------------------ 7.05s 2025-08-29 17:58:40.913043 | orchestrator | Create public subnet ---------------------------------------------------- 4.88s 2025-08-29 17:58:40.913054 | orchestrator | Create default IPv4 subnet pool ----------------------------------------- 4.33s 2025-08-29 17:58:40.913064 | orchestrator | Create manager role ----------------------------------------------------- 3.60s 2025-08-29 17:58:40.913075 | orchestrator | Gathering Facts --------------------------------------------------------- 1.91s 2025-08-29 17:58:43.193960 | orchestrator | 2025-08-29 17:58:43 | INFO  | It takes a moment until task b2438212-87e3-4eea-9716-3941b1050952 (image-manager) has been started and output is visible here. 2025-08-29 17:59:24.502536 | orchestrator | 2025-08-29 17:58:46 | INFO  | Processing image 'Cirros 0.6.2' 2025-08-29 17:59:24.502623 | orchestrator | 2025-08-29 17:58:47 | INFO  | Tested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302 2025-08-29 17:59:24.502635 | orchestrator | 2025-08-29 17:58:47 | INFO  | Importing image Cirros 0.6.2 2025-08-29 17:59:24.502641 | orchestrator | 2025-08-29 17:58:47 | INFO  | Importing from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img 2025-08-29 17:59:24.502648 | orchestrator | 2025-08-29 17:58:48 | INFO  | Waiting for image to leave queued state... 2025-08-29 17:59:24.502654 | orchestrator | 2025-08-29 17:58:50 | INFO  | Waiting for import to complete... 2025-08-29 17:59:24.502659 | orchestrator | 2025-08-29 17:59:01 | INFO  | Import of 'Cirros 0.6.2' successfully completed, reloading images 2025-08-29 17:59:24.502665 | orchestrator | 2025-08-29 17:59:01 | INFO  | Checking parameters of 'Cirros 0.6.2' 2025-08-29 17:59:24.502670 | orchestrator | 2025-08-29 17:59:01 | INFO  | Setting internal_version = 0.6.2 2025-08-29 17:59:24.502675 | orchestrator | 2025-08-29 17:59:01 | INFO  | Setting image_original_user = cirros 2025-08-29 17:59:24.502681 | orchestrator | 2025-08-29 17:59:01 | INFO  | Adding tag os:cirros 2025-08-29 17:59:24.502686 | orchestrator | 2025-08-29 17:59:01 | INFO  | Setting property architecture: x86_64 2025-08-29 17:59:24.502691 | orchestrator | 2025-08-29 17:59:01 | INFO  | Setting property hw_disk_bus: scsi 2025-08-29 17:59:24.502696 | orchestrator | 2025-08-29 17:59:02 | INFO  | Setting property hw_rng_model: virtio 2025-08-29 17:59:24.502701 | orchestrator | 2025-08-29 17:59:02 | INFO  | Setting property hw_scsi_model: virtio-scsi 2025-08-29 17:59:24.502707 | orchestrator | 2025-08-29 17:59:02 | INFO  | Setting property hw_watchdog_action: reset 2025-08-29 17:59:24.502712 | orchestrator | 2025-08-29 17:59:02 | INFO  | Setting property hypervisor_type: qemu 2025-08-29 17:59:24.502717 | orchestrator | 2025-08-29 17:59:02 | INFO  | Setting property os_distro: cirros 2025-08-29 17:59:24.502722 | orchestrator | 2025-08-29 17:59:03 | INFO  | Setting property replace_frequency: never 2025-08-29 17:59:24.502727 | orchestrator | 2025-08-29 17:59:03 | INFO  | Setting property uuid_validity: none 2025-08-29 17:59:24.502732 | orchestrator | 2025-08-29 17:59:03 | INFO  | Setting property provided_until: none 2025-08-29 17:59:24.502754 | orchestrator | 2025-08-29 17:59:03 | INFO  | Setting property image_description: Cirros 2025-08-29 17:59:24.502765 | orchestrator | 2025-08-29 17:59:03 | INFO  | Setting property image_name: Cirros 2025-08-29 17:59:24.502770 | orchestrator | 2025-08-29 17:59:04 | INFO  | Setting property internal_version: 0.6.2 2025-08-29 17:59:24.502778 | orchestrator | 2025-08-29 17:59:04 | INFO  | Setting property image_original_user: cirros 2025-08-29 17:59:24.502783 | orchestrator | 2025-08-29 17:59:04 | INFO  | Setting property os_version: 0.6.2 2025-08-29 17:59:24.502789 | orchestrator | 2025-08-29 17:59:04 | INFO  | Setting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img 2025-08-29 17:59:24.502796 | orchestrator | 2025-08-29 17:59:04 | INFO  | Setting property image_build_date: 2023-05-30 2025-08-29 17:59:24.502801 | orchestrator | 2025-08-29 17:59:05 | INFO  | Checking status of 'Cirros 0.6.2' 2025-08-29 17:59:24.502805 | orchestrator | 2025-08-29 17:59:05 | INFO  | Checking visibility of 'Cirros 0.6.2' 2025-08-29 17:59:24.502811 | orchestrator | 2025-08-29 17:59:05 | INFO  | Setting visibility of 'Cirros 0.6.2' to 'public' 2025-08-29 17:59:24.502816 | orchestrator | 2025-08-29 17:59:05 | INFO  | Processing image 'Cirros 0.6.3' 2025-08-29 17:59:24.502820 | orchestrator | 2025-08-29 17:59:05 | INFO  | Tested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302 2025-08-29 17:59:24.502826 | orchestrator | 2025-08-29 17:59:05 | INFO  | Importing image Cirros 0.6.3 2025-08-29 17:59:24.502831 | orchestrator | 2025-08-29 17:59:05 | INFO  | Importing from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img 2025-08-29 17:59:24.502836 | orchestrator | 2025-08-29 17:59:06 | INFO  | Waiting for image to leave queued state... 2025-08-29 17:59:24.502841 | orchestrator | 2025-08-29 17:59:08 | INFO  | Waiting for import to complete... 2025-08-29 17:59:24.502846 | orchestrator | 2025-08-29 17:59:18 | INFO  | Import of 'Cirros 0.6.3' successfully completed, reloading images 2025-08-29 17:59:24.502861 | orchestrator | 2025-08-29 17:59:19 | INFO  | Checking parameters of 'Cirros 0.6.3' 2025-08-29 17:59:24.502866 | orchestrator | 2025-08-29 17:59:19 | INFO  | Setting internal_version = 0.6.3 2025-08-29 17:59:24.502871 | orchestrator | 2025-08-29 17:59:19 | INFO  | Setting image_original_user = cirros 2025-08-29 17:59:24.502876 | orchestrator | 2025-08-29 17:59:19 | INFO  | Adding tag os:cirros 2025-08-29 17:59:24.502881 | orchestrator | 2025-08-29 17:59:19 | INFO  | Setting property architecture: x86_64 2025-08-29 17:59:24.502887 | orchestrator | 2025-08-29 17:59:19 | INFO  | Setting property hw_disk_bus: scsi 2025-08-29 17:59:24.502892 | orchestrator | 2025-08-29 17:59:20 | INFO  | Setting property hw_rng_model: virtio 2025-08-29 17:59:24.502897 | orchestrator | 2025-08-29 17:59:20 | INFO  | Setting property hw_scsi_model: virtio-scsi 2025-08-29 17:59:24.502902 | orchestrator | 2025-08-29 17:59:20 | INFO  | Setting property hw_watchdog_action: reset 2025-08-29 17:59:24.502907 | orchestrator | 2025-08-29 17:59:20 | INFO  | Setting property hypervisor_type: qemu 2025-08-29 17:59:24.502912 | orchestrator | 2025-08-29 17:59:21 | INFO  | Setting property os_distro: cirros 2025-08-29 17:59:24.502917 | orchestrator | 2025-08-29 17:59:21 | INFO  | Setting property replace_frequency: never 2025-08-29 17:59:24.502922 | orchestrator | 2025-08-29 17:59:21 | INFO  | Setting property uuid_validity: none 2025-08-29 17:59:24.502932 | orchestrator | 2025-08-29 17:59:21 | INFO  | Setting property provided_until: none 2025-08-29 17:59:24.502937 | orchestrator | 2025-08-29 17:59:22 | INFO  | Setting property image_description: Cirros 2025-08-29 17:59:24.502942 | orchestrator | 2025-08-29 17:59:22 | INFO  | Setting property image_name: Cirros 2025-08-29 17:59:24.502948 | orchestrator | 2025-08-29 17:59:22 | INFO  | Setting property internal_version: 0.6.3 2025-08-29 17:59:24.502952 | orchestrator | 2025-08-29 17:59:22 | INFO  | Setting property image_original_user: cirros 2025-08-29 17:59:24.502958 | orchestrator | 2025-08-29 17:59:23 | INFO  | Setting property os_version: 0.6.3 2025-08-29 17:59:24.502971 | orchestrator | 2025-08-29 17:59:23 | INFO  | Setting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img 2025-08-29 17:59:24.502977 | orchestrator | 2025-08-29 17:59:23 | INFO  | Setting property image_build_date: 2024-09-26 2025-08-29 17:59:24.502982 | orchestrator | 2025-08-29 17:59:23 | INFO  | Checking status of 'Cirros 0.6.3' 2025-08-29 17:59:24.502987 | orchestrator | 2025-08-29 17:59:23 | INFO  | Checking visibility of 'Cirros 0.6.3' 2025-08-29 17:59:24.502995 | orchestrator | 2025-08-29 17:59:23 | INFO  | Setting visibility of 'Cirros 0.6.3' to 'public' 2025-08-29 17:59:24.837962 | orchestrator | + sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh 2025-08-29 17:59:26.843428 | orchestrator | 2025-08-29 17:59:26 | INFO  | date: 2025-08-29 2025-08-29 17:59:26.843737 | orchestrator | 2025-08-29 17:59:26 | INFO  | image: octavia-amphora-haproxy-2024.2.20250829.qcow2 2025-08-29 17:59:26.843770 | orchestrator | 2025-08-29 17:59:26 | INFO  | url: https://swift.services.a.regiocloud.tech/swift/v1/AUTH_b182637428444b9aa302bb8d5a5a418c/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20250829.qcow2 2025-08-29 17:59:26.843951 | orchestrator | 2025-08-29 17:59:26 | INFO  | checksum_url: https://swift.services.a.regiocloud.tech/swift/v1/AUTH_b182637428444b9aa302bb8d5a5a418c/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20250829.qcow2.CHECKSUM 2025-08-29 17:59:26.870339 | orchestrator | 2025-08-29 17:59:26 | INFO  | checksum: 9bd11944634778935b43eb626302bc74d657e4c319fdb6fd625fdfeb36ffc69d 2025-08-29 17:59:26.941744 | orchestrator | 2025-08-29 17:59:26 | INFO  | It takes a moment until task 1f92ef0d-9793-4c33-b814-ef9945fd9207 (image-manager) has been started and output is visible here. 2025-08-29 18:00:28.535286 | orchestrator | /usr/local/lib/python3.13/site-packages/openstack_image_manager/__init__.py:5: UserWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html. The pkg_resources package is slated for removal as early as 2025-11-30. Refrain from using this package or pin to Setuptools<81. 2025-08-29 18:00:28.535401 | orchestrator | from pkg_resources import get_distribution, DistributionNotFound 2025-08-29 18:00:28.535410 | orchestrator | 2025-08-29 17:59:29 | INFO  | Processing image 'OpenStack Octavia Amphora 2025-08-29' 2025-08-29 18:00:28.535421 | orchestrator | 2025-08-29 17:59:29 | INFO  | Tested URL https://swift.services.a.regiocloud.tech/swift/v1/AUTH_b182637428444b9aa302bb8d5a5a418c/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20250829.qcow2: 200 2025-08-29 18:00:28.535437 | orchestrator | 2025-08-29 17:59:29 | INFO  | Importing image OpenStack Octavia Amphora 2025-08-29 2025-08-29 18:00:28.535465 | orchestrator | 2025-08-29 17:59:29 | INFO  | Importing from URL https://swift.services.a.regiocloud.tech/swift/v1/AUTH_b182637428444b9aa302bb8d5a5a418c/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20250829.qcow2 2025-08-29 18:00:28.535473 | orchestrator | 2025-08-29 17:59:30 | INFO  | Waiting for image to leave queued state... 2025-08-29 18:00:28.535496 | orchestrator | 2025-08-29 17:59:32 | INFO  | Waiting for import to complete... 2025-08-29 18:00:28.535502 | orchestrator | 2025-08-29 17:59:42 | INFO  | Waiting for import to complete... 2025-08-29 18:00:28.535507 | orchestrator | 2025-08-29 17:59:52 | INFO  | Waiting for import to complete... 2025-08-29 18:00:28.535511 | orchestrator | 2025-08-29 18:00:03 | INFO  | Waiting for import to complete... 2025-08-29 18:00:28.535516 | orchestrator | 2025-08-29 18:00:13 | INFO  | Waiting for import to complete... 2025-08-29 18:00:28.535521 | orchestrator | 2025-08-29 18:00:23 | INFO  | Import of 'OpenStack Octavia Amphora 2025-08-29' successfully completed, reloading images 2025-08-29 18:00:28.535527 | orchestrator | 2025-08-29 18:00:24 | INFO  | Checking parameters of 'OpenStack Octavia Amphora 2025-08-29' 2025-08-29 18:00:28.535532 | orchestrator | 2025-08-29 18:00:24 | INFO  | Setting internal_version = 2025-08-29 2025-08-29 18:00:28.535536 | orchestrator | 2025-08-29 18:00:24 | INFO  | Setting image_original_user = ubuntu 2025-08-29 18:00:28.535541 | orchestrator | 2025-08-29 18:00:24 | INFO  | Adding tag amphora 2025-08-29 18:00:28.535546 | orchestrator | 2025-08-29 18:00:24 | INFO  | Adding tag os:ubuntu 2025-08-29 18:00:28.535551 | orchestrator | 2025-08-29 18:00:24 | INFO  | Setting property architecture: x86_64 2025-08-29 18:00:28.535555 | orchestrator | 2025-08-29 18:00:24 | INFO  | Setting property hw_disk_bus: scsi 2025-08-29 18:00:28.535560 | orchestrator | 2025-08-29 18:00:24 | INFO  | Setting property hw_rng_model: virtio 2025-08-29 18:00:28.535571 | orchestrator | 2025-08-29 18:00:25 | INFO  | Setting property hw_scsi_model: virtio-scsi 2025-08-29 18:00:28.535576 | orchestrator | 2025-08-29 18:00:25 | INFO  | Setting property hw_watchdog_action: reset 2025-08-29 18:00:28.535581 | orchestrator | 2025-08-29 18:00:25 | INFO  | Setting property hypervisor_type: qemu 2025-08-29 18:00:28.535585 | orchestrator | 2025-08-29 18:00:25 | INFO  | Setting property os_distro: ubuntu 2025-08-29 18:00:28.535590 | orchestrator | 2025-08-29 18:00:25 | INFO  | Setting property replace_frequency: quarterly 2025-08-29 18:00:28.535594 | orchestrator | 2025-08-29 18:00:26 | INFO  | Setting property uuid_validity: last-1 2025-08-29 18:00:28.535599 | orchestrator | 2025-08-29 18:00:26 | INFO  | Setting property provided_until: none 2025-08-29 18:00:28.535604 | orchestrator | 2025-08-29 18:00:26 | INFO  | Setting property image_description: OpenStack Octavia Amphora 2025-08-29 18:00:28.535608 | orchestrator | 2025-08-29 18:00:26 | INFO  | Setting property image_name: OpenStack Octavia Amphora 2025-08-29 18:00:28.535613 | orchestrator | 2025-08-29 18:00:27 | INFO  | Setting property internal_version: 2025-08-29 2025-08-29 18:00:28.535618 | orchestrator | 2025-08-29 18:00:27 | INFO  | Setting property image_original_user: ubuntu 2025-08-29 18:00:28.535622 | orchestrator | 2025-08-29 18:00:27 | INFO  | Setting property os_version: 2025-08-29 2025-08-29 18:00:28.535627 | orchestrator | 2025-08-29 18:00:27 | INFO  | Setting property image_source: https://swift.services.a.regiocloud.tech/swift/v1/AUTH_b182637428444b9aa302bb8d5a5a418c/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20250829.qcow2 2025-08-29 18:00:28.535642 | orchestrator | 2025-08-29 18:00:27 | INFO  | Setting property image_build_date: 2025-08-29 2025-08-29 18:00:28.535647 | orchestrator | 2025-08-29 18:00:28 | INFO  | Checking status of 'OpenStack Octavia Amphora 2025-08-29' 2025-08-29 18:00:28.535652 | orchestrator | 2025-08-29 18:00:28 | INFO  | Checking visibility of 'OpenStack Octavia Amphora 2025-08-29' 2025-08-29 18:00:28.535661 | orchestrator | 2025-08-29 18:00:28 | INFO  | Processing image 'Cirros 0.6.3' (removal candidate) 2025-08-29 18:00:28.535665 | orchestrator | 2025-08-29 18:00:28 | WARNING  | No image definition found for 'Cirros 0.6.3', image will be ignored 2025-08-29 18:00:28.535670 | orchestrator | 2025-08-29 18:00:28 | INFO  | Processing image 'Cirros 0.6.2' (removal candidate) 2025-08-29 18:00:28.535675 | orchestrator | 2025-08-29 18:00:28 | WARNING  | No image definition found for 'Cirros 0.6.2', image will be ignored 2025-08-29 18:00:29.162804 | orchestrator | ok: Runtime: 0:03:18.531037 2025-08-29 18:00:29.227974 | 2025-08-29 18:00:29.228130 | TASK [Run checks] 2025-08-29 18:00:29.911933 | orchestrator | + set -e 2025-08-29 18:00:29.912064 | orchestrator | + source /opt/configuration/scripts/include.sh 2025-08-29 18:00:29.912074 | orchestrator | ++ export INTERACTIVE=false 2025-08-29 18:00:29.912083 | orchestrator | ++ INTERACTIVE=false 2025-08-29 18:00:29.912089 | orchestrator | ++ export OSISM_APPLY_RETRY=1 2025-08-29 18:00:29.912093 | orchestrator | ++ OSISM_APPLY_RETRY=1 2025-08-29 18:00:29.912100 | orchestrator | + source /opt/configuration/scripts/manager-version.sh 2025-08-29 18:00:29.913109 | orchestrator | +++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml 2025-08-29 18:00:29.920162 | orchestrator | 2025-08-29 18:00:29.920219 | orchestrator | # CHECK 2025-08-29 18:00:29.920224 | orchestrator | 2025-08-29 18:00:29.920228 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 18:00:29.920236 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 18:00:29.920240 | orchestrator | + echo 2025-08-29 18:00:29.920245 | orchestrator | + echo '# CHECK' 2025-08-29 18:00:29.920249 | orchestrator | + echo 2025-08-29 18:00:29.920269 | orchestrator | + for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2 2025-08-29 18:00:29.921096 | orchestrator | ++ semver 9.2.0 5.0.0 2025-08-29 18:00:29.981219 | orchestrator | 2025-08-29 18:00:29.981351 | orchestrator | ## Containers @ testbed-manager 2025-08-29 18:00:29.981367 | orchestrator | 2025-08-29 18:00:29.981381 | orchestrator | + [[ 1 -eq -1 ]] 2025-08-29 18:00:29.981392 | orchestrator | + echo 2025-08-29 18:00:29.981403 | orchestrator | + echo '## Containers @ testbed-manager' 2025-08-29 18:00:29.981414 | orchestrator | + echo 2025-08-29 18:00:29.981424 | orchestrator | + osism container testbed-manager ps 2025-08-29 18:00:32.330776 | orchestrator | CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 2025-08-29 18:00:32.330895 | orchestrator | 69df229ddffe registry.osism.tech/kolla/release/prometheus-blackbox-exporter:0.25.0.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_blackbox_exporter 2025-08-29 18:00:32.330910 | orchestrator | 1cb9fad55e72 registry.osism.tech/kolla/release/prometheus-alertmanager:0.28.0.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_alertmanager 2025-08-29 18:00:32.330918 | orchestrator | c45afb133e11 registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_cadvisor 2025-08-29 18:00:32.330930 | orchestrator | d609e2e1ac79 registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_node_exporter 2025-08-29 18:00:32.330936 | orchestrator | 4875f2bf0e59 registry.osism.tech/kolla/release/prometheus-v2-server:2.55.1.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_server 2025-08-29 18:00:32.330943 | orchestrator | f55079c7b9c8 registry.osism.tech/osism/cephclient:18.2.7 "/usr/bin/dumb-init …" 18 minutes ago Up 17 minutes cephclient 2025-08-29 18:00:32.330955 | orchestrator | 4bab0b87d95f registry.osism.tech/kolla/release/cron:3.0.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes cron 2025-08-29 18:00:32.330961 | orchestrator | 6884890f8f5d registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes kolla_toolbox 2025-08-29 18:00:32.330986 | orchestrator | aefde6d9d862 registry.osism.tech/kolla/release/fluentd:5.0.7.20250711 "dumb-init --single-…" 31 minutes ago Up 31 minutes fluentd 2025-08-29 18:00:32.330993 | orchestrator | dfcfd943e6e0 phpmyadmin/phpmyadmin:5.2 "/docker-entrypoint.…" 31 minutes ago Up 31 minutes (healthy) 80/tcp phpmyadmin 2025-08-29 18:00:32.330999 | orchestrator | f01b7bda2d8a registry.osism.tech/osism/openstackclient:2024.2 "/usr/bin/dumb-init …" 32 minutes ago Up 32 minutes openstackclient 2025-08-29 18:00:32.331006 | orchestrator | a86130fb336e registry.osism.tech/osism/homer:v25.05.2 "/bin/sh /entrypoint…" 32 minutes ago Up 32 minutes (healthy) 8080/tcp homer 2025-08-29 18:00:32.331012 | orchestrator | e0b7eb01cb4c registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta "entrypoint.sh -f /e…" 54 minutes ago Up 54 minutes (healthy) 192.168.16.5:3128->3128/tcp squid 2025-08-29 18:00:32.331022 | orchestrator | d273b835b89e registry.osism.tech/osism/inventory-reconciler:0.20250711.0 "/sbin/tini -- /entr…" 58 minutes ago Up 38 minutes (healthy) manager-inventory_reconciler-1 2025-08-29 18:00:32.331043 | orchestrator | ba31d756dfe6 registry.osism.tech/osism/kolla-ansible:0.20250711.0 "/entrypoint.sh osis…" 58 minutes ago Up 38 minutes (healthy) kolla-ansible 2025-08-29 18:00:32.331050 | orchestrator | e2f9dcc57a3b registry.osism.tech/osism/ceph-ansible:0.20250711.0 "/entrypoint.sh osis…" 58 minutes ago Up 38 minutes (healthy) ceph-ansible 2025-08-29 18:00:32.331057 | orchestrator | 50a93c79f6f9 registry.osism.tech/osism/osism-ansible:0.20250711.0 "/entrypoint.sh osis…" 58 minutes ago Up 38 minutes (healthy) osism-ansible 2025-08-29 18:00:32.331063 | orchestrator | 0481804f1e25 registry.osism.tech/osism/osism-kubernetes:0.20250711.0 "/entrypoint.sh osis…" 58 minutes ago Up 38 minutes (healthy) osism-kubernetes 2025-08-29 18:00:32.331070 | orchestrator | 54a34abd4d49 registry.osism.tech/osism/ara-server:1.7.2 "sh -c '/wait && /ru…" 58 minutes ago Up 39 minutes (healthy) 8000/tcp manager-ara-server-1 2025-08-29 18:00:32.331076 | orchestrator | ea1eabf01e50 registry.osism.tech/dockerhub/library/mariadb:11.8.2 "docker-entrypoint.s…" 58 minutes ago Up 39 minutes (healthy) 3306/tcp manager-mariadb-1 2025-08-29 18:00:32.331083 | orchestrator | 8b535108d298 registry.osism.tech/dockerhub/library/redis:7.4.5-alpine "docker-entrypoint.s…" 58 minutes ago Up 39 minutes (healthy) 6379/tcp manager-redis-1 2025-08-29 18:00:32.331089 | orchestrator | cfed5465d84d registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- sleep…" 58 minutes ago Up 39 minutes (healthy) osismclient 2025-08-29 18:00:32.331101 | orchestrator | f30be9ff4c39 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" 58 minutes ago Up 39 minutes (healthy) manager-flower-1 2025-08-29 18:00:32.331107 | orchestrator | c452cbeecce1 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" 58 minutes ago Up 39 minutes (healthy) manager-listener-1 2025-08-29 18:00:32.331114 | orchestrator | 12ad90536cda registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" 58 minutes ago Up 39 minutes (healthy) manager-openstack-1 2025-08-29 18:00:32.331120 | orchestrator | 1e1f6139193d registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" 58 minutes ago Up 39 minutes (healthy) manager-beat-1 2025-08-29 18:00:32.331127 | orchestrator | 41fc74335a73 registry.osism.tech/osism/osism:0.20250709.0 "/sbin/tini -- osism…" 58 minutes ago Up 39 minutes (healthy) 192.168.16.5:8000->8000/tcp manager-api-1 2025-08-29 18:00:32.331133 | orchestrator | 5dcec6084a1d registry.osism.tech/dockerhub/library/traefik:v3.4.3 "/entrypoint.sh trae…" About an hour ago Up About an hour (healthy) 192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp traefik 2025-08-29 18:00:32.639971 | orchestrator | 2025-08-29 18:00:32.640044 | orchestrator | ## Images @ testbed-manager 2025-08-29 18:00:32.640053 | orchestrator | 2025-08-29 18:00:32.640060 | orchestrator | + echo 2025-08-29 18:00:32.640067 | orchestrator | + echo '## Images @ testbed-manager' 2025-08-29 18:00:32.640074 | orchestrator | + echo 2025-08-29 18:00:32.640081 | orchestrator | + osism container testbed-manager images 2025-08-29 18:00:34.919191 | orchestrator | REPOSITORY TAG IMAGE ID CREATED SIZE 2025-08-29 18:00:34.919270 | orchestrator | registry.osism.tech/osism/openstackclient 2024.2 e303c4555969 10 hours ago 237MB 2025-08-29 18:00:34.919284 | orchestrator | registry.osism.tech/osism/homer v25.05.2 d3334946e20e 3 weeks ago 11.5MB 2025-08-29 18:00:34.919292 | orchestrator | registry.osism.tech/osism/kolla-ansible 0.20250711.0 fcbac8373342 6 weeks ago 571MB 2025-08-29 18:00:34.919298 | orchestrator | registry.osism.tech/kolla/release/fluentd 5.0.7.20250711 eaa70c1312aa 6 weeks ago 628MB 2025-08-29 18:00:34.919320 | orchestrator | registry.osism.tech/kolla/release/kolla-toolbox 19.5.1.20250711 ad526ea47263 6 weeks ago 746MB 2025-08-29 18:00:34.919327 | orchestrator | registry.osism.tech/kolla/release/cron 3.0.20250711 de0bd651bf89 6 weeks ago 318MB 2025-08-29 18:00:34.919334 | orchestrator | registry.osism.tech/kolla/release/prometheus-v2-server 2.55.1.20250711 cb02c47a5187 6 weeks ago 891MB 2025-08-29 18:00:34.919340 | orchestrator | registry.osism.tech/kolla/release/prometheus-blackbox-exporter 0.25.0.20250711 0ac8facfe451 6 weeks ago 360MB 2025-08-29 18:00:34.919347 | orchestrator | registry.osism.tech/kolla/release/prometheus-alertmanager 0.28.0.20250711 6c4eef6335f5 6 weeks ago 456MB 2025-08-29 18:00:34.919353 | orchestrator | registry.osism.tech/kolla/release/prometheus-cadvisor 0.49.2.20250711 937f4652a0d1 6 weeks ago 410MB 2025-08-29 18:00:34.919360 | orchestrator | registry.osism.tech/kolla/release/prometheus-node-exporter 1.8.2.20250711 361ce2873c65 6 weeks ago 358MB 2025-08-29 18:00:34.919366 | orchestrator | registry.osism.tech/osism/osism-ansible 0.20250711.0 7b0f9e78b4e4 6 weeks ago 575MB 2025-08-29 18:00:34.919387 | orchestrator | registry.osism.tech/osism/ceph-ansible 0.20250711.0 f677f8f8094b 6 weeks ago 535MB 2025-08-29 18:00:34.919394 | orchestrator | registry.osism.tech/osism/inventory-reconciler 0.20250711.0 8fcfa643b744 6 weeks ago 308MB 2025-08-29 18:00:34.919401 | orchestrator | registry.osism.tech/osism/osism-kubernetes 0.20250711.0 267f92fc46f6 6 weeks ago 1.21GB 2025-08-29 18:00:34.919407 | orchestrator | registry.osism.tech/osism/osism 0.20250709.0 ccd699d89870 7 weeks ago 310MB 2025-08-29 18:00:34.919413 | orchestrator | registry.osism.tech/dockerhub/library/redis 7.4.5-alpine f218e591b571 7 weeks ago 41.4MB 2025-08-29 18:00:34.919420 | orchestrator | registry.osism.tech/dockerhub/library/traefik v3.4.3 4113453efcb3 2 months ago 226MB 2025-08-29 18:00:34.919426 | orchestrator | registry.osism.tech/dockerhub/library/mariadb 11.8.2 dae0c92b7b63 2 months ago 329MB 2025-08-29 18:00:34.919433 | orchestrator | registry.osism.tech/osism/cephclient 18.2.7 ae977aa79826 3 months ago 453MB 2025-08-29 18:00:34.919439 | orchestrator | phpmyadmin/phpmyadmin 5.2 0276a66ce322 7 months ago 571MB 2025-08-29 18:00:34.919445 | orchestrator | registry.osism.tech/osism/ara-server 1.7.2 bb44122eb176 11 months ago 300MB 2025-08-29 18:00:34.919452 | orchestrator | registry.osism.tech/dockerhub/ubuntu/squid 6.1-23.10_beta 34b6bbbcf74b 14 months ago 146MB 2025-08-29 18:00:35.140007 | orchestrator | + for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2 2025-08-29 18:00:35.140470 | orchestrator | ++ semver 9.2.0 5.0.0 2025-08-29 18:00:35.199595 | orchestrator | 2025-08-29 18:00:35.199682 | orchestrator | ## Containers @ testbed-node-0 2025-08-29 18:00:35.199698 | orchestrator | 2025-08-29 18:00:35.199710 | orchestrator | + [[ 1 -eq -1 ]] 2025-08-29 18:00:35.199722 | orchestrator | + echo 2025-08-29 18:00:35.199734 | orchestrator | + echo '## Containers @ testbed-node-0' 2025-08-29 18:00:35.199746 | orchestrator | + echo 2025-08-29 18:00:35.199757 | orchestrator | + osism container testbed-node-0 ps 2025-08-29 18:00:37.255558 | orchestrator | CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 2025-08-29 18:00:37.255648 | orchestrator | 5beac4e6c47a registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711 "dumb-init --single-…" 7 minutes ago Up 7 minutes (healthy) nova_novncproxy 2025-08-29 18:00:37.255663 | orchestrator | fb5d14c12a61 registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711 "dumb-init --single-…" 7 minutes ago Up 7 minutes (healthy) nova_conductor 2025-08-29 18:00:37.255675 | orchestrator | 6af7bd864711 registry.osism.tech/kolla/release/nova-api:30.0.1.20250711 "dumb-init --single-…" 9 minutes ago Up 9 minutes (healthy) nova_api 2025-08-29 18:00:37.255687 | orchestrator | dbb7594e7ced registry.osism.tech/kolla/release/grafana:12.0.2.20250711 "dumb-init --single-…" 9 minutes ago Up 9 minutes grafana 2025-08-29 18:00:37.255698 | orchestrator | 537e9c04ed6e registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711 "dumb-init --single-…" 9 minutes ago Up 9 minutes (healthy) nova_scheduler 2025-08-29 18:00:37.255709 | orchestrator | 8fbe6f985468 registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) cinder_scheduler 2025-08-29 18:00:37.255720 | orchestrator | 854eafa2151e registry.osism.tech/kolla/release/glance-api:29.0.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) glance_api 2025-08-29 18:00:37.255731 | orchestrator | 3b35002c1176 registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) cinder_api 2025-08-29 18:00:37.255765 | orchestrator | 1515355e61fe registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_elasticsearch_exporter 2025-08-29 18:00:37.255790 | orchestrator | 67764d9b5d28 registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_cadvisor 2025-08-29 18:00:37.255801 | orchestrator | b695203e2151 registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711 "dumb-init --single-…" 13 minutes ago Up 12 minutes prometheus_memcached_exporter 2025-08-29 18:00:37.255812 | orchestrator | 591fb518541a registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_mysqld_exporter 2025-08-29 18:00:37.255823 | orchestrator | 7201adcb2f0b registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_node_exporter 2025-08-29 18:00:37.255834 | orchestrator | 33cdb78dda14 registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes (healthy) magnum_conductor 2025-08-29 18:00:37.255844 | orchestrator | 7b2ff8a73e3c registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711 "dumb-init --single-…" 14 minutes ago Up 14 minutes (healthy) magnum_api 2025-08-29 18:00:37.255855 | orchestrator | 620e962e3446 registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711 "dumb-init --single-…" 14 minutes ago Up 14 minutes (healthy) neutron_server 2025-08-29 18:00:37.255866 | orchestrator | 20253b2fa1d8 registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_worker 2025-08-29 18:00:37.255877 | orchestrator | fa029bf16914 registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_mdns 2025-08-29 18:00:37.255888 | orchestrator | 558543d873cc registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_producer 2025-08-29 18:00:37.255915 | orchestrator | 2573408d8073 registry.osism.tech/kolla/release/designate-central:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_central 2025-08-29 18:00:37.255927 | orchestrator | b8102072673f registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) barbican_worker 2025-08-29 18:00:37.255938 | orchestrator | 92649948baac registry.osism.tech/kolla/release/designate-api:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_api 2025-08-29 18:00:37.255949 | orchestrator | e63fa33ae0f3 registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) barbican_keystone_listener 2025-08-29 18:00:37.255960 | orchestrator | 2c362975fa73 registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_backend_bind9 2025-08-29 18:00:37.255971 | orchestrator | 9e94f2e19130 registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711 "dumb-init --single-…" 16 minutes ago Up 16 minutes (healthy) barbican_api 2025-08-29 18:00:37.255982 | orchestrator | 0824a91b08db registry.osism.tech/kolla/release/placement-api:12.0.1.20250711 "dumb-init --single-…" 16 minutes ago Up 16 minutes (healthy) placement_api 2025-08-29 18:00:37.256005 | orchestrator | 7170af53ccaf registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-mgr -…" 16 minutes ago Up 16 minutes ceph-mgr-testbed-node-0 2025-08-29 18:00:37.256017 | orchestrator | d249cc2ed427 registry.osism.tech/kolla/release/keystone:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone 2025-08-29 18:00:37.256028 | orchestrator | 587435ca6724 registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone_fernet 2025-08-29 18:00:37.256039 | orchestrator | c16a059b0933 registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone_ssh 2025-08-29 18:00:37.256054 | orchestrator | bcd5e4a5fdca registry.osism.tech/kolla/release/horizon:25.1.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) horizon 2025-08-29 18:00:37.256065 | orchestrator | f1438affd2ba registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711 "dumb-init -- kolla_…" 21 minutes ago Up 21 minutes (healthy) mariadb 2025-08-29 18:00:37.256076 | orchestrator | f3156684908a registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711 "dumb-init --single-…" 22 minutes ago Up 22 minutes (healthy) opensearch_dashboards 2025-08-29 18:00:37.256087 | orchestrator | 95857ff37792 registry.osism.tech/kolla/release/opensearch:2.19.2.20250711 "dumb-init --single-…" 23 minutes ago Up 23 minutes (healthy) opensearch 2025-08-29 18:00:37.256098 | orchestrator | 2f55a2fb9103 registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-crash" 23 minutes ago Up 23 minutes ceph-crash-testbed-node-0 2025-08-29 18:00:37.256109 | orchestrator | 4dc0256e485a registry.osism.tech/kolla/release/keepalived:2.2.7.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes keepalived 2025-08-29 18:00:37.256120 | orchestrator | 02e42c56fb30 registry.osism.tech/kolla/release/proxysql:2.7.3.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes (healthy) proxysql 2025-08-29 18:00:37.256131 | orchestrator | 14fb4a04c0c0 registry.osism.tech/kolla/release/haproxy:2.6.12.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes (healthy) haproxy 2025-08-29 18:00:37.256147 | orchestrator | 2f0d4f798d1f registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 27 minutes ovn_northd 2025-08-29 18:00:37.256158 | orchestrator | cac6506579d4 registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 27 minutes ovn_sb_db 2025-08-29 18:00:37.256176 | orchestrator | f720a1f8133a registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 27 minutes ovn_nb_db 2025-08-29 18:00:37.256187 | orchestrator | 0e5cc467e26a registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-mon -…" 28 minutes ago Up 28 minutes ceph-mon-testbed-node-0 2025-08-29 18:00:37.256199 | orchestrator | cf4873bf2139 registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711 "dumb-init --single-…" 28 minutes ago Up 28 minutes ovn_controller 2025-08-29 18:00:37.256209 | orchestrator | da8b2c7b390f registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) rabbitmq 2025-08-29 18:00:37.256227 | orchestrator | 24be2b9b9fbf registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) openvswitch_vswitchd 2025-08-29 18:00:37.256238 | orchestrator | d659ac1bff94 registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) openvswitch_db 2025-08-29 18:00:37.256249 | orchestrator | ffd3ade4708d registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) redis_sentinel 2025-08-29 18:00:37.256329 | orchestrator | 338ca444a1e0 registry.osism.tech/kolla/release/redis:7.0.15.20250711 "dumb-init --single-…" 30 minutes ago Up 29 minutes (healthy) redis 2025-08-29 18:00:37.256345 | orchestrator | 714697832d59 registry.osism.tech/kolla/release/memcached:1.6.18.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes (healthy) memcached 2025-08-29 18:00:37.256356 | orchestrator | 818717d98618 registry.osism.tech/kolla/release/cron:3.0.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes cron 2025-08-29 18:00:37.256367 | orchestrator | a077ccffbfc7 registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711 "dumb-init --single-…" 31 minutes ago Up 31 minutes kolla_toolbox 2025-08-29 18:00:37.256378 | orchestrator | dcf61f6960ee registry.osism.tech/kolla/release/fluentd:5.0.7.20250711 "dumb-init --single-…" 32 minutes ago Up 31 minutes fluentd 2025-08-29 18:00:37.464416 | orchestrator | 2025-08-29 18:00:37.464469 | orchestrator | ## Images @ testbed-node-0 2025-08-29 18:00:37.464475 | orchestrator | 2025-08-29 18:00:37.464479 | orchestrator | + echo 2025-08-29 18:00:37.464484 | orchestrator | + echo '## Images @ testbed-node-0' 2025-08-29 18:00:37.464488 | orchestrator | + echo 2025-08-29 18:00:37.464492 | orchestrator | + osism container testbed-node-0 images 2025-08-29 18:00:39.491325 | orchestrator | REPOSITORY TAG IMAGE ID CREATED SIZE 2025-08-29 18:00:39.491421 | orchestrator | registry.osism.tech/kolla/release/fluentd 5.0.7.20250711 eaa70c1312aa 6 weeks ago 628MB 2025-08-29 18:00:39.491436 | orchestrator | registry.osism.tech/kolla/release/keepalived 2.2.7.20250711 c7f6abdb2516 6 weeks ago 329MB 2025-08-29 18:00:39.491448 | orchestrator | registry.osism.tech/kolla/release/haproxy 2.6.12.20250711 0a9fd950fe86 6 weeks ago 326MB 2025-08-29 18:00:39.491459 | orchestrator | registry.osism.tech/kolla/release/opensearch 2.19.2.20250711 d8c44fac73c2 6 weeks ago 1.59GB 2025-08-29 18:00:39.491469 | orchestrator | registry.osism.tech/kolla/release/opensearch-dashboards 2.19.2.20250711 db87020f3b90 6 weeks ago 1.55GB 2025-08-29 18:00:39.491480 | orchestrator | registry.osism.tech/kolla/release/proxysql 2.7.3.20250711 4c6eaa052643 6 weeks ago 417MB 2025-08-29 18:00:39.491490 | orchestrator | registry.osism.tech/kolla/release/memcached 1.6.18.20250711 cd87896ace76 6 weeks ago 318MB 2025-08-29 18:00:39.491501 | orchestrator | registry.osism.tech/kolla/release/kolla-toolbox 19.5.1.20250711 ad526ea47263 6 weeks ago 746MB 2025-08-29 18:00:39.491512 | orchestrator | registry.osism.tech/kolla/release/rabbitmq 3.13.7.20250711 4ce47f209c9b 6 weeks ago 375MB 2025-08-29 18:00:39.491523 | orchestrator | registry.osism.tech/kolla/release/grafana 12.0.2.20250711 f4164dfd1b02 6 weeks ago 1.01GB 2025-08-29 18:00:39.491533 | orchestrator | registry.osism.tech/kolla/release/cron 3.0.20250711 de0bd651bf89 6 weeks ago 318MB 2025-08-29 18:00:39.491544 | orchestrator | registry.osism.tech/kolla/release/openvswitch-db-server 3.4.2.20250711 15f29551e6ce 6 weeks ago 361MB 2025-08-29 18:00:39.491577 | orchestrator | registry.osism.tech/kolla/release/openvswitch-vswitchd 3.4.2.20250711 ea9ea8f197d8 6 weeks ago 361MB 2025-08-29 18:00:39.491588 | orchestrator | registry.osism.tech/kolla/release/horizon 25.1.1.20250711 d4ae4a297d3b 6 weeks ago 1.21GB 2025-08-29 18:00:39.491599 | orchestrator | registry.osism.tech/kolla/release/prometheus-mysqld-exporter 0.16.0.20250711 142dafde994c 6 weeks ago 353MB 2025-08-29 18:00:39.491609 | orchestrator | registry.osism.tech/kolla/release/prometheus-cadvisor 0.49.2.20250711 937f4652a0d1 6 weeks ago 410MB 2025-08-29 18:00:39.491619 | orchestrator | registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter 1.8.0.20250711 62e13ec7689a 6 weeks ago 344MB 2025-08-29 18:00:39.491630 | orchestrator | registry.osism.tech/kolla/release/prometheus-node-exporter 1.8.2.20250711 361ce2873c65 6 weeks ago 358MB 2025-08-29 18:00:39.491641 | orchestrator | registry.osism.tech/kolla/release/prometheus-memcached-exporter 0.15.0.20250711 834c4c2dcd78 6 weeks ago 351MB 2025-08-29 18:00:39.491668 | orchestrator | registry.osism.tech/kolla/release/redis-sentinel 7.0.15.20250711 534f393a19e2 6 weeks ago 324MB 2025-08-29 18:00:39.491679 | orchestrator | registry.osism.tech/kolla/release/redis 7.0.15.20250711 d7d5c3586026 6 weeks ago 324MB 2025-08-29 18:00:39.491690 | orchestrator | registry.osism.tech/kolla/release/mariadb-server 10.11.13.20250711 5892b19e1064 6 weeks ago 590MB 2025-08-29 18:00:39.491701 | orchestrator | registry.osism.tech/kolla/release/ovn-controller 24.9.2.20250711 65e36d1176bd 6 weeks ago 947MB 2025-08-29 18:00:39.491711 | orchestrator | registry.osism.tech/kolla/release/ovn-sb-db-server 24.9.2.20250711 28654474dfe5 6 weeks ago 946MB 2025-08-29 18:00:39.491722 | orchestrator | registry.osism.tech/kolla/release/ovn-northd 24.9.2.20250711 58ad45688234 6 weeks ago 947MB 2025-08-29 18:00:39.491732 | orchestrator | registry.osism.tech/kolla/release/ovn-nb-db-server 24.9.2.20250711 affa47a97549 6 weeks ago 946MB 2025-08-29 18:00:39.491743 | orchestrator | registry.osism.tech/kolla/release/ceilometer-central 23.0.0.20250711 05a4552273f6 6 weeks ago 1.04GB 2025-08-29 18:00:39.491754 | orchestrator | registry.osism.tech/kolla/release/ceilometer-notification 23.0.0.20250711 41f8c34132c7 6 weeks ago 1.04GB 2025-08-29 18:00:39.491764 | orchestrator | registry.osism.tech/kolla/release/octavia-health-manager 15.0.1.20250711 06deffb77b4f 6 weeks ago 1.1GB 2025-08-29 18:00:39.491775 | orchestrator | registry.osism.tech/kolla/release/octavia-worker 15.0.1.20250711 02867223fb33 6 weeks ago 1.1GB 2025-08-29 18:00:39.491785 | orchestrator | registry.osism.tech/kolla/release/octavia-api 15.0.1.20250711 6146c08f2b76 6 weeks ago 1.12GB 2025-08-29 18:00:39.491814 | orchestrator | registry.osism.tech/kolla/release/octavia-housekeeping 15.0.1.20250711 6d529ee19c1c 6 weeks ago 1.1GB 2025-08-29 18:00:39.491825 | orchestrator | registry.osism.tech/kolla/release/octavia-driver-agent 15.0.1.20250711 b1ed239b634f 6 weeks ago 1.12GB 2025-08-29 18:00:39.491836 | orchestrator | registry.osism.tech/kolla/release/glance-api 29.0.1.20250711 65a4d0afbb1c 6 weeks ago 1.15GB 2025-08-29 18:00:39.491847 | orchestrator | registry.osism.tech/kolla/release/placement-api 12.0.1.20250711 2b6bd346ad18 6 weeks ago 1.04GB 2025-08-29 18:00:39.491857 | orchestrator | registry.osism.tech/kolla/release/barbican-api 19.0.1.20250711 1b7dd2682590 6 weeks ago 1.06GB 2025-08-29 18:00:39.491868 | orchestrator | registry.osism.tech/kolla/release/barbican-worker 19.0.1.20250711 e475391ce44d 6 weeks ago 1.06GB 2025-08-29 18:00:39.491878 | orchestrator | registry.osism.tech/kolla/release/barbican-keystone-listener 19.0.1.20250711 09290580fa03 6 weeks ago 1.06GB 2025-08-29 18:00:39.491889 | orchestrator | registry.osism.tech/kolla/release/cinder-scheduler 25.2.1.20250711 a09a8be1b711 6 weeks ago 1.41GB 2025-08-29 18:00:39.491907 | orchestrator | registry.osism.tech/kolla/release/cinder-api 25.2.1.20250711 c0d28e8febb9 6 weeks ago 1.41GB 2025-08-29 18:00:39.491918 | orchestrator | registry.osism.tech/kolla/release/nova-scheduler 30.0.1.20250711 e0ad0ae52bef 6 weeks ago 1.29GB 2025-08-29 18:00:39.491928 | orchestrator | registry.osism.tech/kolla/release/nova-novncproxy 30.0.1.20250711 b395cfe7f13f 6 weeks ago 1.42GB 2025-08-29 18:00:39.491939 | orchestrator | registry.osism.tech/kolla/release/nova-api 30.0.1.20250711 ee83c124eb76 6 weeks ago 1.29GB 2025-08-29 18:00:39.491949 | orchestrator | registry.osism.tech/kolla/release/nova-conductor 30.0.1.20250711 44e25b162470 6 weeks ago 1.29GB 2025-08-29 18:00:39.491965 | orchestrator | registry.osism.tech/kolla/release/magnum-api 19.0.1.20250711 71f47d2b2def 6 weeks ago 1.2GB 2025-08-29 18:00:39.491976 | orchestrator | registry.osism.tech/kolla/release/magnum-conductor 19.0.1.20250711 13b61cb4a5d2 6 weeks ago 1.31GB 2025-08-29 18:00:39.491987 | orchestrator | registry.osism.tech/kolla/release/designate-mdns 19.0.1.20250711 a030b794eaa9 6 weeks ago 1.05GB 2025-08-29 18:00:39.491997 | orchestrator | registry.osism.tech/kolla/release/designate-producer 19.0.1.20250711 2d0954c30848 6 weeks ago 1.05GB 2025-08-29 18:00:39.492008 | orchestrator | registry.osism.tech/kolla/release/designate-central 19.0.1.20250711 f7fa0bcabe47 6 weeks ago 1.05GB 2025-08-29 18:00:39.492018 | orchestrator | registry.osism.tech/kolla/release/designate-worker 19.0.1.20250711 4de726ebba0e 6 weeks ago 1.06GB 2025-08-29 18:00:39.492029 | orchestrator | registry.osism.tech/kolla/release/designate-backend-bind9 19.0.1.20250711 a14c6ace0b24 6 weeks ago 1.06GB 2025-08-29 18:00:39.492040 | orchestrator | registry.osism.tech/kolla/release/designate-api 19.0.1.20250711 2a2b32cdb83f 6 weeks ago 1.05GB 2025-08-29 18:00:39.492050 | orchestrator | registry.osism.tech/kolla/release/skyline-console 5.0.1.20250711 f2e37439c6b7 6 weeks ago 1.11GB 2025-08-29 18:00:39.492061 | orchestrator | registry.osism.tech/kolla/release/skyline-apiserver 5.0.1.20250711 b3d19c53d4de 6 weeks ago 1.11GB 2025-08-29 18:00:39.492071 | orchestrator | registry.osism.tech/kolla/release/keystone-fernet 26.0.1.20250711 53889b0cb73d 6 weeks ago 1.11GB 2025-08-29 18:00:39.492082 | orchestrator | registry.osism.tech/kolla/release/keystone 26.0.1.20250711 caf4f12b4799 6 weeks ago 1.13GB 2025-08-29 18:00:39.492093 | orchestrator | registry.osism.tech/kolla/release/keystone-ssh 26.0.1.20250711 3ba6da1abaea 6 weeks ago 1.11GB 2025-08-29 18:00:39.492103 | orchestrator | registry.osism.tech/kolla/release/neutron-server 25.2.1.20250711 8377b7d24f73 6 weeks ago 1.24GB 2025-08-29 18:00:39.492114 | orchestrator | registry.osism.tech/kolla/release/aodh-evaluator 19.0.0.20250711 c26d685bbc69 6 weeks ago 1.04GB 2025-08-29 18:00:39.492124 | orchestrator | registry.osism.tech/kolla/release/aodh-listener 19.0.0.20250711 55a7448b63ad 6 weeks ago 1.04GB 2025-08-29 18:00:39.492135 | orchestrator | registry.osism.tech/kolla/release/aodh-api 19.0.0.20250711 b8a4d60cb725 6 weeks ago 1.04GB 2025-08-29 18:00:39.492146 | orchestrator | registry.osism.tech/kolla/release/aodh-notifier 19.0.0.20250711 c0822bfcb81c 6 weeks ago 1.04GB 2025-08-29 18:00:39.492156 | orchestrator | registry.osism.tech/osism/ceph-daemon 18.2.7 5f92363b1f93 3 months ago 1.27GB 2025-08-29 18:00:39.689909 | orchestrator | + for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2 2025-08-29 18:00:39.690538 | orchestrator | ++ semver 9.2.0 5.0.0 2025-08-29 18:00:39.732049 | orchestrator | 2025-08-29 18:00:39.732131 | orchestrator | ## Containers @ testbed-node-1 2025-08-29 18:00:39.732169 | orchestrator | 2025-08-29 18:00:39.732181 | orchestrator | + [[ 1 -eq -1 ]] 2025-08-29 18:00:39.732193 | orchestrator | + echo 2025-08-29 18:00:39.732206 | orchestrator | + echo '## Containers @ testbed-node-1' 2025-08-29 18:00:39.732218 | orchestrator | + echo 2025-08-29 18:00:39.732228 | orchestrator | + osism container testbed-node-1 ps 2025-08-29 18:00:41.770738 | orchestrator | CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 2025-08-29 18:00:41.770808 | orchestrator | 63c96fc509ac registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711 "dumb-init --single-…" 7 minutes ago Up 7 minutes (healthy) nova_novncproxy 2025-08-29 18:00:41.770818 | orchestrator | e1d7574255a5 registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711 "dumb-init --single-…" 7 minutes ago Up 7 minutes (healthy) nova_conductor 2025-08-29 18:00:41.770827 | orchestrator | 29432452ac3c registry.osism.tech/kolla/release/grafana:12.0.2.20250711 "dumb-init --single-…" 8 minutes ago Up 8 minutes grafana 2025-08-29 18:00:41.770834 | orchestrator | bc8da13c8f9a registry.osism.tech/kolla/release/nova-api:30.0.1.20250711 "dumb-init --single-…" 9 minutes ago Up 9 minutes (healthy) nova_api 2025-08-29 18:00:41.770854 | orchestrator | 80ab64e8cf59 registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711 "dumb-init --single-…" 9 minutes ago Up 9 minutes (healthy) nova_scheduler 2025-08-29 18:00:41.770863 | orchestrator | 81bcac17c1ed registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) cinder_scheduler 2025-08-29 18:00:41.770870 | orchestrator | b40c07f9de68 registry.osism.tech/kolla/release/glance-api:29.0.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) glance_api 2025-08-29 18:00:41.770878 | orchestrator | 6a4652881094 registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) cinder_api 2025-08-29 18:00:41.770886 | orchestrator | cefe42d8eac1 registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_elasticsearch_exporter 2025-08-29 18:00:41.770895 | orchestrator | 9a4ea8aebb5e registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_cadvisor 2025-08-29 18:00:41.770903 | orchestrator | 3d606b4e64a0 registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_memcached_exporter 2025-08-29 18:00:41.770911 | orchestrator | 82d7ee2f6b90 registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_mysqld_exporter 2025-08-29 18:00:41.770918 | orchestrator | 3c0389397b89 registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_node_exporter 2025-08-29 18:00:41.770926 | orchestrator | 433c472810e0 registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes (healthy) magnum_conductor 2025-08-29 18:00:41.770933 | orchestrator | e98ce41f0cb2 registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711 "dumb-init --single-…" 14 minutes ago Up 14 minutes (healthy) magnum_api 2025-08-29 18:00:41.770941 | orchestrator | cccccbd3058e registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711 "dumb-init --single-…" 14 minutes ago Up 14 minutes (healthy) neutron_server 2025-08-29 18:00:41.770963 | orchestrator | 7128089253fa registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_worker 2025-08-29 18:00:41.770971 | orchestrator | db661202b5ff registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_mdns 2025-08-29 18:00:41.770998 | orchestrator | fe351a41cfba registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_producer 2025-08-29 18:00:41.771017 | orchestrator | 616a18de0a7c registry.osism.tech/kolla/release/designate-central:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_central 2025-08-29 18:00:41.771025 | orchestrator | a6113706c2e0 registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) barbican_worker 2025-08-29 18:00:41.771032 | orchestrator | 1841eeb84d17 registry.osism.tech/kolla/release/designate-api:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_api 2025-08-29 18:00:41.771039 | orchestrator | 98fc0b5e9e07 registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) barbican_keystone_listener 2025-08-29 18:00:41.771050 | orchestrator | 5a4125d7f97c registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_backend_bind9 2025-08-29 18:00:41.771061 | orchestrator | bdb7c7dc499d registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711 "dumb-init --single-…" 16 minutes ago Up 16 minutes (healthy) barbican_api 2025-08-29 18:00:41.771068 | orchestrator | 46245eaa59ec registry.osism.tech/kolla/release/placement-api:12.0.1.20250711 "dumb-init --single-…" 16 minutes ago Up 16 minutes (healthy) placement_api 2025-08-29 18:00:41.771076 | orchestrator | c5745a8e0a7a registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-mgr -…" 16 minutes ago Up 16 minutes ceph-mgr-testbed-node-1 2025-08-29 18:00:41.771083 | orchestrator | 786dcbe48989 registry.osism.tech/kolla/release/keystone:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone 2025-08-29 18:00:41.771090 | orchestrator | 976d1ce48572 registry.osism.tech/kolla/release/horizon:25.1.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) horizon 2025-08-29 18:00:41.771194 | orchestrator | 87669d0b9611 registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone_fernet 2025-08-29 18:00:41.771205 | orchestrator | 4641daa2646e registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone_ssh 2025-08-29 18:00:41.771212 | orchestrator | 8e4b4c7b57a9 registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711 "dumb-init --single-…" 21 minutes ago Up 21 minutes (healthy) opensearch_dashboards 2025-08-29 18:00:41.771219 | orchestrator | 248e72274246 registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711 "dumb-init -- kolla_…" 22 minutes ago Up 22 minutes (healthy) mariadb 2025-08-29 18:00:41.771226 | orchestrator | f5521884c461 registry.osism.tech/kolla/release/opensearch:2.19.2.20250711 "dumb-init --single-…" 22 minutes ago Up 22 minutes (healthy) opensearch 2025-08-29 18:00:41.771240 | orchestrator | abcff14f1264 registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-crash" 23 minutes ago Up 23 minutes ceph-crash-testbed-node-1 2025-08-29 18:00:41.771247 | orchestrator | c2546a32eae8 registry.osism.tech/kolla/release/keepalived:2.2.7.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes keepalived 2025-08-29 18:00:41.771254 | orchestrator | 5182e721e1e1 registry.osism.tech/kolla/release/proxysql:2.7.3.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes (healthy) proxysql 2025-08-29 18:00:41.771285 | orchestrator | 53a612fbe034 registry.osism.tech/kolla/release/haproxy:2.6.12.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes (healthy) haproxy 2025-08-29 18:00:41.771293 | orchestrator | bebbcea458fe registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 26 minutes ovn_northd 2025-08-29 18:00:41.771301 | orchestrator | 49591f4085f0 registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 26 minutes ovn_sb_db 2025-08-29 18:00:41.771308 | orchestrator | f1232a9a1b0f registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 26 minutes ovn_nb_db 2025-08-29 18:00:41.771315 | orchestrator | eb7855ed62d9 registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711 "dumb-init --single-…" 28 minutes ago Up 28 minutes ovn_controller 2025-08-29 18:00:41.771322 | orchestrator | c118f3cd6fd8 registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711 "dumb-init --single-…" 28 minutes ago Up 28 minutes (healthy) rabbitmq 2025-08-29 18:00:41.771330 | orchestrator | 9d107394c3ff registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-mon -…" 28 minutes ago Up 28 minutes ceph-mon-testbed-node-1 2025-08-29 18:00:41.771337 | orchestrator | 2618c9de3eb6 registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) openvswitch_vswitchd 2025-08-29 18:00:41.771345 | orchestrator | 3bd7e0e9d2b1 registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) openvswitch_db 2025-08-29 18:00:41.771352 | orchestrator | d2310e5c2725 registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) redis_sentinel 2025-08-29 18:00:41.771359 | orchestrator | 37df45bfb21d registry.osism.tech/kolla/release/redis:7.0.15.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes (healthy) redis 2025-08-29 18:00:41.771370 | orchestrator | 85cc5cf68376 registry.osism.tech/kolla/release/memcached:1.6.18.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes (healthy) memcached 2025-08-29 18:00:41.771378 | orchestrator | 7aca55f71a28 registry.osism.tech/kolla/release/cron:3.0.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes cron 2025-08-29 18:00:41.771389 | orchestrator | a699691be21e registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes kolla_toolbox 2025-08-29 18:00:41.771397 | orchestrator | 75bc082d46da registry.osism.tech/kolla/release/fluentd:5.0.7.20250711 "dumb-init --single-…" 31 minutes ago Up 31 minutes fluentd 2025-08-29 18:00:41.973664 | orchestrator | 2025-08-29 18:00:41.973735 | orchestrator | ## Images @ testbed-node-1 2025-08-29 18:00:41.973747 | orchestrator | 2025-08-29 18:00:41.973773 | orchestrator | + echo 2025-08-29 18:00:41.973783 | orchestrator | + echo '## Images @ testbed-node-1' 2025-08-29 18:00:41.973793 | orchestrator | + echo 2025-08-29 18:00:41.973802 | orchestrator | + osism container testbed-node-1 images 2025-08-29 18:00:44.195703 | orchestrator | REPOSITORY TAG IMAGE ID CREATED SIZE 2025-08-29 18:00:44.195791 | orchestrator | registry.osism.tech/kolla/release/fluentd 5.0.7.20250711 eaa70c1312aa 6 weeks ago 628MB 2025-08-29 18:00:44.195799 | orchestrator | registry.osism.tech/kolla/release/keepalived 2.2.7.20250711 c7f6abdb2516 6 weeks ago 329MB 2025-08-29 18:00:44.195807 | orchestrator | registry.osism.tech/kolla/release/haproxy 2.6.12.20250711 0a9fd950fe86 6 weeks ago 326MB 2025-08-29 18:00:44.195814 | orchestrator | registry.osism.tech/kolla/release/opensearch 2.19.2.20250711 d8c44fac73c2 6 weeks ago 1.59GB 2025-08-29 18:00:44.195821 | orchestrator | registry.osism.tech/kolla/release/opensearch-dashboards 2.19.2.20250711 db87020f3b90 6 weeks ago 1.55GB 2025-08-29 18:00:44.195827 | orchestrator | registry.osism.tech/kolla/release/proxysql 2.7.3.20250711 4c6eaa052643 6 weeks ago 417MB 2025-08-29 18:00:44.195834 | orchestrator | registry.osism.tech/kolla/release/memcached 1.6.18.20250711 cd87896ace76 6 weeks ago 318MB 2025-08-29 18:00:44.195841 | orchestrator | registry.osism.tech/kolla/release/rabbitmq 3.13.7.20250711 4ce47f209c9b 6 weeks ago 375MB 2025-08-29 18:00:44.195847 | orchestrator | registry.osism.tech/kolla/release/kolla-toolbox 19.5.1.20250711 ad526ea47263 6 weeks ago 746MB 2025-08-29 18:00:44.195854 | orchestrator | registry.osism.tech/kolla/release/grafana 12.0.2.20250711 f4164dfd1b02 6 weeks ago 1.01GB 2025-08-29 18:00:44.195861 | orchestrator | registry.osism.tech/kolla/release/cron 3.0.20250711 de0bd651bf89 6 weeks ago 318MB 2025-08-29 18:00:44.195867 | orchestrator | registry.osism.tech/kolla/release/openvswitch-db-server 3.4.2.20250711 15f29551e6ce 6 weeks ago 361MB 2025-08-29 18:00:44.195874 | orchestrator | registry.osism.tech/kolla/release/openvswitch-vswitchd 3.4.2.20250711 ea9ea8f197d8 6 weeks ago 361MB 2025-08-29 18:00:44.195880 | orchestrator | registry.osism.tech/kolla/release/horizon 25.1.1.20250711 d4ae4a297d3b 6 weeks ago 1.21GB 2025-08-29 18:00:44.195887 | orchestrator | registry.osism.tech/kolla/release/prometheus-mysqld-exporter 0.16.0.20250711 142dafde994c 6 weeks ago 353MB 2025-08-29 18:00:44.195894 | orchestrator | registry.osism.tech/kolla/release/prometheus-cadvisor 0.49.2.20250711 937f4652a0d1 6 weeks ago 410MB 2025-08-29 18:00:44.195900 | orchestrator | registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter 1.8.0.20250711 62e13ec7689a 6 weeks ago 344MB 2025-08-29 18:00:44.195907 | orchestrator | registry.osism.tech/kolla/release/prometheus-node-exporter 1.8.2.20250711 361ce2873c65 6 weeks ago 358MB 2025-08-29 18:00:44.195913 | orchestrator | registry.osism.tech/kolla/release/prometheus-memcached-exporter 0.15.0.20250711 834c4c2dcd78 6 weeks ago 351MB 2025-08-29 18:00:44.195920 | orchestrator | registry.osism.tech/kolla/release/redis-sentinel 7.0.15.20250711 534f393a19e2 6 weeks ago 324MB 2025-08-29 18:00:44.195926 | orchestrator | registry.osism.tech/kolla/release/redis 7.0.15.20250711 d7d5c3586026 6 weeks ago 324MB 2025-08-29 18:00:44.195933 | orchestrator | registry.osism.tech/kolla/release/mariadb-server 10.11.13.20250711 5892b19e1064 6 weeks ago 590MB 2025-08-29 18:00:44.195940 | orchestrator | registry.osism.tech/kolla/release/ovn-controller 24.9.2.20250711 65e36d1176bd 6 weeks ago 947MB 2025-08-29 18:00:44.195946 | orchestrator | registry.osism.tech/kolla/release/ovn-sb-db-server 24.9.2.20250711 28654474dfe5 6 weeks ago 946MB 2025-08-29 18:00:44.195972 | orchestrator | registry.osism.tech/kolla/release/ovn-northd 24.9.2.20250711 58ad45688234 6 weeks ago 947MB 2025-08-29 18:00:44.195978 | orchestrator | registry.osism.tech/kolla/release/ovn-nb-db-server 24.9.2.20250711 affa47a97549 6 weeks ago 946MB 2025-08-29 18:00:44.195985 | orchestrator | registry.osism.tech/kolla/release/glance-api 29.0.1.20250711 65a4d0afbb1c 6 weeks ago 1.15GB 2025-08-29 18:00:44.195991 | orchestrator | registry.osism.tech/kolla/release/placement-api 12.0.1.20250711 2b6bd346ad18 6 weeks ago 1.04GB 2025-08-29 18:00:44.195998 | orchestrator | registry.osism.tech/kolla/release/barbican-api 19.0.1.20250711 1b7dd2682590 6 weeks ago 1.06GB 2025-08-29 18:00:44.196004 | orchestrator | registry.osism.tech/kolla/release/barbican-worker 19.0.1.20250711 e475391ce44d 6 weeks ago 1.06GB 2025-08-29 18:00:44.196011 | orchestrator | registry.osism.tech/kolla/release/barbican-keystone-listener 19.0.1.20250711 09290580fa03 6 weeks ago 1.06GB 2025-08-29 18:00:44.196030 | orchestrator | registry.osism.tech/kolla/release/cinder-scheduler 25.2.1.20250711 a09a8be1b711 6 weeks ago 1.41GB 2025-08-29 18:00:44.196036 | orchestrator | registry.osism.tech/kolla/release/cinder-api 25.2.1.20250711 c0d28e8febb9 6 weeks ago 1.41GB 2025-08-29 18:00:44.196057 | orchestrator | registry.osism.tech/kolla/release/nova-scheduler 30.0.1.20250711 e0ad0ae52bef 6 weeks ago 1.29GB 2025-08-29 18:00:44.196064 | orchestrator | registry.osism.tech/kolla/release/nova-novncproxy 30.0.1.20250711 b395cfe7f13f 6 weeks ago 1.42GB 2025-08-29 18:00:44.196070 | orchestrator | registry.osism.tech/kolla/release/nova-api 30.0.1.20250711 ee83c124eb76 6 weeks ago 1.29GB 2025-08-29 18:00:44.196077 | orchestrator | registry.osism.tech/kolla/release/nova-conductor 30.0.1.20250711 44e25b162470 6 weeks ago 1.29GB 2025-08-29 18:00:44.196083 | orchestrator | registry.osism.tech/kolla/release/magnum-api 19.0.1.20250711 71f47d2b2def 6 weeks ago 1.2GB 2025-08-29 18:00:44.196089 | orchestrator | registry.osism.tech/kolla/release/magnum-conductor 19.0.1.20250711 13b61cb4a5d2 6 weeks ago 1.31GB 2025-08-29 18:00:44.196098 | orchestrator | registry.osism.tech/kolla/release/designate-mdns 19.0.1.20250711 a030b794eaa9 6 weeks ago 1.05GB 2025-08-29 18:00:44.196105 | orchestrator | registry.osism.tech/kolla/release/designate-producer 19.0.1.20250711 2d0954c30848 6 weeks ago 1.05GB 2025-08-29 18:00:44.196111 | orchestrator | registry.osism.tech/kolla/release/designate-central 19.0.1.20250711 f7fa0bcabe47 6 weeks ago 1.05GB 2025-08-29 18:00:44.196118 | orchestrator | registry.osism.tech/kolla/release/designate-worker 19.0.1.20250711 4de726ebba0e 6 weeks ago 1.06GB 2025-08-29 18:00:44.196124 | orchestrator | registry.osism.tech/kolla/release/designate-backend-bind9 19.0.1.20250711 a14c6ace0b24 6 weeks ago 1.06GB 2025-08-29 18:00:44.196130 | orchestrator | registry.osism.tech/kolla/release/designate-api 19.0.1.20250711 2a2b32cdb83f 6 weeks ago 1.05GB 2025-08-29 18:00:44.196136 | orchestrator | registry.osism.tech/kolla/release/keystone-fernet 26.0.1.20250711 53889b0cb73d 6 weeks ago 1.11GB 2025-08-29 18:00:44.196142 | orchestrator | registry.osism.tech/kolla/release/keystone 26.0.1.20250711 caf4f12b4799 6 weeks ago 1.13GB 2025-08-29 18:00:44.196148 | orchestrator | registry.osism.tech/kolla/release/keystone-ssh 26.0.1.20250711 3ba6da1abaea 6 weeks ago 1.11GB 2025-08-29 18:00:44.196154 | orchestrator | registry.osism.tech/kolla/release/neutron-server 25.2.1.20250711 8377b7d24f73 6 weeks ago 1.24GB 2025-08-29 18:00:44.196160 | orchestrator | registry.osism.tech/osism/ceph-daemon 18.2.7 5f92363b1f93 3 months ago 1.27GB 2025-08-29 18:00:44.496687 | orchestrator | + for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2 2025-08-29 18:00:44.497571 | orchestrator | ++ semver 9.2.0 5.0.0 2025-08-29 18:00:44.540167 | orchestrator | 2025-08-29 18:00:44.540317 | orchestrator | ## Containers @ testbed-node-2 2025-08-29 18:00:44.540335 | orchestrator | 2025-08-29 18:00:44.540347 | orchestrator | + [[ 1 -eq -1 ]] 2025-08-29 18:00:44.540359 | orchestrator | + echo 2025-08-29 18:00:44.540371 | orchestrator | + echo '## Containers @ testbed-node-2' 2025-08-29 18:00:44.540383 | orchestrator | + echo 2025-08-29 18:00:44.540394 | orchestrator | + osism container testbed-node-2 ps 2025-08-29 18:00:46.743639 | orchestrator | CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES 2025-08-29 18:00:46.743748 | orchestrator | 176e34546bf4 registry.osism.tech/kolla/release/nova-novncproxy:30.0.1.20250711 "dumb-init --single-…" 7 minutes ago Up 7 minutes (healthy) nova_novncproxy 2025-08-29 18:00:46.743764 | orchestrator | 531f37bdd2f9 registry.osism.tech/kolla/release/nova-conductor:30.0.1.20250711 "dumb-init --single-…" 7 minutes ago Up 7 minutes (healthy) nova_conductor 2025-08-29 18:00:46.743776 | orchestrator | 1692ebeac793 registry.osism.tech/kolla/release/grafana:12.0.2.20250711 "dumb-init --single-…" 8 minutes ago Up 8 minutes grafana 2025-08-29 18:00:46.743787 | orchestrator | ce452fa5dc10 registry.osism.tech/kolla/release/nova-api:30.0.1.20250711 "dumb-init --single-…" 9 minutes ago Up 9 minutes (healthy) nova_api 2025-08-29 18:00:46.743798 | orchestrator | 04dc53ff502e registry.osism.tech/kolla/release/nova-scheduler:30.0.1.20250711 "dumb-init --single-…" 9 minutes ago Up 9 minutes (healthy) nova_scheduler 2025-08-29 18:00:46.743809 | orchestrator | de0cc58e4b21 registry.osism.tech/kolla/release/cinder-scheduler:25.2.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) cinder_scheduler 2025-08-29 18:00:46.743820 | orchestrator | 09790030445a registry.osism.tech/kolla/release/cinder-api:25.2.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) cinder_api 2025-08-29 18:00:46.743831 | orchestrator | d22a4284508a registry.osism.tech/kolla/release/glance-api:29.0.1.20250711 "dumb-init --single-…" 11 minutes ago Up 11 minutes (healthy) glance_api 2025-08-29 18:00:46.743842 | orchestrator | d055547db3d5 registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter:1.8.0.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_elasticsearch_exporter 2025-08-29 18:00:46.743855 | orchestrator | 4ac56aed0fe5 registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711 "dumb-init --single-…" 12 minutes ago Up 12 minutes prometheus_cadvisor 2025-08-29 18:00:46.743866 | orchestrator | 113a71729a9b registry.osism.tech/kolla/release/prometheus-memcached-exporter:0.15.0.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_memcached_exporter 2025-08-29 18:00:46.743877 | orchestrator | b61aa21dd698 registry.osism.tech/kolla/release/prometheus-mysqld-exporter:0.16.0.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_mysqld_exporter 2025-08-29 18:00:46.743888 | orchestrator | 2cc3f3345a0b registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes prometheus_node_exporter 2025-08-29 18:00:46.743899 | orchestrator | bd04308d4d12 registry.osism.tech/kolla/release/magnum-conductor:19.0.1.20250711 "dumb-init --single-…" 13 minutes ago Up 13 minutes (healthy) magnum_conductor 2025-08-29 18:00:46.743910 | orchestrator | 11bfd1e10270 registry.osism.tech/kolla/release/magnum-api:19.0.1.20250711 "dumb-init --single-…" 14 minutes ago Up 14 minutes (healthy) magnum_api 2025-08-29 18:00:46.743991 | orchestrator | ec46b0e12ac4 registry.osism.tech/kolla/release/neutron-server:25.2.1.20250711 "dumb-init --single-…" 14 minutes ago Up 14 minutes (healthy) neutron_server 2025-08-29 18:00:46.744021 | orchestrator | 9e43bf39571d registry.osism.tech/kolla/release/designate-worker:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_worker 2025-08-29 18:00:46.744033 | orchestrator | 955f9d404b30 registry.osism.tech/kolla/release/designate-mdns:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_mdns 2025-08-29 18:00:46.744044 | orchestrator | 91bfc0876061 registry.osism.tech/kolla/release/designate-producer:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_producer 2025-08-29 18:00:46.744072 | orchestrator | c2e990b1aa76 registry.osism.tech/kolla/release/designate-central:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_central 2025-08-29 18:00:46.744084 | orchestrator | 305cbbf3318a registry.osism.tech/kolla/release/barbican-worker:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) barbican_worker 2025-08-29 18:00:46.744095 | orchestrator | 1d21be31ce93 registry.osism.tech/kolla/release/designate-api:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) designate_api 2025-08-29 18:00:46.744106 | orchestrator | d44ffefd21a5 registry.osism.tech/kolla/release/barbican-keystone-listener:19.0.1.20250711 "dumb-init --single-…" 15 minutes ago Up 15 minutes (healthy) barbican_keystone_listener 2025-08-29 18:00:46.744117 | orchestrator | 9de56d080a69 registry.osism.tech/kolla/release/designate-backend-bind9:19.0.1.20250711 "dumb-init --single-…" 16 minutes ago Up 16 minutes (healthy) designate_backend_bind9 2025-08-29 18:00:46.744127 | orchestrator | 8b848bd7046b registry.osism.tech/kolla/release/barbican-api:19.0.1.20250711 "dumb-init --single-…" 16 minutes ago Up 16 minutes (healthy) barbican_api 2025-08-29 18:00:46.744141 | orchestrator | 212162559c54 registry.osism.tech/kolla/release/placement-api:12.0.1.20250711 "dumb-init --single-…" 16 minutes ago Up 16 minutes (healthy) placement_api 2025-08-29 18:00:46.744160 | orchestrator | d5f03da06452 registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-mgr -…" 16 minutes ago Up 16 minutes ceph-mgr-testbed-node-2 2025-08-29 18:00:46.744179 | orchestrator | 2f0fb5c0ea9e registry.osism.tech/kolla/release/keystone:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone 2025-08-29 18:00:46.744199 | orchestrator | 0faecb2ea6a6 registry.osism.tech/kolla/release/horizon:25.1.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) horizon 2025-08-29 18:00:46.744219 | orchestrator | 1bbfb304f213 registry.osism.tech/kolla/release/keystone-fernet:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone_fernet 2025-08-29 18:00:46.744239 | orchestrator | 989ea699d8f0 registry.osism.tech/kolla/release/keystone-ssh:26.0.1.20250711 "dumb-init --single-…" 19 minutes ago Up 19 minutes (healthy) keystone_ssh 2025-08-29 18:00:46.744288 | orchestrator | c301abc50d55 registry.osism.tech/kolla/release/opensearch-dashboards:2.19.2.20250711 "dumb-init --single-…" 21 minutes ago Up 21 minutes (healthy) opensearch_dashboards 2025-08-29 18:00:46.744311 | orchestrator | d902d892635d registry.osism.tech/kolla/release/mariadb-server:10.11.13.20250711 "dumb-init -- kolla_…" 21 minutes ago Up 21 minutes (healthy) mariadb 2025-08-29 18:00:46.744325 | orchestrator | 8f7b1c8ab71a registry.osism.tech/kolla/release/opensearch:2.19.2.20250711 "dumb-init --single-…" 22 minutes ago Up 22 minutes (healthy) opensearch 2025-08-29 18:00:46.744349 | orchestrator | a642b706c538 registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-crash" 23 minutes ago Up 23 minutes ceph-crash-testbed-node-2 2025-08-29 18:00:46.744362 | orchestrator | e7cfd72c50b1 registry.osism.tech/kolla/release/keepalived:2.2.7.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes keepalived 2025-08-29 18:00:46.744374 | orchestrator | fa0b2828b251 registry.osism.tech/kolla/release/proxysql:2.7.3.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes (healthy) proxysql 2025-08-29 18:00:46.744387 | orchestrator | 9e304c2d5e49 registry.osism.tech/kolla/release/haproxy:2.6.12.20250711 "dumb-init --single-…" 24 minutes ago Up 24 minutes (healthy) haproxy 2025-08-29 18:00:46.744399 | orchestrator | 2056527cb438 registry.osism.tech/kolla/release/ovn-northd:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 26 minutes ovn_northd 2025-08-29 18:00:46.744411 | orchestrator | ed3973971cbd registry.osism.tech/kolla/release/ovn-sb-db-server:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 26 minutes ovn_sb_db 2025-08-29 18:00:46.744432 | orchestrator | 5a7e5c042fe2 registry.osism.tech/kolla/release/ovn-nb-db-server:24.9.2.20250711 "dumb-init --single-…" 27 minutes ago Up 27 minutes ovn_nb_db 2025-08-29 18:00:46.744445 | orchestrator | abb40394d5d4 registry.osism.tech/kolla/release/rabbitmq:3.13.7.20250711 "dumb-init --single-…" 28 minutes ago Up 28 minutes (healthy) rabbitmq 2025-08-29 18:00:46.744457 | orchestrator | fd565fa089c2 registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711 "dumb-init --single-…" 28 minutes ago Up 28 minutes ovn_controller 2025-08-29 18:00:46.744470 | orchestrator | bc3d2c50e18a registry.osism.tech/osism/ceph-daemon:18.2.7 "/usr/bin/ceph-mon -…" 28 minutes ago Up 28 minutes ceph-mon-testbed-node-2 2025-08-29 18:00:46.744483 | orchestrator | a7aa8504f1f6 registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) openvswitch_vswitchd 2025-08-29 18:00:46.744495 | orchestrator | 8549a9665cd0 registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711 "dumb-init --single-…" 29 minutes ago Up 29 minutes (healthy) openvswitch_db 2025-08-29 18:00:46.744508 | orchestrator | 8b30fa530b28 registry.osism.tech/kolla/release/redis-sentinel:7.0.15.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes (healthy) redis_sentinel 2025-08-29 18:00:46.744520 | orchestrator | a989680a7862 registry.osism.tech/kolla/release/redis:7.0.15.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes (healthy) redis 2025-08-29 18:00:46.744533 | orchestrator | cba1b50deaf3 registry.osism.tech/kolla/release/memcached:1.6.18.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes (healthy) memcached 2025-08-29 18:00:46.744553 | orchestrator | fa06b54a3736 registry.osism.tech/kolla/release/cron:3.0.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes cron 2025-08-29 18:00:46.744564 | orchestrator | 1be99c78c71e registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711 "dumb-init --single-…" 30 minutes ago Up 30 minutes kolla_toolbox 2025-08-29 18:00:46.744575 | orchestrator | 0a145c35e9fd registry.osism.tech/kolla/release/fluentd:5.0.7.20250711 "dumb-init --single-…" 31 minutes ago Up 31 minutes fluentd 2025-08-29 18:00:47.049162 | orchestrator | 2025-08-29 18:00:47.049339 | orchestrator | ## Images @ testbed-node-2 2025-08-29 18:00:47.049360 | orchestrator | 2025-08-29 18:00:47.049372 | orchestrator | + echo 2025-08-29 18:00:47.049384 | orchestrator | + echo '## Images @ testbed-node-2' 2025-08-29 18:00:47.049396 | orchestrator | + echo 2025-08-29 18:00:47.049407 | orchestrator | + osism container testbed-node-2 images 2025-08-29 18:00:49.317325 | orchestrator | REPOSITORY TAG IMAGE ID CREATED SIZE 2025-08-29 18:00:49.317430 | orchestrator | registry.osism.tech/kolla/release/fluentd 5.0.7.20250711 eaa70c1312aa 6 weeks ago 628MB 2025-08-29 18:00:49.317462 | orchestrator | registry.osism.tech/kolla/release/keepalived 2.2.7.20250711 c7f6abdb2516 6 weeks ago 329MB 2025-08-29 18:00:49.317471 | orchestrator | registry.osism.tech/kolla/release/haproxy 2.6.12.20250711 0a9fd950fe86 6 weeks ago 326MB 2025-08-29 18:00:49.317484 | orchestrator | registry.osism.tech/kolla/release/opensearch 2.19.2.20250711 d8c44fac73c2 6 weeks ago 1.59GB 2025-08-29 18:00:49.317495 | orchestrator | registry.osism.tech/kolla/release/opensearch-dashboards 2.19.2.20250711 db87020f3b90 6 weeks ago 1.55GB 2025-08-29 18:00:49.317505 | orchestrator | registry.osism.tech/kolla/release/proxysql 2.7.3.20250711 4c6eaa052643 6 weeks ago 417MB 2025-08-29 18:00:49.317517 | orchestrator | registry.osism.tech/kolla/release/memcached 1.6.18.20250711 cd87896ace76 6 weeks ago 318MB 2025-08-29 18:00:49.317529 | orchestrator | registry.osism.tech/kolla/release/rabbitmq 3.13.7.20250711 4ce47f209c9b 6 weeks ago 375MB 2025-08-29 18:00:49.317539 | orchestrator | registry.osism.tech/kolla/release/kolla-toolbox 19.5.1.20250711 ad526ea47263 6 weeks ago 746MB 2025-08-29 18:00:49.317550 | orchestrator | registry.osism.tech/kolla/release/grafana 12.0.2.20250711 f4164dfd1b02 6 weeks ago 1.01GB 2025-08-29 18:00:49.317561 | orchestrator | registry.osism.tech/kolla/release/cron 3.0.20250711 de0bd651bf89 6 weeks ago 318MB 2025-08-29 18:00:49.317572 | orchestrator | registry.osism.tech/kolla/release/openvswitch-db-server 3.4.2.20250711 15f29551e6ce 6 weeks ago 361MB 2025-08-29 18:00:49.317583 | orchestrator | registry.osism.tech/kolla/release/openvswitch-vswitchd 3.4.2.20250711 ea9ea8f197d8 6 weeks ago 361MB 2025-08-29 18:00:49.317593 | orchestrator | registry.osism.tech/kolla/release/horizon 25.1.1.20250711 d4ae4a297d3b 6 weeks ago 1.21GB 2025-08-29 18:00:49.317603 | orchestrator | registry.osism.tech/kolla/release/prometheus-mysqld-exporter 0.16.0.20250711 142dafde994c 6 weeks ago 353MB 2025-08-29 18:00:49.317615 | orchestrator | registry.osism.tech/kolla/release/prometheus-cadvisor 0.49.2.20250711 937f4652a0d1 6 weeks ago 410MB 2025-08-29 18:00:49.317625 | orchestrator | registry.osism.tech/kolla/release/prometheus-elasticsearch-exporter 1.8.0.20250711 62e13ec7689a 6 weeks ago 344MB 2025-08-29 18:00:49.317634 | orchestrator | registry.osism.tech/kolla/release/prometheus-node-exporter 1.8.2.20250711 361ce2873c65 6 weeks ago 358MB 2025-08-29 18:00:49.317645 | orchestrator | registry.osism.tech/kolla/release/prometheus-memcached-exporter 0.15.0.20250711 834c4c2dcd78 6 weeks ago 351MB 2025-08-29 18:00:49.317655 | orchestrator | registry.osism.tech/kolla/release/redis-sentinel 7.0.15.20250711 534f393a19e2 6 weeks ago 324MB 2025-08-29 18:00:49.317664 | orchestrator | registry.osism.tech/kolla/release/redis 7.0.15.20250711 d7d5c3586026 6 weeks ago 324MB 2025-08-29 18:00:49.317674 | orchestrator | registry.osism.tech/kolla/release/mariadb-server 10.11.13.20250711 5892b19e1064 6 weeks ago 590MB 2025-08-29 18:00:49.317684 | orchestrator | registry.osism.tech/kolla/release/ovn-sb-db-server 24.9.2.20250711 28654474dfe5 6 weeks ago 946MB 2025-08-29 18:00:49.317718 | orchestrator | registry.osism.tech/kolla/release/ovn-controller 24.9.2.20250711 65e36d1176bd 6 weeks ago 947MB 2025-08-29 18:00:49.317729 | orchestrator | registry.osism.tech/kolla/release/ovn-northd 24.9.2.20250711 58ad45688234 6 weeks ago 947MB 2025-08-29 18:00:49.317741 | orchestrator | registry.osism.tech/kolla/release/ovn-nb-db-server 24.9.2.20250711 affa47a97549 6 weeks ago 946MB 2025-08-29 18:00:49.317751 | orchestrator | registry.osism.tech/kolla/release/glance-api 29.0.1.20250711 65a4d0afbb1c 6 weeks ago 1.15GB 2025-08-29 18:00:49.317760 | orchestrator | registry.osism.tech/kolla/release/placement-api 12.0.1.20250711 2b6bd346ad18 6 weeks ago 1.04GB 2025-08-29 18:00:49.317770 | orchestrator | registry.osism.tech/kolla/release/barbican-api 19.0.1.20250711 1b7dd2682590 6 weeks ago 1.06GB 2025-08-29 18:00:49.317780 | orchestrator | registry.osism.tech/kolla/release/barbican-worker 19.0.1.20250711 e475391ce44d 6 weeks ago 1.06GB 2025-08-29 18:00:49.317791 | orchestrator | registry.osism.tech/kolla/release/barbican-keystone-listener 19.0.1.20250711 09290580fa03 6 weeks ago 1.06GB 2025-08-29 18:00:49.317821 | orchestrator | registry.osism.tech/kolla/release/cinder-scheduler 25.2.1.20250711 a09a8be1b711 6 weeks ago 1.41GB 2025-08-29 18:00:49.317833 | orchestrator | registry.osism.tech/kolla/release/cinder-api 25.2.1.20250711 c0d28e8febb9 6 weeks ago 1.41GB 2025-08-29 18:00:49.317844 | orchestrator | registry.osism.tech/kolla/release/nova-scheduler 30.0.1.20250711 e0ad0ae52bef 6 weeks ago 1.29GB 2025-08-29 18:00:49.317855 | orchestrator | registry.osism.tech/kolla/release/nova-novncproxy 30.0.1.20250711 b395cfe7f13f 6 weeks ago 1.42GB 2025-08-29 18:00:49.317865 | orchestrator | registry.osism.tech/kolla/release/nova-api 30.0.1.20250711 ee83c124eb76 6 weeks ago 1.29GB 2025-08-29 18:00:49.317876 | orchestrator | registry.osism.tech/kolla/release/nova-conductor 30.0.1.20250711 44e25b162470 6 weeks ago 1.29GB 2025-08-29 18:00:49.317887 | orchestrator | registry.osism.tech/kolla/release/magnum-api 19.0.1.20250711 71f47d2b2def 6 weeks ago 1.2GB 2025-08-29 18:00:49.317904 | orchestrator | registry.osism.tech/kolla/release/magnum-conductor 19.0.1.20250711 13b61cb4a5d2 6 weeks ago 1.31GB 2025-08-29 18:00:49.317916 | orchestrator | registry.osism.tech/kolla/release/designate-mdns 19.0.1.20250711 a030b794eaa9 6 weeks ago 1.05GB 2025-08-29 18:00:49.317928 | orchestrator | registry.osism.tech/kolla/release/designate-producer 19.0.1.20250711 2d0954c30848 6 weeks ago 1.05GB 2025-08-29 18:00:49.317939 | orchestrator | registry.osism.tech/kolla/release/designate-central 19.0.1.20250711 f7fa0bcabe47 6 weeks ago 1.05GB 2025-08-29 18:00:49.317950 | orchestrator | registry.osism.tech/kolla/release/designate-worker 19.0.1.20250711 4de726ebba0e 6 weeks ago 1.06GB 2025-08-29 18:00:49.317961 | orchestrator | registry.osism.tech/kolla/release/designate-backend-bind9 19.0.1.20250711 a14c6ace0b24 6 weeks ago 1.06GB 2025-08-29 18:00:49.317972 | orchestrator | registry.osism.tech/kolla/release/designate-api 19.0.1.20250711 2a2b32cdb83f 6 weeks ago 1.05GB 2025-08-29 18:00:49.317984 | orchestrator | registry.osism.tech/kolla/release/keystone-fernet 26.0.1.20250711 53889b0cb73d 6 weeks ago 1.11GB 2025-08-29 18:00:49.317994 | orchestrator | registry.osism.tech/kolla/release/keystone 26.0.1.20250711 caf4f12b4799 6 weeks ago 1.13GB 2025-08-29 18:00:49.318006 | orchestrator | registry.osism.tech/kolla/release/keystone-ssh 26.0.1.20250711 3ba6da1abaea 6 weeks ago 1.11GB 2025-08-29 18:00:49.318015 | orchestrator | registry.osism.tech/kolla/release/neutron-server 25.2.1.20250711 8377b7d24f73 6 weeks ago 1.24GB 2025-08-29 18:00:49.318074 | orchestrator | registry.osism.tech/osism/ceph-daemon 18.2.7 5f92363b1f93 3 months ago 1.27GB 2025-08-29 18:00:49.598657 | orchestrator | + sh -c /opt/configuration/scripts/check-services.sh 2025-08-29 18:00:49.605170 | orchestrator | + set -e 2025-08-29 18:00:49.605246 | orchestrator | + source /opt/manager-vars.sh 2025-08-29 18:00:49.606675 | orchestrator | ++ export NUMBER_OF_NODES=6 2025-08-29 18:00:49.606724 | orchestrator | ++ NUMBER_OF_NODES=6 2025-08-29 18:00:49.606745 | orchestrator | ++ export CEPH_VERSION=reef 2025-08-29 18:00:49.607636 | orchestrator | ++ CEPH_VERSION=reef 2025-08-29 18:00:49.607681 | orchestrator | ++ export CONFIGURATION_VERSION=main 2025-08-29 18:00:49.607716 | orchestrator | ++ CONFIGURATION_VERSION=main 2025-08-29 18:00:49.607729 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 18:00:49.607740 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 18:00:49.607751 | orchestrator | ++ export OPENSTACK_VERSION=2024.2 2025-08-29 18:00:49.607762 | orchestrator | ++ OPENSTACK_VERSION=2024.2 2025-08-29 18:00:49.607774 | orchestrator | ++ export ARA=false 2025-08-29 18:00:49.607785 | orchestrator | ++ ARA=false 2025-08-29 18:00:49.607797 | orchestrator | ++ export DEPLOY_MODE=manager 2025-08-29 18:00:49.607808 | orchestrator | ++ DEPLOY_MODE=manager 2025-08-29 18:00:49.607819 | orchestrator | ++ export TEMPEST=false 2025-08-29 18:00:49.607830 | orchestrator | ++ TEMPEST=false 2025-08-29 18:00:49.607842 | orchestrator | ++ export IS_ZUUL=true 2025-08-29 18:00:49.607852 | orchestrator | ++ IS_ZUUL=true 2025-08-29 18:00:49.607864 | orchestrator | ++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 18:00:49.607875 | orchestrator | ++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 18:00:49.607885 | orchestrator | ++ export EXTERNAL_API=false 2025-08-29 18:00:49.607896 | orchestrator | ++ EXTERNAL_API=false 2025-08-29 18:00:49.607906 | orchestrator | ++ export IMAGE_USER=ubuntu 2025-08-29 18:00:49.607917 | orchestrator | ++ IMAGE_USER=ubuntu 2025-08-29 18:00:49.607928 | orchestrator | ++ export IMAGE_NODE_USER=ubuntu 2025-08-29 18:00:49.607939 | orchestrator | ++ IMAGE_NODE_USER=ubuntu 2025-08-29 18:00:49.607949 | orchestrator | ++ export CEPH_STACK=ceph-ansible 2025-08-29 18:00:49.607960 | orchestrator | ++ CEPH_STACK=ceph-ansible 2025-08-29 18:00:49.607971 | orchestrator | + [[ ceph-ansible == \c\e\p\h\-\a\n\s\i\b\l\e ]] 2025-08-29 18:00:49.607982 | orchestrator | + sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh 2025-08-29 18:00:49.616243 | orchestrator | + set -e 2025-08-29 18:00:49.616327 | orchestrator | + source /opt/configuration/scripts/include.sh 2025-08-29 18:00:49.616340 | orchestrator | ++ export INTERACTIVE=false 2025-08-29 18:00:49.616351 | orchestrator | ++ INTERACTIVE=false 2025-08-29 18:00:49.616362 | orchestrator | ++ export OSISM_APPLY_RETRY=1 2025-08-29 18:00:49.616373 | orchestrator | ++ OSISM_APPLY_RETRY=1 2025-08-29 18:00:49.616384 | orchestrator | + source /opt/configuration/scripts/manager-version.sh 2025-08-29 18:00:49.617139 | orchestrator | +++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml 2025-08-29 18:00:49.622368 | orchestrator | 2025-08-29 18:00:49.622420 | orchestrator | # Ceph status 2025-08-29 18:00:49.622439 | orchestrator | 2025-08-29 18:00:49.622455 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 18:00:49.622473 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 18:00:49.622492 | orchestrator | + echo 2025-08-29 18:00:49.622518 | orchestrator | + echo '# Ceph status' 2025-08-29 18:00:49.622536 | orchestrator | + echo 2025-08-29 18:00:49.622555 | orchestrator | + ceph -s 2025-08-29 18:00:50.225580 | orchestrator | cluster: 2025-08-29 18:00:50.225687 | orchestrator | id: 11111111-1111-1111-1111-111111111111 2025-08-29 18:00:50.225703 | orchestrator | health: HEALTH_OK 2025-08-29 18:00:50.225717 | orchestrator | 2025-08-29 18:00:50.225729 | orchestrator | services: 2025-08-29 18:00:50.225740 | orchestrator | mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 28m) 2025-08-29 18:00:50.225769 | orchestrator | mgr: testbed-node-2(active, since 16m), standbys: testbed-node-1, testbed-node-0 2025-08-29 18:00:50.225789 | orchestrator | mds: 1/1 daemons up, 2 standby 2025-08-29 18:00:50.225808 | orchestrator | osd: 6 osds: 6 up (since 24m), 6 in (since 25m) 2025-08-29 18:00:50.225829 | orchestrator | rgw: 3 daemons active (3 hosts, 1 zones) 2025-08-29 18:00:50.225849 | orchestrator | 2025-08-29 18:00:50.225868 | orchestrator | data: 2025-08-29 18:00:50.225880 | orchestrator | volumes: 1/1 healthy 2025-08-29 18:00:50.225891 | orchestrator | pools: 14 pools, 401 pgs 2025-08-29 18:00:50.225903 | orchestrator | objects: 522 objects, 2.2 GiB 2025-08-29 18:00:50.225914 | orchestrator | usage: 7.1 GiB used, 113 GiB / 120 GiB avail 2025-08-29 18:00:50.225925 | orchestrator | pgs: 401 active+clean 2025-08-29 18:00:50.225936 | orchestrator | 2025-08-29 18:00:50.291335 | orchestrator | 2025-08-29 18:00:50.291431 | orchestrator | # Ceph versions 2025-08-29 18:00:50.291443 | orchestrator | 2025-08-29 18:00:50.291451 | orchestrator | + echo 2025-08-29 18:00:50.291459 | orchestrator | + echo '# Ceph versions' 2025-08-29 18:00:50.291467 | orchestrator | + echo 2025-08-29 18:00:50.291499 | orchestrator | + ceph versions 2025-08-29 18:00:50.898889 | orchestrator | { 2025-08-29 18:00:50.898992 | orchestrator | "mon": { 2025-08-29 18:00:50.899008 | orchestrator | "ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)": 3 2025-08-29 18:00:50.899020 | orchestrator | }, 2025-08-29 18:00:50.899032 | orchestrator | "mgr": { 2025-08-29 18:00:50.899043 | orchestrator | "ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)": 3 2025-08-29 18:00:50.899054 | orchestrator | }, 2025-08-29 18:00:50.899065 | orchestrator | "osd": { 2025-08-29 18:00:50.899076 | orchestrator | "ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)": 6 2025-08-29 18:00:50.899088 | orchestrator | }, 2025-08-29 18:00:50.899098 | orchestrator | "mds": { 2025-08-29 18:00:50.899110 | orchestrator | "ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)": 3 2025-08-29 18:00:50.899120 | orchestrator | }, 2025-08-29 18:00:50.899131 | orchestrator | "rgw": { 2025-08-29 18:00:50.899142 | orchestrator | "ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)": 3 2025-08-29 18:00:50.899153 | orchestrator | }, 2025-08-29 18:00:50.899164 | orchestrator | "overall": { 2025-08-29 18:00:50.899177 | orchestrator | "ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)": 18 2025-08-29 18:00:50.899188 | orchestrator | } 2025-08-29 18:00:50.899199 | orchestrator | } 2025-08-29 18:00:50.940016 | orchestrator | 2025-08-29 18:00:50.940127 | orchestrator | # Ceph OSD tree 2025-08-29 18:00:50.940143 | orchestrator | 2025-08-29 18:00:50.940155 | orchestrator | + echo 2025-08-29 18:00:50.940167 | orchestrator | + echo '# Ceph OSD tree' 2025-08-29 18:00:50.940179 | orchestrator | + echo 2025-08-29 18:00:50.940191 | orchestrator | + ceph osd df tree 2025-08-29 18:00:51.494838 | orchestrator | ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL %USE VAR PGS STATUS TYPE NAME 2025-08-29 18:00:51.494957 | orchestrator | -1 0.11691 - 120 GiB 7.1 GiB 6.7 GiB 6 KiB 430 MiB 113 GiB 5.92 1.00 - root default 2025-08-29 18:00:51.494973 | orchestrator | -3 0.03897 - 40 GiB 2.4 GiB 2.2 GiB 2 KiB 143 MiB 38 GiB 5.92 1.00 - host testbed-node-3 2025-08-29 18:00:51.494985 | orchestrator | 0 hdd 0.01949 1.00000 20 GiB 1.5 GiB 1.4 GiB 1 KiB 70 MiB 19 GiB 7.27 1.23 201 up osd.0 2025-08-29 18:00:51.494996 | orchestrator | 5 hdd 0.01949 1.00000 20 GiB 932 MiB 859 MiB 1 KiB 74 MiB 19 GiB 4.56 0.77 189 up osd.5 2025-08-29 18:00:51.495008 | orchestrator | -7 0.03897 - 40 GiB 2.4 GiB 2.2 GiB 2 KiB 143 MiB 38 GiB 5.92 1.00 - host testbed-node-4 2025-08-29 18:00:51.495019 | orchestrator | 2 hdd 0.01949 1.00000 20 GiB 1.4 GiB 1.3 GiB 1 KiB 70 MiB 19 GiB 6.96 1.18 188 up osd.2 2025-08-29 18:00:51.495030 | orchestrator | 3 hdd 0.01949 1.00000 20 GiB 996 MiB 923 MiB 1 KiB 74 MiB 19 GiB 4.87 0.82 200 up osd.3 2025-08-29 18:00:51.495041 | orchestrator | -5 0.03897 - 40 GiB 2.4 GiB 2.2 GiB 2 KiB 143 MiB 38 GiB 5.92 1.00 - host testbed-node-5 2025-08-29 18:00:51.495052 | orchestrator | 1 hdd 0.01949 1.00000 20 GiB 1.2 GiB 1.2 GiB 1 KiB 74 MiB 19 GiB 6.16 1.04 190 up osd.1 2025-08-29 18:00:51.495063 | orchestrator | 4 hdd 0.01949 1.00000 20 GiB 1.1 GiB 1.1 GiB 1 KiB 70 MiB 19 GiB 5.67 0.96 202 up osd.4 2025-08-29 18:00:51.495074 | orchestrator | TOTAL 120 GiB 7.1 GiB 6.7 GiB 9.3 KiB 430 MiB 113 GiB 5.92 2025-08-29 18:00:51.495086 | orchestrator | MIN/MAX VAR: 0.77/1.23 STDDEV: 1.00 2025-08-29 18:00:51.537248 | orchestrator | 2025-08-29 18:00:51.537423 | orchestrator | # Ceph monitor status 2025-08-29 18:00:51.537449 | orchestrator | 2025-08-29 18:00:51.537471 | orchestrator | + echo 2025-08-29 18:00:51.537490 | orchestrator | + echo '# Ceph monitor status' 2025-08-29 18:00:51.537513 | orchestrator | + echo 2025-08-29 18:00:51.537525 | orchestrator | + ceph mon stat 2025-08-29 18:00:52.132452 | orchestrator | e1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 8, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2 2025-08-29 18:00:52.174523 | orchestrator | 2025-08-29 18:00:52.174655 | orchestrator | # Ceph quorum status 2025-08-29 18:00:52.174671 | orchestrator | 2025-08-29 18:00:52.174683 | orchestrator | + echo 2025-08-29 18:00:52.174694 | orchestrator | + echo '# Ceph quorum status' 2025-08-29 18:00:52.174706 | orchestrator | + echo 2025-08-29 18:00:52.174784 | orchestrator | + jq 2025-08-29 18:00:52.174799 | orchestrator | + ceph quorum_status 2025-08-29 18:00:52.820765 | orchestrator | { 2025-08-29 18:00:52.820903 | orchestrator | "election_epoch": 8, 2025-08-29 18:00:52.820934 | orchestrator | "quorum": [ 2025-08-29 18:00:52.820955 | orchestrator | 0, 2025-08-29 18:00:52.820973 | orchestrator | 1, 2025-08-29 18:00:52.820992 | orchestrator | 2 2025-08-29 18:00:52.821011 | orchestrator | ], 2025-08-29 18:00:52.821030 | orchestrator | "quorum_names": [ 2025-08-29 18:00:52.821050 | orchestrator | "testbed-node-0", 2025-08-29 18:00:52.821068 | orchestrator | "testbed-node-1", 2025-08-29 18:00:52.821088 | orchestrator | "testbed-node-2" 2025-08-29 18:00:52.821102 | orchestrator | ], 2025-08-29 18:00:52.821116 | orchestrator | "quorum_leader_name": "testbed-node-0", 2025-08-29 18:00:52.821130 | orchestrator | "quorum_age": 1713, 2025-08-29 18:00:52.821143 | orchestrator | "features": { 2025-08-29 18:00:52.821156 | orchestrator | "quorum_con": "4540138322906710015", 2025-08-29 18:00:52.821169 | orchestrator | "quorum_mon": [ 2025-08-29 18:00:52.821181 | orchestrator | "kraken", 2025-08-29 18:00:52.821193 | orchestrator | "luminous", 2025-08-29 18:00:52.821206 | orchestrator | "mimic", 2025-08-29 18:00:52.821218 | orchestrator | "osdmap-prune", 2025-08-29 18:00:52.821231 | orchestrator | "nautilus", 2025-08-29 18:00:52.821243 | orchestrator | "octopus", 2025-08-29 18:00:52.821255 | orchestrator | "pacific", 2025-08-29 18:00:52.821309 | orchestrator | "elector-pinging", 2025-08-29 18:00:52.821331 | orchestrator | "quincy", 2025-08-29 18:00:52.821351 | orchestrator | "reef" 2025-08-29 18:00:52.821369 | orchestrator | ] 2025-08-29 18:00:52.821382 | orchestrator | }, 2025-08-29 18:00:52.821395 | orchestrator | "monmap": { 2025-08-29 18:00:52.821407 | orchestrator | "epoch": 1, 2025-08-29 18:00:52.821420 | orchestrator | "fsid": "11111111-1111-1111-1111-111111111111", 2025-08-29 18:00:52.821433 | orchestrator | "modified": "2025-08-29T17:32:02.205963Z", 2025-08-29 18:00:52.821444 | orchestrator | "created": "2025-08-29T17:32:02.205963Z", 2025-08-29 18:00:52.821455 | orchestrator | "min_mon_release": 18, 2025-08-29 18:00:52.821466 | orchestrator | "min_mon_release_name": "reef", 2025-08-29 18:00:52.821477 | orchestrator | "election_strategy": 1, 2025-08-29 18:00:52.821488 | orchestrator | "disallowed_leaders: ": "", 2025-08-29 18:00:52.821499 | orchestrator | "stretch_mode": false, 2025-08-29 18:00:52.821510 | orchestrator | "tiebreaker_mon": "", 2025-08-29 18:00:52.821520 | orchestrator | "removed_ranks: ": "", 2025-08-29 18:00:52.821531 | orchestrator | "features": { 2025-08-29 18:00:52.821542 | orchestrator | "persistent": [ 2025-08-29 18:00:52.821553 | orchestrator | "kraken", 2025-08-29 18:00:52.821563 | orchestrator | "luminous", 2025-08-29 18:00:52.821574 | orchestrator | "mimic", 2025-08-29 18:00:52.821585 | orchestrator | "osdmap-prune", 2025-08-29 18:00:52.821595 | orchestrator | "nautilus", 2025-08-29 18:00:52.821606 | orchestrator | "octopus", 2025-08-29 18:00:52.821616 | orchestrator | "pacific", 2025-08-29 18:00:52.821627 | orchestrator | "elector-pinging", 2025-08-29 18:00:52.821638 | orchestrator | "quincy", 2025-08-29 18:00:52.821649 | orchestrator | "reef" 2025-08-29 18:00:52.821659 | orchestrator | ], 2025-08-29 18:00:52.821671 | orchestrator | "optional": [] 2025-08-29 18:00:52.821682 | orchestrator | }, 2025-08-29 18:00:52.821692 | orchestrator | "mons": [ 2025-08-29 18:00:52.821703 | orchestrator | { 2025-08-29 18:00:52.821714 | orchestrator | "rank": 0, 2025-08-29 18:00:52.821725 | orchestrator | "name": "testbed-node-0", 2025-08-29 18:00:52.821736 | orchestrator | "public_addrs": { 2025-08-29 18:00:52.821746 | orchestrator | "addrvec": [ 2025-08-29 18:00:52.821757 | orchestrator | { 2025-08-29 18:00:52.821768 | orchestrator | "type": "v2", 2025-08-29 18:00:52.821779 | orchestrator | "addr": "192.168.16.10:3300", 2025-08-29 18:00:52.821790 | orchestrator | "nonce": 0 2025-08-29 18:00:52.821800 | orchestrator | }, 2025-08-29 18:00:52.821811 | orchestrator | { 2025-08-29 18:00:52.821822 | orchestrator | "type": "v1", 2025-08-29 18:00:52.821857 | orchestrator | "addr": "192.168.16.10:6789", 2025-08-29 18:00:52.821869 | orchestrator | "nonce": 0 2025-08-29 18:00:52.821879 | orchestrator | } 2025-08-29 18:00:52.821890 | orchestrator | ] 2025-08-29 18:00:52.821901 | orchestrator | }, 2025-08-29 18:00:52.821911 | orchestrator | "addr": "192.168.16.10:6789/0", 2025-08-29 18:00:52.821922 | orchestrator | "public_addr": "192.168.16.10:6789/0", 2025-08-29 18:00:52.821933 | orchestrator | "priority": 0, 2025-08-29 18:00:52.821943 | orchestrator | "weight": 0, 2025-08-29 18:00:52.821954 | orchestrator | "crush_location": "{}" 2025-08-29 18:00:52.821965 | orchestrator | }, 2025-08-29 18:00:52.821976 | orchestrator | { 2025-08-29 18:00:52.821986 | orchestrator | "rank": 1, 2025-08-29 18:00:52.821997 | orchestrator | "name": "testbed-node-1", 2025-08-29 18:00:52.822008 | orchestrator | "public_addrs": { 2025-08-29 18:00:52.822076 | orchestrator | "addrvec": [ 2025-08-29 18:00:52.822091 | orchestrator | { 2025-08-29 18:00:52.822102 | orchestrator | "type": "v2", 2025-08-29 18:00:52.822113 | orchestrator | "addr": "192.168.16.11:3300", 2025-08-29 18:00:52.822124 | orchestrator | "nonce": 0 2025-08-29 18:00:52.822135 | orchestrator | }, 2025-08-29 18:00:52.822146 | orchestrator | { 2025-08-29 18:00:52.822157 | orchestrator | "type": "v1", 2025-08-29 18:00:52.822168 | orchestrator | "addr": "192.168.16.11:6789", 2025-08-29 18:00:52.822179 | orchestrator | "nonce": 0 2025-08-29 18:00:52.822190 | orchestrator | } 2025-08-29 18:00:52.822201 | orchestrator | ] 2025-08-29 18:00:52.822211 | orchestrator | }, 2025-08-29 18:00:52.822222 | orchestrator | "addr": "192.168.16.11:6789/0", 2025-08-29 18:00:52.822234 | orchestrator | "public_addr": "192.168.16.11:6789/0", 2025-08-29 18:00:52.822244 | orchestrator | "priority": 0, 2025-08-29 18:00:52.822255 | orchestrator | "weight": 0, 2025-08-29 18:00:52.822315 | orchestrator | "crush_location": "{}" 2025-08-29 18:00:52.822326 | orchestrator | }, 2025-08-29 18:00:52.822337 | orchestrator | { 2025-08-29 18:00:52.822348 | orchestrator | "rank": 2, 2025-08-29 18:00:52.822359 | orchestrator | "name": "testbed-node-2", 2025-08-29 18:00:52.822369 | orchestrator | "public_addrs": { 2025-08-29 18:00:52.822380 | orchestrator | "addrvec": [ 2025-08-29 18:00:52.822391 | orchestrator | { 2025-08-29 18:00:52.822402 | orchestrator | "type": "v2", 2025-08-29 18:00:52.822412 | orchestrator | "addr": "192.168.16.12:3300", 2025-08-29 18:00:52.822423 | orchestrator | "nonce": 0 2025-08-29 18:00:52.822434 | orchestrator | }, 2025-08-29 18:00:52.822444 | orchestrator | { 2025-08-29 18:00:52.822455 | orchestrator | "type": "v1", 2025-08-29 18:00:52.822466 | orchestrator | "addr": "192.168.16.12:6789", 2025-08-29 18:00:52.822477 | orchestrator | "nonce": 0 2025-08-29 18:00:52.822488 | orchestrator | } 2025-08-29 18:00:52.822498 | orchestrator | ] 2025-08-29 18:00:52.822509 | orchestrator | }, 2025-08-29 18:00:52.822520 | orchestrator | "addr": "192.168.16.12:6789/0", 2025-08-29 18:00:52.822531 | orchestrator | "public_addr": "192.168.16.12:6789/0", 2025-08-29 18:00:52.822541 | orchestrator | "priority": 0, 2025-08-29 18:00:52.822552 | orchestrator | "weight": 0, 2025-08-29 18:00:52.822563 | orchestrator | "crush_location": "{}" 2025-08-29 18:00:52.822574 | orchestrator | } 2025-08-29 18:00:52.822585 | orchestrator | ] 2025-08-29 18:00:52.822596 | orchestrator | } 2025-08-29 18:00:52.822606 | orchestrator | } 2025-08-29 18:00:52.822634 | orchestrator | 2025-08-29 18:00:52.822645 | orchestrator | # Ceph free space status 2025-08-29 18:00:52.822656 | orchestrator | 2025-08-29 18:00:52.822667 | orchestrator | + echo 2025-08-29 18:00:52.822679 | orchestrator | + echo '# Ceph free space status' 2025-08-29 18:00:52.822689 | orchestrator | + echo 2025-08-29 18:00:52.822700 | orchestrator | + ceph df 2025-08-29 18:00:53.421088 | orchestrator | --- RAW STORAGE --- 2025-08-29 18:00:53.421171 | orchestrator | CLASS SIZE AVAIL USED RAW USED %RAW USED 2025-08-29 18:00:53.421192 | orchestrator | hdd 120 GiB 113 GiB 7.1 GiB 7.1 GiB 5.92 2025-08-29 18:00:53.421200 | orchestrator | TOTAL 120 GiB 113 GiB 7.1 GiB 7.1 GiB 5.92 2025-08-29 18:00:53.421207 | orchestrator | 2025-08-29 18:00:53.421214 | orchestrator | --- POOLS --- 2025-08-29 18:00:53.421222 | orchestrator | POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL 2025-08-29 18:00:53.421230 | orchestrator | .mgr 1 1 577 KiB 2 1.1 MiB 0 53 GiB 2025-08-29 18:00:53.421236 | orchestrator | cephfs_data 2 32 0 B 0 0 B 0 35 GiB 2025-08-29 18:00:53.421305 | orchestrator | cephfs_metadata 3 16 4.4 KiB 22 96 KiB 0 35 GiB 2025-08-29 18:00:53.421313 | orchestrator | default.rgw.buckets.data 4 32 0 B 0 0 B 0 35 GiB 2025-08-29 18:00:53.421320 | orchestrator | default.rgw.buckets.index 5 32 0 B 0 0 B 0 35 GiB 2025-08-29 18:00:53.421327 | orchestrator | default.rgw.control 6 32 0 B 8 0 B 0 35 GiB 2025-08-29 18:00:53.421333 | orchestrator | default.rgw.log 7 32 3.6 KiB 177 408 KiB 0 35 GiB 2025-08-29 18:00:53.421339 | orchestrator | default.rgw.meta 8 32 0 B 0 0 B 0 35 GiB 2025-08-29 18:00:53.421346 | orchestrator | .rgw.root 9 32 2.6 KiB 6 48 KiB 0 53 GiB 2025-08-29 18:00:53.421352 | orchestrator | backups 10 32 19 B 2 12 KiB 0 35 GiB 2025-08-29 18:00:53.421358 | orchestrator | volumes 11 32 19 B 2 12 KiB 0 35 GiB 2025-08-29 18:00:53.421364 | orchestrator | images 12 32 2.2 GiB 299 6.7 GiB 5.95 35 GiB 2025-08-29 18:00:53.421371 | orchestrator | metrics 13 32 19 B 2 12 KiB 0 35 GiB 2025-08-29 18:00:53.421377 | orchestrator | vms 14 32 19 B 2 12 KiB 0 35 GiB 2025-08-29 18:00:53.464219 | orchestrator | ++ semver 9.2.0 5.0.0 2025-08-29 18:00:53.518176 | orchestrator | + [[ 1 -eq -1 ]] 2025-08-29 18:00:53.518296 | orchestrator | + [[ ! -e /etc/redhat-release ]] 2025-08-29 18:00:53.518312 | orchestrator | + osism apply facts 2025-08-29 18:01:05.472958 | orchestrator | 2025-08-29 18:01:05 | INFO  | Task 273ab17e-5216-4f68-863a-ff3f87aa9e0c (facts) was prepared for execution. 2025-08-29 18:01:05.473079 | orchestrator | 2025-08-29 18:01:05 | INFO  | It takes a moment until task 273ab17e-5216-4f68-863a-ff3f87aa9e0c (facts) has been started and output is visible here. 2025-08-29 18:01:19.449971 | orchestrator | 2025-08-29 18:01:19.450098 | orchestrator | PLAY [Apply role facts] ******************************************************** 2025-08-29 18:01:19.450109 | orchestrator | 2025-08-29 18:01:19.450116 | orchestrator | TASK [osism.commons.facts : Create custom facts directory] ********************* 2025-08-29 18:01:19.450123 | orchestrator | Friday 29 August 2025 18:01:09 +0000 (0:00:00.274) 0:00:00.274 ********* 2025-08-29 18:01:19.450129 | orchestrator | ok: [testbed-manager] 2025-08-29 18:01:19.450136 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:19.450143 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:01:19.450149 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:01:19.450155 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:01:19.450161 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:01:19.450167 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:01:19.450173 | orchestrator | 2025-08-29 18:01:19.450179 | orchestrator | TASK [osism.commons.facts : Copy fact files] *********************************** 2025-08-29 18:01:19.450199 | orchestrator | Friday 29 August 2025 18:01:10 +0000 (0:00:01.385) 0:00:01.659 ********* 2025-08-29 18:01:19.450205 | orchestrator | skipping: [testbed-manager] 2025-08-29 18:01:19.450212 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:19.450218 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:01:19.450224 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:01:19.450229 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:01:19.450235 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:01:19.450241 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:01:19.450247 | orchestrator | 2025-08-29 18:01:19.450253 | orchestrator | PLAY [Gather facts for all hosts] ********************************************** 2025-08-29 18:01:19.450292 | orchestrator | 2025-08-29 18:01:19.450299 | orchestrator | TASK [Gathers facts about hosts] *********************************************** 2025-08-29 18:01:19.450305 | orchestrator | Friday 29 August 2025 18:01:12 +0000 (0:00:01.130) 0:00:02.789 ********* 2025-08-29 18:01:19.450311 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:01:19.450317 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:01:19.450323 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:19.450329 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:01:19.450352 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:01:19.450358 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:01:19.450364 | orchestrator | ok: [testbed-manager] 2025-08-29 18:01:19.450370 | orchestrator | 2025-08-29 18:01:19.450376 | orchestrator | PLAY [Gather facts for all hosts if using --limit] ***************************** 2025-08-29 18:01:19.450381 | orchestrator | 2025-08-29 18:01:19.450387 | orchestrator | TASK [Gather facts for all hosts] ********************************************** 2025-08-29 18:01:19.450393 | orchestrator | Friday 29 August 2025 18:01:18 +0000 (0:00:06.431) 0:00:09.221 ********* 2025-08-29 18:01:19.450399 | orchestrator | skipping: [testbed-manager] 2025-08-29 18:01:19.450405 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:19.450411 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:01:19.450416 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:01:19.450422 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:01:19.450428 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:01:19.450434 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:01:19.450439 | orchestrator | 2025-08-29 18:01:19.450445 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 18:01:19.450451 | orchestrator | testbed-manager : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:19.450458 | orchestrator | testbed-node-0 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:19.450464 | orchestrator | testbed-node-1 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:19.450470 | orchestrator | testbed-node-2 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:19.450475 | orchestrator | testbed-node-3 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:19.450481 | orchestrator | testbed-node-4 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:19.450487 | orchestrator | testbed-node-5 : ok=2  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:19.450493 | orchestrator | 2025-08-29 18:01:19.450499 | orchestrator | 2025-08-29 18:01:19.450505 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 18:01:19.450510 | orchestrator | Friday 29 August 2025 18:01:19 +0000 (0:00:00.556) 0:00:09.777 ********* 2025-08-29 18:01:19.450517 | orchestrator | =============================================================================== 2025-08-29 18:01:19.450522 | orchestrator | Gathers facts about hosts ----------------------------------------------- 6.43s 2025-08-29 18:01:19.450528 | orchestrator | osism.commons.facts : Create custom facts directory --------------------- 1.39s 2025-08-29 18:01:19.450535 | orchestrator | osism.commons.facts : Copy fact files ----------------------------------- 1.13s 2025-08-29 18:01:19.450542 | orchestrator | Gather facts for all hosts ---------------------------------------------- 0.56s 2025-08-29 18:01:19.828143 | orchestrator | + osism validate ceph-mons 2025-08-29 18:01:51.867661 | orchestrator | 2025-08-29 18:01:51.867771 | orchestrator | PLAY [Ceph validate mons] ****************************************************** 2025-08-29 18:01:51.867788 | orchestrator | 2025-08-29 18:01:51.867800 | orchestrator | TASK [Get timestamp for report file] ******************************************* 2025-08-29 18:01:51.867812 | orchestrator | Friday 29 August 2025 18:01:36 +0000 (0:00:00.428) 0:00:00.428 ********* 2025-08-29 18:01:51.867824 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:01:51.867836 | orchestrator | 2025-08-29 18:01:51.867847 | orchestrator | TASK [Create report output directory] ****************************************** 2025-08-29 18:01:51.867858 | orchestrator | Friday 29 August 2025 18:01:36 +0000 (0:00:00.650) 0:00:01.079 ********* 2025-08-29 18:01:51.867888 | orchestrator | changed: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:01:51.867899 | orchestrator | 2025-08-29 18:01:51.867911 | orchestrator | TASK [Define report vars] ****************************************************** 2025-08-29 18:01:51.867921 | orchestrator | Friday 29 August 2025 18:01:37 +0000 (0:00:00.903) 0:00:01.982 ********* 2025-08-29 18:01:51.867932 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.867944 | orchestrator | 2025-08-29 18:01:51.867955 | orchestrator | TASK [Prepare test data for container existance test] ************************** 2025-08-29 18:01:51.867966 | orchestrator | Friday 29 August 2025 18:01:37 +0000 (0:00:00.245) 0:00:02.228 ********* 2025-08-29 18:01:51.867977 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.867988 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:01:51.868000 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:01:51.868011 | orchestrator | 2025-08-29 18:01:51.868022 | orchestrator | TASK [Get container info] ****************************************************** 2025-08-29 18:01:51.868033 | orchestrator | Friday 29 August 2025 18:01:38 +0000 (0:00:00.326) 0:00:02.555 ********* 2025-08-29 18:01:51.868044 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:01:51.868055 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.868066 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:01:51.868077 | orchestrator | 2025-08-29 18:01:51.868088 | orchestrator | TASK [Set test result to failed if container is missing] *********************** 2025-08-29 18:01:51.868100 | orchestrator | Friday 29 August 2025 18:01:39 +0000 (0:00:00.981) 0:00:03.536 ********* 2025-08-29 18:01:51.868111 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868122 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:01:51.868133 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:01:51.868144 | orchestrator | 2025-08-29 18:01:51.868155 | orchestrator | TASK [Set test result to passed if container is existing] ********************** 2025-08-29 18:01:51.868166 | orchestrator | Friday 29 August 2025 18:01:39 +0000 (0:00:00.283) 0:00:03.819 ********* 2025-08-29 18:01:51.868177 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.868188 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:01:51.868200 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:01:51.868213 | orchestrator | 2025-08-29 18:01:51.868226 | orchestrator | TASK [Prepare test data] ******************************************************* 2025-08-29 18:01:51.868238 | orchestrator | Friday 29 August 2025 18:01:39 +0000 (0:00:00.450) 0:00:04.270 ********* 2025-08-29 18:01:51.868251 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.868305 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:01:51.868318 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:01:51.868330 | orchestrator | 2025-08-29 18:01:51.868343 | orchestrator | TASK [Set test result to failed if ceph-mon is not running] ******************** 2025-08-29 18:01:51.868356 | orchestrator | Friday 29 August 2025 18:01:40 +0000 (0:00:00.308) 0:00:04.578 ********* 2025-08-29 18:01:51.868369 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868380 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:01:51.868391 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:01:51.868402 | orchestrator | 2025-08-29 18:01:51.868413 | orchestrator | TASK [Set test result to passed if ceph-mon is running] ************************ 2025-08-29 18:01:51.868424 | orchestrator | Friday 29 August 2025 18:01:40 +0000 (0:00:00.281) 0:00:04.860 ********* 2025-08-29 18:01:51.868434 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.868445 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:01:51.868456 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:01:51.868467 | orchestrator | 2025-08-29 18:01:51.868478 | orchestrator | TASK [Aggregate test results step one] ***************************************** 2025-08-29 18:01:51.868489 | orchestrator | Friday 29 August 2025 18:01:40 +0000 (0:00:00.288) 0:00:05.148 ********* 2025-08-29 18:01:51.868500 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868511 | orchestrator | 2025-08-29 18:01:51.868522 | orchestrator | TASK [Aggregate test results step two] ***************************************** 2025-08-29 18:01:51.868533 | orchestrator | Friday 29 August 2025 18:01:41 +0000 (0:00:00.251) 0:00:05.399 ********* 2025-08-29 18:01:51.868544 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868562 | orchestrator | 2025-08-29 18:01:51.868573 | orchestrator | TASK [Aggregate test results step three] *************************************** 2025-08-29 18:01:51.868584 | orchestrator | Friday 29 August 2025 18:01:41 +0000 (0:00:00.664) 0:00:06.064 ********* 2025-08-29 18:01:51.868595 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868606 | orchestrator | 2025-08-29 18:01:51.868616 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:01:51.868627 | orchestrator | Friday 29 August 2025 18:01:42 +0000 (0:00:00.297) 0:00:06.361 ********* 2025-08-29 18:01:51.868638 | orchestrator | 2025-08-29 18:01:51.868649 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:01:51.868660 | orchestrator | Friday 29 August 2025 18:01:42 +0000 (0:00:00.086) 0:00:06.448 ********* 2025-08-29 18:01:51.868671 | orchestrator | 2025-08-29 18:01:51.868682 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:01:51.868693 | orchestrator | Friday 29 August 2025 18:01:42 +0000 (0:00:00.081) 0:00:06.530 ********* 2025-08-29 18:01:51.868704 | orchestrator | 2025-08-29 18:01:51.868715 | orchestrator | TASK [Print report file information] ******************************************* 2025-08-29 18:01:51.868726 | orchestrator | Friday 29 August 2025 18:01:42 +0000 (0:00:00.072) 0:00:06.602 ********* 2025-08-29 18:01:51.868737 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868748 | orchestrator | 2025-08-29 18:01:51.868759 | orchestrator | TASK [Fail due to missing containers] ****************************************** 2025-08-29 18:01:51.868770 | orchestrator | Friday 29 August 2025 18:01:42 +0000 (0:00:00.271) 0:00:06.874 ********* 2025-08-29 18:01:51.868781 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868792 | orchestrator | 2025-08-29 18:01:51.868820 | orchestrator | TASK [Prepare quorum test vars] ************************************************ 2025-08-29 18:01:51.868842 | orchestrator | Friday 29 August 2025 18:01:42 +0000 (0:00:00.235) 0:00:07.110 ********* 2025-08-29 18:01:51.868854 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.868865 | orchestrator | 2025-08-29 18:01:51.868875 | orchestrator | TASK [Get monmap info from one mon container] ********************************** 2025-08-29 18:01:51.868886 | orchestrator | Friday 29 August 2025 18:01:42 +0000 (0:00:00.113) 0:00:07.224 ********* 2025-08-29 18:01:51.868897 | orchestrator | changed: [testbed-node-0] 2025-08-29 18:01:51.868907 | orchestrator | 2025-08-29 18:01:51.868918 | orchestrator | TASK [Set quorum test data] **************************************************** 2025-08-29 18:01:51.868929 | orchestrator | Friday 29 August 2025 18:01:44 +0000 (0:00:01.555) 0:00:08.779 ********* 2025-08-29 18:01:51.868939 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.868950 | orchestrator | 2025-08-29 18:01:51.868960 | orchestrator | TASK [Fail quorum test if not all monitors are in quorum] ********************** 2025-08-29 18:01:51.868971 | orchestrator | Friday 29 August 2025 18:01:44 +0000 (0:00:00.342) 0:00:09.122 ********* 2025-08-29 18:01:51.868981 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.868992 | orchestrator | 2025-08-29 18:01:51.869003 | orchestrator | TASK [Pass quorum test if all monitors are in quorum] ************************** 2025-08-29 18:01:51.869013 | orchestrator | Friday 29 August 2025 18:01:44 +0000 (0:00:00.127) 0:00:09.249 ********* 2025-08-29 18:01:51.869024 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.869035 | orchestrator | 2025-08-29 18:01:51.869049 | orchestrator | TASK [Set fsid test vars] ****************************************************** 2025-08-29 18:01:51.869060 | orchestrator | Friday 29 August 2025 18:01:45 +0000 (0:00:00.528) 0:00:09.778 ********* 2025-08-29 18:01:51.869070 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.869081 | orchestrator | 2025-08-29 18:01:51.869092 | orchestrator | TASK [Fail Cluster FSID test if FSID does not match configuration] ************* 2025-08-29 18:01:51.869102 | orchestrator | Friday 29 August 2025 18:01:45 +0000 (0:00:00.295) 0:00:10.073 ********* 2025-08-29 18:01:51.869113 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.869124 | orchestrator | 2025-08-29 18:01:51.869134 | orchestrator | TASK [Pass Cluster FSID test if it matches configuration] ********************** 2025-08-29 18:01:51.869145 | orchestrator | Friday 29 August 2025 18:01:45 +0000 (0:00:00.133) 0:00:10.206 ********* 2025-08-29 18:01:51.869161 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.869172 | orchestrator | 2025-08-29 18:01:51.869183 | orchestrator | TASK [Prepare status test vars] ************************************************ 2025-08-29 18:01:51.869194 | orchestrator | Friday 29 August 2025 18:01:46 +0000 (0:00:00.116) 0:00:10.323 ********* 2025-08-29 18:01:51.869204 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.869215 | orchestrator | 2025-08-29 18:01:51.869226 | orchestrator | TASK [Gather status data] ****************************************************** 2025-08-29 18:01:51.869237 | orchestrator | Friday 29 August 2025 18:01:46 +0000 (0:00:00.132) 0:00:10.455 ********* 2025-08-29 18:01:51.869247 | orchestrator | changed: [testbed-node-0] 2025-08-29 18:01:51.869258 | orchestrator | 2025-08-29 18:01:51.869289 | orchestrator | TASK [Set health test data] **************************************************** 2025-08-29 18:01:51.869300 | orchestrator | Friday 29 August 2025 18:01:47 +0000 (0:00:01.345) 0:00:11.801 ********* 2025-08-29 18:01:51.869311 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.869321 | orchestrator | 2025-08-29 18:01:51.869332 | orchestrator | TASK [Fail cluster-health if health is not acceptable] ************************* 2025-08-29 18:01:51.869343 | orchestrator | Friday 29 August 2025 18:01:47 +0000 (0:00:00.297) 0:00:12.098 ********* 2025-08-29 18:01:51.869354 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.869364 | orchestrator | 2025-08-29 18:01:51.869375 | orchestrator | TASK [Pass cluster-health if health is acceptable] ***************************** 2025-08-29 18:01:51.869386 | orchestrator | Friday 29 August 2025 18:01:47 +0000 (0:00:00.148) 0:00:12.247 ********* 2025-08-29 18:01:51.869397 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:01:51.869407 | orchestrator | 2025-08-29 18:01:51.869418 | orchestrator | TASK [Fail cluster-health if health is not acceptable (strict)] **************** 2025-08-29 18:01:51.869429 | orchestrator | Friday 29 August 2025 18:01:48 +0000 (0:00:00.168) 0:00:12.415 ********* 2025-08-29 18:01:51.869440 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.869450 | orchestrator | 2025-08-29 18:01:51.869461 | orchestrator | TASK [Pass cluster-health if status is OK (strict)] **************************** 2025-08-29 18:01:51.869472 | orchestrator | Friday 29 August 2025 18:01:48 +0000 (0:00:00.134) 0:00:12.550 ********* 2025-08-29 18:01:51.869483 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.869493 | orchestrator | 2025-08-29 18:01:51.869504 | orchestrator | TASK [Set validation result to passed if no test failed] *********************** 2025-08-29 18:01:51.869515 | orchestrator | Friday 29 August 2025 18:01:48 +0000 (0:00:00.122) 0:00:12.672 ********* 2025-08-29 18:01:51.869525 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:01:51.869537 | orchestrator | 2025-08-29 18:01:51.869548 | orchestrator | TASK [Set validation result to failed if a test failed] ************************ 2025-08-29 18:01:51.869558 | orchestrator | Friday 29 August 2025 18:01:48 +0000 (0:00:00.448) 0:00:13.120 ********* 2025-08-29 18:01:51.869569 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:01:51.869580 | orchestrator | 2025-08-29 18:01:51.869591 | orchestrator | TASK [Aggregate test results step one] ***************************************** 2025-08-29 18:01:51.869601 | orchestrator | Friday 29 August 2025 18:01:49 +0000 (0:00:00.638) 0:00:13.759 ********* 2025-08-29 18:01:51.869612 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:01:51.869623 | orchestrator | 2025-08-29 18:01:51.869634 | orchestrator | TASK [Aggregate test results step two] ***************************************** 2025-08-29 18:01:51.869645 | orchestrator | Friday 29 August 2025 18:01:51 +0000 (0:00:01.653) 0:00:15.412 ********* 2025-08-29 18:01:51.869655 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:01:51.869666 | orchestrator | 2025-08-29 18:01:51.869677 | orchestrator | TASK [Aggregate test results step three] *************************************** 2025-08-29 18:01:51.869687 | orchestrator | Friday 29 August 2025 18:01:51 +0000 (0:00:00.259) 0:00:15.672 ********* 2025-08-29 18:01:51.869698 | orchestrator | changed: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:01:51.869709 | orchestrator | 2025-08-29 18:01:51.869727 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:01:54.028066 | orchestrator | Friday 29 August 2025 18:01:51 +0000 (0:00:00.261) 0:00:15.933 ********* 2025-08-29 18:01:54.028180 | orchestrator | 2025-08-29 18:01:54.028196 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:01:54.028209 | orchestrator | Friday 29 August 2025 18:01:51 +0000 (0:00:00.068) 0:00:16.002 ********* 2025-08-29 18:01:54.028219 | orchestrator | 2025-08-29 18:01:54.028230 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:01:54.028241 | orchestrator | Friday 29 August 2025 18:01:51 +0000 (0:00:00.068) 0:00:16.071 ********* 2025-08-29 18:01:54.028252 | orchestrator | 2025-08-29 18:01:54.028326 | orchestrator | RUNNING HANDLER [Write report file] ******************************************** 2025-08-29 18:01:54.028339 | orchestrator | Friday 29 August 2025 18:01:51 +0000 (0:00:00.073) 0:00:16.144 ********* 2025-08-29 18:01:54.028350 | orchestrator | changed: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:01:54.028361 | orchestrator | 2025-08-29 18:01:54.028372 | orchestrator | TASK [Print report file information] ******************************************* 2025-08-29 18:01:54.028383 | orchestrator | Friday 29 August 2025 18:01:53 +0000 (0:00:01.325) 0:00:17.470 ********* 2025-08-29 18:01:54.028394 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => { 2025-08-29 18:01:54.028405 | orchestrator |  "msg": [ 2025-08-29 18:01:54.028418 | orchestrator |  "Validator run completed.", 2025-08-29 18:01:54.028453 | orchestrator |  "You can find the report file here:", 2025-08-29 18:01:54.028465 | orchestrator |  "/opt/reports/validator/ceph-mons-validator-2025-08-29T18:01:36+00:00-report.json", 2025-08-29 18:01:54.028477 | orchestrator |  "on the following host:", 2025-08-29 18:01:54.028488 | orchestrator |  "testbed-manager" 2025-08-29 18:01:54.028499 | orchestrator |  ] 2025-08-29 18:01:54.028510 | orchestrator | } 2025-08-29 18:01:54.028522 | orchestrator | 2025-08-29 18:01:54.028533 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 18:01:54.028545 | orchestrator | testbed-node-0 : ok=24  changed=5  unreachable=0 failed=0 skipped=13  rescued=0 ignored=0 2025-08-29 18:01:54.028557 | orchestrator | testbed-node-1 : ok=5  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:54.028569 | orchestrator | testbed-node-2 : ok=5  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:01:54.028581 | orchestrator | 2025-08-29 18:01:54.028592 | orchestrator | 2025-08-29 18:01:54.028603 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 18:01:54.028616 | orchestrator | Friday 29 August 2025 18:01:53 +0000 (0:00:00.405) 0:00:17.875 ********* 2025-08-29 18:01:54.028634 | orchestrator | =============================================================================== 2025-08-29 18:01:54.028647 | orchestrator | Aggregate test results step one ----------------------------------------- 1.65s 2025-08-29 18:01:54.028660 | orchestrator | Get monmap info from one mon container ---------------------------------- 1.56s 2025-08-29 18:01:54.028672 | orchestrator | Gather status data ------------------------------------------------------ 1.35s 2025-08-29 18:01:54.028685 | orchestrator | Write report file ------------------------------------------------------- 1.33s 2025-08-29 18:01:54.028697 | orchestrator | Get container info ------------------------------------------------------ 0.98s 2025-08-29 18:01:54.028710 | orchestrator | Create report output directory ------------------------------------------ 0.90s 2025-08-29 18:01:54.028722 | orchestrator | Aggregate test results step two ----------------------------------------- 0.66s 2025-08-29 18:01:54.028734 | orchestrator | Get timestamp for report file ------------------------------------------- 0.65s 2025-08-29 18:01:54.028747 | orchestrator | Set validation result to failed if a test failed ------------------------ 0.64s 2025-08-29 18:01:54.028760 | orchestrator | Pass quorum test if all monitors are in quorum -------------------------- 0.53s 2025-08-29 18:01:54.028797 | orchestrator | Set test result to passed if container is existing ---------------------- 0.45s 2025-08-29 18:01:54.028808 | orchestrator | Set validation result to passed if no test failed ----------------------- 0.45s 2025-08-29 18:01:54.028819 | orchestrator | Print report file information ------------------------------------------- 0.41s 2025-08-29 18:01:54.028830 | orchestrator | Set quorum test data ---------------------------------------------------- 0.34s 2025-08-29 18:01:54.028841 | orchestrator | Prepare test data for container existance test -------------------------- 0.33s 2025-08-29 18:01:54.028851 | orchestrator | Prepare test data ------------------------------------------------------- 0.31s 2025-08-29 18:01:54.028862 | orchestrator | Aggregate test results step three --------------------------------------- 0.30s 2025-08-29 18:01:54.028873 | orchestrator | Set health test data ---------------------------------------------------- 0.30s 2025-08-29 18:01:54.028884 | orchestrator | Set fsid test vars ------------------------------------------------------ 0.30s 2025-08-29 18:01:54.028895 | orchestrator | Set test result to passed if ceph-mon is running ------------------------ 0.29s 2025-08-29 18:01:54.336098 | orchestrator | + osism validate ceph-mgrs 2025-08-29 18:02:25.791126 | orchestrator | 2025-08-29 18:02:25.791244 | orchestrator | PLAY [Ceph validate mgrs] ****************************************************** 2025-08-29 18:02:25.791347 | orchestrator | 2025-08-29 18:02:25.791361 | orchestrator | TASK [Get timestamp for report file] ******************************************* 2025-08-29 18:02:25.791372 | orchestrator | Friday 29 August 2025 18:02:10 +0000 (0:00:00.444) 0:00:00.444 ********* 2025-08-29 18:02:25.791384 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:25.791395 | orchestrator | 2025-08-29 18:02:25.791405 | orchestrator | TASK [Create report output directory] ****************************************** 2025-08-29 18:02:25.791416 | orchestrator | Friday 29 August 2025 18:02:11 +0000 (0:00:00.685) 0:00:01.129 ********* 2025-08-29 18:02:25.791427 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:25.791437 | orchestrator | 2025-08-29 18:02:25.791447 | orchestrator | TASK [Define report vars] ****************************************************** 2025-08-29 18:02:25.791458 | orchestrator | Friday 29 August 2025 18:02:12 +0000 (0:00:00.809) 0:00:01.939 ********* 2025-08-29 18:02:25.791468 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.791480 | orchestrator | 2025-08-29 18:02:25.791491 | orchestrator | TASK [Prepare test data for container existance test] ************************** 2025-08-29 18:02:25.791501 | orchestrator | Friday 29 August 2025 18:02:12 +0000 (0:00:00.250) 0:00:02.189 ********* 2025-08-29 18:02:25.791512 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.791522 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:02:25.791532 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:02:25.791542 | orchestrator | 2025-08-29 18:02:25.791553 | orchestrator | TASK [Get container info] ****************************************************** 2025-08-29 18:02:25.791563 | orchestrator | Friday 29 August 2025 18:02:12 +0000 (0:00:00.286) 0:00:02.475 ********* 2025-08-29 18:02:25.791573 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:02:25.791584 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.791594 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:02:25.791605 | orchestrator | 2025-08-29 18:02:25.791615 | orchestrator | TASK [Set test result to failed if container is missing] *********************** 2025-08-29 18:02:25.791626 | orchestrator | Friday 29 August 2025 18:02:13 +0000 (0:00:00.941) 0:00:03.417 ********* 2025-08-29 18:02:25.791636 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.791647 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:02:25.791657 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:02:25.791668 | orchestrator | 2025-08-29 18:02:25.791678 | orchestrator | TASK [Set test result to passed if container is existing] ********************** 2025-08-29 18:02:25.791689 | orchestrator | Friday 29 August 2025 18:02:14 +0000 (0:00:00.409) 0:00:03.827 ********* 2025-08-29 18:02:25.791700 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.791711 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:02:25.791721 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:02:25.791732 | orchestrator | 2025-08-29 18:02:25.791742 | orchestrator | TASK [Prepare test data] ******************************************************* 2025-08-29 18:02:25.791776 | orchestrator | Friday 29 August 2025 18:02:14 +0000 (0:00:00.489) 0:00:04.316 ********* 2025-08-29 18:02:25.791787 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.791798 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:02:25.791808 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:02:25.791818 | orchestrator | 2025-08-29 18:02:25.791828 | orchestrator | TASK [Set test result to failed if ceph-mgr is not running] ******************** 2025-08-29 18:02:25.791839 | orchestrator | Friday 29 August 2025 18:02:15 +0000 (0:00:00.309) 0:00:04.626 ********* 2025-08-29 18:02:25.791849 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.791859 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:02:25.791869 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:02:25.791880 | orchestrator | 2025-08-29 18:02:25.791890 | orchestrator | TASK [Set test result to passed if ceph-mgr is running] ************************ 2025-08-29 18:02:25.791901 | orchestrator | Friday 29 August 2025 18:02:15 +0000 (0:00:00.301) 0:00:04.927 ********* 2025-08-29 18:02:25.791911 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.791922 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:02:25.791933 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:02:25.791945 | orchestrator | 2025-08-29 18:02:25.791977 | orchestrator | TASK [Aggregate test results step one] ***************************************** 2025-08-29 18:02:25.791990 | orchestrator | Friday 29 August 2025 18:02:15 +0000 (0:00:00.345) 0:00:05.272 ********* 2025-08-29 18:02:25.792003 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.792014 | orchestrator | 2025-08-29 18:02:25.792026 | orchestrator | TASK [Aggregate test results step two] ***************************************** 2025-08-29 18:02:25.792035 | orchestrator | Friday 29 August 2025 18:02:15 +0000 (0:00:00.239) 0:00:05.512 ********* 2025-08-29 18:02:25.792043 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.792051 | orchestrator | 2025-08-29 18:02:25.792057 | orchestrator | TASK [Aggregate test results step three] *************************************** 2025-08-29 18:02:25.792064 | orchestrator | Friday 29 August 2025 18:02:16 +0000 (0:00:00.631) 0:00:06.144 ********* 2025-08-29 18:02:25.792071 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.792077 | orchestrator | 2025-08-29 18:02:25.792084 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:25.792090 | orchestrator | Friday 29 August 2025 18:02:16 +0000 (0:00:00.239) 0:00:06.384 ********* 2025-08-29 18:02:25.792097 | orchestrator | 2025-08-29 18:02:25.792104 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:25.792110 | orchestrator | Friday 29 August 2025 18:02:16 +0000 (0:00:00.067) 0:00:06.451 ********* 2025-08-29 18:02:25.792117 | orchestrator | 2025-08-29 18:02:25.792123 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:25.792135 | orchestrator | Friday 29 August 2025 18:02:16 +0000 (0:00:00.066) 0:00:06.518 ********* 2025-08-29 18:02:25.792146 | orchestrator | 2025-08-29 18:02:25.792156 | orchestrator | TASK [Print report file information] ******************************************* 2025-08-29 18:02:25.792167 | orchestrator | Friday 29 August 2025 18:02:16 +0000 (0:00:00.071) 0:00:06.589 ********* 2025-08-29 18:02:25.792178 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.792189 | orchestrator | 2025-08-29 18:02:25.792200 | orchestrator | TASK [Fail due to missing containers] ****************************************** 2025-08-29 18:02:25.792211 | orchestrator | Friday 29 August 2025 18:02:17 +0000 (0:00:00.275) 0:00:06.864 ********* 2025-08-29 18:02:25.792222 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.792233 | orchestrator | 2025-08-29 18:02:25.792284 | orchestrator | TASK [Define mgr module test vars] ********************************************* 2025-08-29 18:02:25.792297 | orchestrator | Friday 29 August 2025 18:02:17 +0000 (0:00:00.309) 0:00:07.174 ********* 2025-08-29 18:02:25.792308 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.792319 | orchestrator | 2025-08-29 18:02:25.792330 | orchestrator | TASK [Gather list of mgr modules] ********************************************** 2025-08-29 18:02:25.792340 | orchestrator | Friday 29 August 2025 18:02:17 +0000 (0:00:00.113) 0:00:07.287 ********* 2025-08-29 18:02:25.792360 | orchestrator | changed: [testbed-node-0] 2025-08-29 18:02:25.792370 | orchestrator | 2025-08-29 18:02:25.792381 | orchestrator | TASK [Parse mgr module list from json] ***************************************** 2025-08-29 18:02:25.792387 | orchestrator | Friday 29 August 2025 18:02:19 +0000 (0:00:02.144) 0:00:09.432 ********* 2025-08-29 18:02:25.792394 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.792400 | orchestrator | 2025-08-29 18:02:25.792406 | orchestrator | TASK [Extract list of enabled mgr modules] ************************************* 2025-08-29 18:02:25.792413 | orchestrator | Friday 29 August 2025 18:02:20 +0000 (0:00:00.249) 0:00:09.682 ********* 2025-08-29 18:02:25.792419 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.792425 | orchestrator | 2025-08-29 18:02:25.792431 | orchestrator | TASK [Fail test if mgr modules are disabled that should be enabled] ************ 2025-08-29 18:02:25.792437 | orchestrator | Friday 29 August 2025 18:02:20 +0000 (0:00:00.309) 0:00:09.991 ********* 2025-08-29 18:02:25.792443 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.792449 | orchestrator | 2025-08-29 18:02:25.792456 | orchestrator | TASK [Pass test if required mgr modules are enabled] *************************** 2025-08-29 18:02:25.792462 | orchestrator | Friday 29 August 2025 18:02:20 +0000 (0:00:00.311) 0:00:10.303 ********* 2025-08-29 18:02:25.792468 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:02:25.792474 | orchestrator | 2025-08-29 18:02:25.792480 | orchestrator | TASK [Set validation result to passed if no test failed] *********************** 2025-08-29 18:02:25.792486 | orchestrator | Friday 29 August 2025 18:02:20 +0000 (0:00:00.174) 0:00:10.478 ********* 2025-08-29 18:02:25.792492 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:25.792499 | orchestrator | 2025-08-29 18:02:25.792509 | orchestrator | TASK [Set validation result to failed if a test failed] ************************ 2025-08-29 18:02:25.792515 | orchestrator | Friday 29 August 2025 18:02:21 +0000 (0:00:00.258) 0:00:10.736 ********* 2025-08-29 18:02:25.792522 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:02:25.792528 | orchestrator | 2025-08-29 18:02:25.792534 | orchestrator | TASK [Aggregate test results step one] ***************************************** 2025-08-29 18:02:25.792540 | orchestrator | Friday 29 August 2025 18:02:21 +0000 (0:00:00.236) 0:00:10.972 ********* 2025-08-29 18:02:25.792546 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:25.792552 | orchestrator | 2025-08-29 18:02:25.792558 | orchestrator | TASK [Aggregate test results step two] ***************************************** 2025-08-29 18:02:25.792565 | orchestrator | Friday 29 August 2025 18:02:22 +0000 (0:00:01.230) 0:00:12.202 ********* 2025-08-29 18:02:25.792575 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:25.792585 | orchestrator | 2025-08-29 18:02:25.792595 | orchestrator | TASK [Aggregate test results step three] *************************************** 2025-08-29 18:02:25.792605 | orchestrator | Friday 29 August 2025 18:02:22 +0000 (0:00:00.313) 0:00:12.515 ********* 2025-08-29 18:02:25.792614 | orchestrator | changed: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:25.792624 | orchestrator | 2025-08-29 18:02:25.792634 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:25.792643 | orchestrator | Friday 29 August 2025 18:02:23 +0000 (0:00:00.254) 0:00:12.769 ********* 2025-08-29 18:02:25.792653 | orchestrator | 2025-08-29 18:02:25.792662 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:25.792672 | orchestrator | Friday 29 August 2025 18:02:23 +0000 (0:00:00.069) 0:00:12.839 ********* 2025-08-29 18:02:25.792681 | orchestrator | 2025-08-29 18:02:25.792691 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:25.792700 | orchestrator | Friday 29 August 2025 18:02:23 +0000 (0:00:00.069) 0:00:12.908 ********* 2025-08-29 18:02:25.792710 | orchestrator | 2025-08-29 18:02:25.792719 | orchestrator | RUNNING HANDLER [Write report file] ******************************************** 2025-08-29 18:02:25.792729 | orchestrator | Friday 29 August 2025 18:02:23 +0000 (0:00:00.075) 0:00:12.983 ********* 2025-08-29 18:02:25.792745 | orchestrator | changed: [testbed-node-0 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:25.792755 | orchestrator | 2025-08-29 18:02:25.792764 | orchestrator | TASK [Print report file information] ******************************************* 2025-08-29 18:02:25.792773 | orchestrator | Friday 29 August 2025 18:02:24 +0000 (0:00:01.500) 0:00:14.484 ********* 2025-08-29 18:02:25.792783 | orchestrator | ok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => { 2025-08-29 18:02:25.792793 | orchestrator |  "msg": [ 2025-08-29 18:02:25.792803 | orchestrator |  "Validator run completed.", 2025-08-29 18:02:25.792813 | orchestrator |  "You can find the report file here:", 2025-08-29 18:02:25.792822 | orchestrator |  "/opt/reports/validator/ceph-mgrs-validator-2025-08-29T18:02:11+00:00-report.json", 2025-08-29 18:02:25.792833 | orchestrator |  "on the following host:", 2025-08-29 18:02:25.792843 | orchestrator |  "testbed-manager" 2025-08-29 18:02:25.792853 | orchestrator |  ] 2025-08-29 18:02:25.792863 | orchestrator | } 2025-08-29 18:02:25.792872 | orchestrator | 2025-08-29 18:02:25.792882 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 18:02:25.792893 | orchestrator | testbed-node-0 : ok=19  changed=3  unreachable=0 failed=0 skipped=9  rescued=0 ignored=0 2025-08-29 18:02:25.792903 | orchestrator | testbed-node-1 : ok=5  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:02:25.792920 | orchestrator | testbed-node-2 : ok=5  changed=0 unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:02:26.106956 | orchestrator | 2025-08-29 18:02:26.107051 | orchestrator | 2025-08-29 18:02:26.107065 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 18:02:26.107079 | orchestrator | Friday 29 August 2025 18:02:25 +0000 (0:00:00.902) 0:00:15.386 ********* 2025-08-29 18:02:26.107090 | orchestrator | =============================================================================== 2025-08-29 18:02:26.107101 | orchestrator | Gather list of mgr modules ---------------------------------------------- 2.14s 2025-08-29 18:02:26.107112 | orchestrator | Write report file ------------------------------------------------------- 1.50s 2025-08-29 18:02:26.107123 | orchestrator | Aggregate test results step one ----------------------------------------- 1.23s 2025-08-29 18:02:26.107133 | orchestrator | Get container info ------------------------------------------------------ 0.94s 2025-08-29 18:02:26.107144 | orchestrator | Print report file information ------------------------------------------- 0.90s 2025-08-29 18:02:26.107155 | orchestrator | Create report output directory ------------------------------------------ 0.81s 2025-08-29 18:02:26.107166 | orchestrator | Get timestamp for report file ------------------------------------------- 0.69s 2025-08-29 18:02:26.107176 | orchestrator | Aggregate test results step two ----------------------------------------- 0.63s 2025-08-29 18:02:26.107187 | orchestrator | Set test result to passed if container is existing ---------------------- 0.49s 2025-08-29 18:02:26.107197 | orchestrator | Set test result to failed if container is missing ----------------------- 0.41s 2025-08-29 18:02:26.107208 | orchestrator | Set test result to passed if ceph-mgr is running ------------------------ 0.35s 2025-08-29 18:02:26.107219 | orchestrator | Aggregate test results step two ----------------------------------------- 0.31s 2025-08-29 18:02:26.107230 | orchestrator | Fail test if mgr modules are disabled that should be enabled ------------ 0.31s 2025-08-29 18:02:26.107241 | orchestrator | Fail due to missing containers ------------------------------------------ 0.31s 2025-08-29 18:02:26.107309 | orchestrator | Prepare test data ------------------------------------------------------- 0.31s 2025-08-29 18:02:26.107323 | orchestrator | Extract list of enabled mgr modules ------------------------------------- 0.31s 2025-08-29 18:02:26.107334 | orchestrator | Set test result to failed if ceph-mgr is not running -------------------- 0.30s 2025-08-29 18:02:26.107344 | orchestrator | Prepare test data for container existance test -------------------------- 0.29s 2025-08-29 18:02:26.107355 | orchestrator | Print report file information ------------------------------------------- 0.28s 2025-08-29 18:02:26.107389 | orchestrator | Set validation result to passed if no test failed ----------------------- 0.26s 2025-08-29 18:02:26.417417 | orchestrator | + osism validate ceph-osds 2025-08-29 18:02:36.925697 | orchestrator | 2025-08-29 18:02:36.925777 | orchestrator | PLAY [Ceph validate OSDs] ****************************************************** 2025-08-29 18:02:36.925784 | orchestrator | 2025-08-29 18:02:36.925789 | orchestrator | TASK [Get timestamp for report file] ******************************************* 2025-08-29 18:02:36.925793 | orchestrator | Friday 29 August 2025 18:02:32 +0000 (0:00:00.415) 0:00:00.415 ********* 2025-08-29 18:02:36.925798 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:36.925802 | orchestrator | 2025-08-29 18:02:36.925806 | orchestrator | TASK [Get extra vars for Ceph configuration] *********************************** 2025-08-29 18:02:36.925811 | orchestrator | Friday 29 August 2025 18:02:33 +0000 (0:00:00.628) 0:00:01.044 ********* 2025-08-29 18:02:36.925815 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:36.925819 | orchestrator | 2025-08-29 18:02:36.925822 | orchestrator | TASK [Create report output directory] ****************************************** 2025-08-29 18:02:36.925826 | orchestrator | Friday 29 August 2025 18:02:33 +0000 (0:00:00.252) 0:00:01.296 ********* 2025-08-29 18:02:36.925830 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:36.925834 | orchestrator | 2025-08-29 18:02:36.925838 | orchestrator | TASK [Define report vars] ****************************************************** 2025-08-29 18:02:36.925841 | orchestrator | Friday 29 August 2025 18:02:34 +0000 (0:00:01.013) 0:00:02.310 ********* 2025-08-29 18:02:36.925845 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:36.925850 | orchestrator | 2025-08-29 18:02:36.925854 | orchestrator | TASK [Define OSD test variables] *********************************************** 2025-08-29 18:02:36.925858 | orchestrator | Friday 29 August 2025 18:02:34 +0000 (0:00:00.111) 0:00:02.422 ********* 2025-08-29 18:02:36.925862 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:36.925866 | orchestrator | 2025-08-29 18:02:36.925869 | orchestrator | TASK [Calculate OSD devices for each host] ************************************* 2025-08-29 18:02:36.925873 | orchestrator | Friday 29 August 2025 18:02:34 +0000 (0:00:00.132) 0:00:02.554 ********* 2025-08-29 18:02:36.925877 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:36.925881 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:36.925885 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:36.925888 | orchestrator | 2025-08-29 18:02:36.925892 | orchestrator | TASK [Define OSD test variables] *********************************************** 2025-08-29 18:02:36.925896 | orchestrator | Friday 29 August 2025 18:02:35 +0000 (0:00:00.301) 0:00:02.856 ********* 2025-08-29 18:02:36.925900 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:36.925903 | orchestrator | 2025-08-29 18:02:36.925907 | orchestrator | TASK [Calculate OSD devices for each host] ************************************* 2025-08-29 18:02:36.925911 | orchestrator | Friday 29 August 2025 18:02:35 +0000 (0:00:00.147) 0:00:03.003 ********* 2025-08-29 18:02:36.925915 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:36.925919 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:36.925922 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:36.925926 | orchestrator | 2025-08-29 18:02:36.925930 | orchestrator | TASK [Calculate total number of OSDs in cluster] ******************************* 2025-08-29 18:02:36.925934 | orchestrator | Friday 29 August 2025 18:02:35 +0000 (0:00:00.304) 0:00:03.308 ********* 2025-08-29 18:02:36.925938 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:36.925941 | orchestrator | 2025-08-29 18:02:36.925945 | orchestrator | TASK [Prepare test data] ******************************************************* 2025-08-29 18:02:36.925949 | orchestrator | Friday 29 August 2025 18:02:36 +0000 (0:00:00.552) 0:00:03.861 ********* 2025-08-29 18:02:36.925953 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:36.925957 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:36.925960 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:36.925964 | orchestrator | 2025-08-29 18:02:36.925968 | orchestrator | TASK [Get list of ceph-osd containers on host] ********************************* 2025-08-29 18:02:36.925988 | orchestrator | Friday 29 August 2025 18:02:36 +0000 (0:00:00.449) 0:00:04.310 ********* 2025-08-29 18:02:36.925994 | orchestrator | skipping: [testbed-node-3] => (item={'id': '831db196fd44b17afa950ddeca82e1c84858cd7b85a187549ea1c16eb2984c3f', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'})  2025-08-29 18:02:36.926000 | orchestrator | skipping: [testbed-node-3] => (item={'id': '43db33169d13aa0428ead7c8bcc33f3ddcb651e2d39ed645855f608e2e3716c6', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'})  2025-08-29 18:02:36.926004 | orchestrator | skipping: [testbed-node-3] => (item={'id': 'a6a6d2983ce44971aeff502c04cac16d83e8942efa28d0972e703b077be17772', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'})  2025-08-29 18:02:36.926010 | orchestrator | skipping: [testbed-node-3] => (item={'id': 'dd2bfca62512c71e4842410b8b7da7854e9fc2e705799e091144b9eaf0b19a41', 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'name': '/cinder_backup', 'state': 'running', 'status': 'Up 10 minutes (healthy)'})  2025-08-29 18:02:36.926042 | orchestrator | skipping: [testbed-node-3] => (item={'id': '1f8d386c333791106ae35223308f5a1d2c26c745df436379937ca62f74794df6', 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'name': '/cinder_volume', 'state': 'running', 'status': 'Up 10 minutes (healthy)'})  2025-08-29 18:02:36.926068 | orchestrator | skipping: [testbed-node-3] => (item={'id': 'dc265fa1a6e7d9e4cce4e74f6247812b212ef88825d2d04f50d462f0f59f94ee', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'})  2025-08-29 18:02:36.926092 | orchestrator | skipping: [testbed-node-3] => (item={'id': '73d1f740374ac3a6370ee655c432cf79f1b0497496f696a24b3d82f10d5db734', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'})  2025-08-29 18:02:36.926101 | orchestrator | skipping: [testbed-node-3] => (item={'id': 'fbf07d662364d8ba249dc2fb1a3756692e5f33b9d97e57598ea52652ee35f3d3', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'})  2025-08-29 18:02:36.926107 | orchestrator | skipping: [testbed-node-3] => (item={'id': '88d0d067ba7ed979832971f0beee9ce14bcb2ce5b1fecfe4036184d7be0802b8', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'})  2025-08-29 18:02:36.926111 | orchestrator | skipping: [testbed-node-3] => (item={'id': '9415c637c14f836422d459140629614d2adf6fa0f1b18dfebd926f3908fe2c35', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 22 minutes'})  2025-08-29 18:02:36.926116 | orchestrator | skipping: [testbed-node-3] => (item={'id': '70d277b5c802e1dbd7ab5b98adb60818042b27e904da19c0009f3dc7c2cc9a74', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 23 minutes'})  2025-08-29 18:02:36.926120 | orchestrator | skipping: [testbed-node-3] => (item={'id': 'aa300e7c318b46c149df5fe127c223ccb55d102ef8592384517488090dab05aa', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 24 minutes'})  2025-08-29 18:02:36.926125 | orchestrator | ok: [testbed-node-3] => (item={'id': '97d656564b3848ce24ec894d05d57bd6acb81b9aeb818630d899dc8615d4bde2', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 25 minutes'}) 2025-08-29 18:02:36.926135 | orchestrator | ok: [testbed-node-3] => (item={'id': 'f0edcf22f4d3fdb79bd3256316c29865cc8ba2253f3eddabfa710ce3e58e6729', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 25 minutes'}) 2025-08-29 18:02:36.926139 | orchestrator | skipping: [testbed-node-3] => (item={'id': '8b7600d7a6e52b5b10cb9da125feb22c9b0122ab84899c1c3dfba68e876c4a44', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'})  2025-08-29 18:02:36.926143 | orchestrator | skipping: [testbed-node-3] => (item={'id': '8813a89613324d51a687a59295318f960d33da26d0d17a5e784b97fad705e3fc', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'})  2025-08-29 18:02:36.926147 | orchestrator | skipping: [testbed-node-3] => (item={'id': '371c0de7a12e8350a1c931282b4a9869c51e79cad275474305bfaeabf585fc5e', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 30 minutes (healthy)'})  2025-08-29 18:02:36.926151 | orchestrator | skipping: [testbed-node-3] => (item={'id': '42140f71ed33fb98035a61e94b9b2dde95c0d0c91d09e64387e90674b021eb92', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'name': '/cron', 'state': 'running', 'status': 'Up 31 minutes'})  2025-08-29 18:02:36.926158 | orchestrator | skipping: [testbed-node-3] => (item={'id': '3694d7914e00ae4b2bea57975063ca5a51d7ba51c46bb4950c9a882c171417b7', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 31 minutes'})  2025-08-29 18:02:36.926162 | orchestrator | skipping: [testbed-node-3] => (item={'id': '1c6d8c66f691a53905d71e5dbe58616ebd908599313cb814da68fd7a4c1f5119', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'name': '/fluentd', 'state': 'running', 'status': 'Up 32 minutes'})  2025-08-29 18:02:36.926169 | orchestrator | skipping: [testbed-node-4] => (item={'id': '6b3770b63ae6b01da9c8fc1b6ea7053906cb7f39f88aa60b41710e5ba222e175', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'})  2025-08-29 18:02:37.200122 | orchestrator | skipping: [testbed-node-4] => (item={'id': '4ad5568d7b6b2abcc6468e1a63dd1e5a7849660f98b7ddc6255e3eca54ab795e', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'})  2025-08-29 18:02:37.200206 | orchestrator | skipping: [testbed-node-4] => (item={'id': '73e1e400e81637d8033c13ad572ca1f2302d78000b87204b756fd2abea510e9e', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'})  2025-08-29 18:02:37.200218 | orchestrator | skipping: [testbed-node-4] => (item={'id': '2bf6f8c514c9889579c47fcabb4cabd8430f65206aa4c5fbd0c9b9d99cd6bd31', 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'name': '/cinder_backup', 'state': 'running', 'status': 'Up 10 minutes (healthy)'})  2025-08-29 18:02:37.200225 | orchestrator | skipping: [testbed-node-4] => (item={'id': 'a0d4fd32a671c9e93c13cbc9a86f866ad3e4c5c5bd7c2e563524c623f37881ae', 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'name': '/cinder_volume', 'state': 'running', 'status': 'Up 11 minutes (healthy)'})  2025-08-29 18:02:37.200232 | orchestrator | skipping: [testbed-node-4] => (item={'id': '9dbddffeb8c3cccd7ff2932c41c3d817c5024dc2b6d2b6aee511b50ceb347bc0', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'})  2025-08-29 18:02:37.200241 | orchestrator | skipping: [testbed-node-4] => (item={'id': '48f451e40ebbee1e40012274f79cffb90856a2717bcba2192d60e5e78920536a', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'})  2025-08-29 18:02:37.200311 | orchestrator | skipping: [testbed-node-4] => (item={'id': '2188a27afcc7abbede1f0fbbcd50bab399e6b8b35615d096ae8d44c09ca22947', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'})  2025-08-29 18:02:37.200322 | orchestrator | skipping: [testbed-node-4] => (item={'id': '533fffb4cc79d90ca2fbe3b83bd49d4d39f24d9e4c4d3738b9d0bc74ad327c40', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'})  2025-08-29 18:02:37.200329 | orchestrator | skipping: [testbed-node-4] => (item={'id': '96dda61ad0ab6d9c345fcea72006b365d615d75ec7930b4afb6213a44dfe5f1a', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 22 minutes'})  2025-08-29 18:02:37.200338 | orchestrator | skipping: [testbed-node-4] => (item={'id': '26794a61ef3b0c34052c4c8a7a24a88c05ba4f5dfd006a81ec894c5e62aa429a', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 23 minutes'})  2025-08-29 18:02:37.200346 | orchestrator | skipping: [testbed-node-4] => (item={'id': '9780ec9c20c72a0eecb3094462974806cb2469ff155fa71eb67225aa020c38f3', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 24 minutes'})  2025-08-29 18:02:37.200356 | orchestrator | ok: [testbed-node-4] => (item={'id': 'bb2389a531af30573b57ea1371fcafcb41cdce0a3dc8926fd77334fa9f59a5d9', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 25 minutes'}) 2025-08-29 18:02:37.200376 | orchestrator | ok: [testbed-node-4] => (item={'id': '83d938f10495e87503fef96df4c7fddc8a59c60ad29a2eb5e2c0dbf82bde4216', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 25 minutes'}) 2025-08-29 18:02:37.200385 | orchestrator | skipping: [testbed-node-4] => (item={'id': '819e56d2b4289e61d8fe2c184498f67228e38874a8d7e017801ee5775bba04ec', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'})  2025-08-29 18:02:37.200407 | orchestrator | skipping: [testbed-node-4] => (item={'id': 'cbe6626888e55714a88a09825bfcbb2ac80b9072f425a74e45f4e17bc9ed3920', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'})  2025-08-29 18:02:37.200415 | orchestrator | skipping: [testbed-node-4] => (item={'id': '47246fb53d0833b8672ad50f1944670303c2040dae792b317ef987fc77418bc9', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 30 minutes (healthy)'})  2025-08-29 18:02:37.200423 | orchestrator | skipping: [testbed-node-4] => (item={'id': '445853a2786ee3a7ab98e10bf3294d285adeb42ced6966c72c50e7a76cd5a4c4', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'name': '/cron', 'state': 'running', 'status': 'Up 31 minutes'})  2025-08-29 18:02:37.200430 | orchestrator | skipping: [testbed-node-4] => (item={'id': '8337624f2559c70025e33fe48ee0a60c8944f2a8f5c5b63419582b1cd5aa509c', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 31 minutes'})  2025-08-29 18:02:37.200438 | orchestrator | skipping: [testbed-node-4] => (item={'id': '521eeb9b19510a1e067a5103783d2e383141e4085e873dbd8278a5d3af2620c1', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'name': '/fluentd', 'state': 'running', 'status': 'Up 32 minutes'})  2025-08-29 18:02:37.200452 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'c6a5f985a6d6783bb89a2acc7d39954c4636c4f44cbc900da1b4929d1777d035', 'image': 'registry.osism.tech/kolla/release/nova-compute:30.0.1.20250711', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 6 minutes (healthy)'})  2025-08-29 18:02:37.200460 | orchestrator | skipping: [testbed-node-5] => (item={'id': '8d74968abfc5e746b2713da3b8c98b2e61e3b35dbed043d775c329030fe196f1', 'image': 'registry.osism.tech/kolla/release/nova-libvirt:9.0.0.20250711', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 6 minutes (healthy)'})  2025-08-29 18:02:37.200468 | orchestrator | skipping: [testbed-node-5] => (item={'id': '2a3c04001189548aaedcbba046c2aceca0705929bbdb5627f52e0b69ed8fa622', 'image': 'registry.osism.tech/kolla/release/nova-ssh:30.0.1.20250711', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 7 minutes (healthy)'})  2025-08-29 18:02:37.200475 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'aa5262d0ba1e1d29a6db48425c613d8289af2e0c55857dc8f120e7dc719c75b9', 'image': 'registry.osism.tech/kolla/release/cinder-backup:25.2.1.20250711', 'name': '/cinder_backup', 'state': 'running', 'status': 'Up 10 minutes (healthy)'})  2025-08-29 18:02:37.200483 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'f14ccdb0d4be60d1ce247661bc77143ab2c35c4237208015067342bcb5d27e38', 'image': 'registry.osism.tech/kolla/release/cinder-volume:25.2.1.20250711', 'name': '/cinder_volume', 'state': 'running', 'status': 'Up 11 minutes (healthy)'})  2025-08-29 18:02:37.200490 | orchestrator | skipping: [testbed-node-5] => (item={'id': '71402882281c025e3583274db5c6945e2b33ec02fcc73e15a9b778448bc53d2b', 'image': 'registry.osism.tech/kolla/release/prometheus-libvirt-exporter:0.20250711.0.20250711', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 12 minutes'})  2025-08-29 18:02:37.200498 | orchestrator | skipping: [testbed-node-5] => (item={'id': '7dddd71ee8a9ebfb2739d83cbc27898444d39847585268d35784165dc2853fd5', 'image': 'registry.osism.tech/kolla/release/prometheus-cadvisor:0.49.2.20250711', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'})  2025-08-29 18:02:37.200505 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'c7974fb7b40433d6f7e454c77471b484b6f88690accf4a02a80a238d75d63446', 'image': 'registry.osism.tech/kolla/release/neutron-metadata-agent:25.2.1.20250711', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 13 minutes (healthy)'})  2025-08-29 18:02:37.200513 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'e323ea688a93346967e666e7d3f435bdb0fa0ba37e0a5230d2f0e0d2a468ef36', 'image': 'registry.osism.tech/kolla/release/prometheus-node-exporter:1.8.2.20250711', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 13 minutes'})  2025-08-29 18:02:37.200520 | orchestrator | skipping: [testbed-node-5] => (item={'id': '5843a8c7c2dbbfea80b706b7e863b9da0f850169d420e13e2e93b2dad8146bde', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 22 minutes'})  2025-08-29 18:02:37.200532 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'b64257a1f29d8a5c9545ea0a6495ef9fd9c5ddf6d5d84c23aa5339c465e1c9db', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 23 minutes'})  2025-08-29 18:02:44.567439 | orchestrator | skipping: [testbed-node-5] => (item={'id': '29156ef1b9db93c9f2d3dd80ba0a8becff853f6359ac46ee86e87fccaf7ddd9a', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 24 minutes'})  2025-08-29 18:02:44.567549 | orchestrator | ok: [testbed-node-5] => (item={'id': '47bafb143a264df0934c8490e7d28dadee83465a4ecaf9db300fb88b2ac596bb', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 25 minutes'}) 2025-08-29 18:02:44.567587 | orchestrator | ok: [testbed-node-5] => (item={'id': '49873dde383ea466b5ed05a4c6f06f749e3c7fb2434a5a485d1a509b219cf7fc', 'image': 'registry.osism.tech/osism/ceph-daemon:18.2.7', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 25 minutes'}) 2025-08-29 18:02:44.567626 | orchestrator | skipping: [testbed-node-5] => (item={'id': '54100880ab97fb343bcc44578ddcbc6d3390dcebf53ac5b486f141c3ec29479c', 'image': 'registry.osism.tech/kolla/release/ovn-controller:24.9.2.20250711', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 28 minutes'})  2025-08-29 18:02:44.567640 | orchestrator | skipping: [testbed-node-5] => (item={'id': '2c57bd67d9ce3746c4aae70a4047fa49a90c5f09a0bf4bc9ac3b68357a8bac29', 'image': 'registry.osism.tech/kolla/release/openvswitch-vswitchd:3.4.2.20250711', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 29 minutes (healthy)'})  2025-08-29 18:02:44.567653 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'a4a4d21c549b33006ae9e03556b170992662a8e6aac8b223b720093dfd9bc4a3', 'image': 'registry.osism.tech/kolla/release/openvswitch-db-server:3.4.2.20250711', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 30 minutes (healthy)'})  2025-08-29 18:02:44.567665 | orchestrator | skipping: [testbed-node-5] => (item={'id': '12e159d426d157159ff64c019529d774f3827585a428e6982aea356ca61bfd00', 'image': 'registry.osism.tech/kolla/release/cron:3.0.20250711', 'name': '/cron', 'state': 'running', 'status': 'Up 31 minutes'})  2025-08-29 18:02:44.567677 | orchestrator | skipping: [testbed-node-5] => (item={'id': 'd40952b6843dcc838fd0b1d4e8a47e41403a93b5c64dd9ac9d7f399ca68b2296', 'image': 'registry.osism.tech/kolla/release/kolla-toolbox:19.5.1.20250711', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 31 minutes'})  2025-08-29 18:02:44.567688 | orchestrator | skipping: [testbed-node-5] => (item={'id': '0ab264b03cf832c0569786ee66f70a448fbc7fa900c359bce6f20a9aafa7829b', 'image': 'registry.osism.tech/kolla/release/fluentd:5.0.7.20250711', 'name': '/fluentd', 'state': 'running', 'status': 'Up 32 minutes'})  2025-08-29 18:02:44.567699 | orchestrator | 2025-08-29 18:02:44.567712 | orchestrator | TASK [Get count of ceph-osd containers on host] ******************************** 2025-08-29 18:02:44.567725 | orchestrator | Friday 29 August 2025 18:02:37 +0000 (0:00:00.566) 0:00:04.876 ********* 2025-08-29 18:02:44.567736 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.567747 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:44.567758 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:44.567769 | orchestrator | 2025-08-29 18:02:44.567780 | orchestrator | TASK [Set test result to failed when count of containers is wrong] ************* 2025-08-29 18:02:44.567791 | orchestrator | Friday 29 August 2025 18:02:37 +0000 (0:00:00.299) 0:00:05.176 ********* 2025-08-29 18:02:44.567802 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.567814 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:44.567825 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:44.567836 | orchestrator | 2025-08-29 18:02:44.567847 | orchestrator | TASK [Set test result to passed if count matches] ****************************** 2025-08-29 18:02:44.567858 | orchestrator | Friday 29 August 2025 18:02:37 +0000 (0:00:00.294) 0:00:05.471 ********* 2025-08-29 18:02:44.567869 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.567880 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:44.567891 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:44.567902 | orchestrator | 2025-08-29 18:02:44.567913 | orchestrator | TASK [Prepare test data] ******************************************************* 2025-08-29 18:02:44.567929 | orchestrator | Friday 29 August 2025 18:02:38 +0000 (0:00:00.492) 0:00:05.963 ********* 2025-08-29 18:02:44.567940 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.567951 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:44.567962 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:44.567972 | orchestrator | 2025-08-29 18:02:44.567983 | orchestrator | TASK [Get list of ceph-osd containers that are not running] ******************** 2025-08-29 18:02:44.567994 | orchestrator | Friday 29 August 2025 18:02:38 +0000 (0:00:00.284) 0:00:06.247 ********* 2025-08-29 18:02:44.568005 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'})  2025-08-29 18:02:44.568025 | orchestrator | skipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'})  2025-08-29 18:02:44.568037 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568048 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'})  2025-08-29 18:02:44.568059 | orchestrator | skipping: [testbed-node-4] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'})  2025-08-29 18:02:44.568088 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:44.568100 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'})  2025-08-29 18:02:44.568111 | orchestrator | skipping: [testbed-node-5] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'})  2025-08-29 18:02:44.568122 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:44.568133 | orchestrator | 2025-08-29 18:02:44.568144 | orchestrator | TASK [Get count of ceph-osd containers that are not running] ******************* 2025-08-29 18:02:44.568155 | orchestrator | Friday 29 August 2025 18:02:38 +0000 (0:00:00.305) 0:00:06.552 ********* 2025-08-29 18:02:44.568166 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.568177 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:44.568187 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:44.568198 | orchestrator | 2025-08-29 18:02:44.568209 | orchestrator | TASK [Set test result to failed if an OSD is not running] ********************** 2025-08-29 18:02:44.568220 | orchestrator | Friday 29 August 2025 18:02:39 +0000 (0:00:00.298) 0:00:06.851 ********* 2025-08-29 18:02:44.568231 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568242 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:44.568253 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:44.568292 | orchestrator | 2025-08-29 18:02:44.568305 | orchestrator | TASK [Set test result to failed if an OSD is not running] ********************** 2025-08-29 18:02:44.568316 | orchestrator | Friday 29 August 2025 18:02:39 +0000 (0:00:00.464) 0:00:07.315 ********* 2025-08-29 18:02:44.568327 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568337 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:44.568348 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:44.568359 | orchestrator | 2025-08-29 18:02:44.568370 | orchestrator | TASK [Set test result to passed if all containers are running] ***************** 2025-08-29 18:02:44.568381 | orchestrator | Friday 29 August 2025 18:02:39 +0000 (0:00:00.280) 0:00:07.596 ********* 2025-08-29 18:02:44.568392 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.568403 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:44.568414 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:44.568424 | orchestrator | 2025-08-29 18:02:44.568435 | orchestrator | TASK [Aggregate test results step one] ***************************************** 2025-08-29 18:02:44.568446 | orchestrator | Friday 29 August 2025 18:02:40 +0000 (0:00:00.284) 0:00:07.880 ********* 2025-08-29 18:02:44.568457 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568468 | orchestrator | 2025-08-29 18:02:44.568480 | orchestrator | TASK [Aggregate test results step two] ***************************************** 2025-08-29 18:02:44.568491 | orchestrator | Friday 29 August 2025 18:02:40 +0000 (0:00:00.235) 0:00:08.116 ********* 2025-08-29 18:02:44.568502 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568513 | orchestrator | 2025-08-29 18:02:44.568523 | orchestrator | TASK [Aggregate test results step three] *************************************** 2025-08-29 18:02:44.568534 | orchestrator | Friday 29 August 2025 18:02:40 +0000 (0:00:00.246) 0:00:08.363 ********* 2025-08-29 18:02:44.568545 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568556 | orchestrator | 2025-08-29 18:02:44.568567 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:44.568578 | orchestrator | Friday 29 August 2025 18:02:40 +0000 (0:00:00.238) 0:00:08.602 ********* 2025-08-29 18:02:44.568589 | orchestrator | 2025-08-29 18:02:44.568600 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:44.568618 | orchestrator | Friday 29 August 2025 18:02:40 +0000 (0:00:00.066) 0:00:08.669 ********* 2025-08-29 18:02:44.568629 | orchestrator | 2025-08-29 18:02:44.568640 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:44.568651 | orchestrator | Friday 29 August 2025 18:02:41 +0000 (0:00:00.070) 0:00:08.739 ********* 2025-08-29 18:02:44.568662 | orchestrator | 2025-08-29 18:02:44.568673 | orchestrator | TASK [Print report file information] ******************************************* 2025-08-29 18:02:44.568684 | orchestrator | Friday 29 August 2025 18:02:41 +0000 (0:00:00.256) 0:00:08.995 ********* 2025-08-29 18:02:44.568695 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568706 | orchestrator | 2025-08-29 18:02:44.568717 | orchestrator | TASK [Fail early due to containers not running] ******************************** 2025-08-29 18:02:44.568727 | orchestrator | Friday 29 August 2025 18:02:41 +0000 (0:00:00.251) 0:00:09.247 ********* 2025-08-29 18:02:44.568738 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:44.568749 | orchestrator | 2025-08-29 18:02:44.568760 | orchestrator | TASK [Prepare test data] ******************************************************* 2025-08-29 18:02:44.568771 | orchestrator | Friday 29 August 2025 18:02:41 +0000 (0:00:00.270) 0:00:09.517 ********* 2025-08-29 18:02:44.568782 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.568793 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:44.568804 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:44.568815 | orchestrator | 2025-08-29 18:02:44.568826 | orchestrator | TASK [Set _mon_hostname fact] ************************************************** 2025-08-29 18:02:44.568837 | orchestrator | Friday 29 August 2025 18:02:42 +0000 (0:00:00.335) 0:00:09.852 ********* 2025-08-29 18:02:44.568849 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.568868 | orchestrator | 2025-08-29 18:02:44.568894 | orchestrator | TASK [Get ceph osd tree] ******************************************************* 2025-08-29 18:02:44.568928 | orchestrator | Friday 29 August 2025 18:02:42 +0000 (0:00:00.216) 0:00:10.069 ********* 2025-08-29 18:02:44.568947 | orchestrator | changed: [testbed-node-3 -> testbed-node-0(192.168.16.10)] 2025-08-29 18:02:44.568965 | orchestrator | 2025-08-29 18:02:44.568983 | orchestrator | TASK [Parse osd tree from JSON] ************************************************ 2025-08-29 18:02:44.569001 | orchestrator | Friday 29 August 2025 18:02:43 +0000 (0:00:01.581) 0:00:11.651 ********* 2025-08-29 18:02:44.569017 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.569035 | orchestrator | 2025-08-29 18:02:44.569051 | orchestrator | TASK [Get OSDs that are not up or in] ****************************************** 2025-08-29 18:02:44.569069 | orchestrator | Friday 29 August 2025 18:02:44 +0000 (0:00:00.157) 0:00:11.809 ********* 2025-08-29 18:02:44.569086 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:44.569103 | orchestrator | 2025-08-29 18:02:44.569120 | orchestrator | TASK [Fail test if OSDs are not up or in] ************************************** 2025-08-29 18:02:44.569138 | orchestrator | Friday 29 August 2025 18:02:44 +0000 (0:00:00.321) 0:00:12.130 ********* 2025-08-29 18:02:44.569169 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:57.367101 | orchestrator | 2025-08-29 18:02:57.367177 | orchestrator | TASK [Pass test if OSDs are all up and in] ************************************* 2025-08-29 18:02:57.367189 | orchestrator | Friday 29 August 2025 18:02:44 +0000 (0:00:00.120) 0:00:12.251 ********* 2025-08-29 18:02:57.367198 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367207 | orchestrator | 2025-08-29 18:02:57.367215 | orchestrator | TASK [Prepare test data] ******************************************************* 2025-08-29 18:02:57.367223 | orchestrator | Friday 29 August 2025 18:02:44 +0000 (0:00:00.130) 0:00:12.381 ********* 2025-08-29 18:02:57.367231 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367239 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367247 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367255 | orchestrator | 2025-08-29 18:02:57.367293 | orchestrator | TASK [List ceph LVM volumes and collect data] ********************************** 2025-08-29 18:02:57.367303 | orchestrator | Friday 29 August 2025 18:02:45 +0000 (0:00:00.496) 0:00:12.878 ********* 2025-08-29 18:02:57.367311 | orchestrator | changed: [testbed-node-3] 2025-08-29 18:02:57.367335 | orchestrator | changed: [testbed-node-4] 2025-08-29 18:02:57.367344 | orchestrator | changed: [testbed-node-5] 2025-08-29 18:02:57.367352 | orchestrator | 2025-08-29 18:02:57.367360 | orchestrator | TASK [Parse LVM data as JSON] ************************************************** 2025-08-29 18:02:57.367368 | orchestrator | Friday 29 August 2025 18:02:47 +0000 (0:00:02.542) 0:00:15.420 ********* 2025-08-29 18:02:57.367376 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367384 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367392 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367400 | orchestrator | 2025-08-29 18:02:57.367408 | orchestrator | TASK [Get unencrypted and encrypted OSDs] ************************************** 2025-08-29 18:02:57.367416 | orchestrator | Friday 29 August 2025 18:02:48 +0000 (0:00:00.283) 0:00:15.704 ********* 2025-08-29 18:02:57.367423 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367431 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367439 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367447 | orchestrator | 2025-08-29 18:02:57.367455 | orchestrator | TASK [Fail if count of encrypted OSDs does not match] ************************** 2025-08-29 18:02:57.367473 | orchestrator | Friday 29 August 2025 18:02:48 +0000 (0:00:00.479) 0:00:16.183 ********* 2025-08-29 18:02:57.367488 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:57.367497 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:57.367505 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:57.367513 | orchestrator | 2025-08-29 18:02:57.367521 | orchestrator | TASK [Pass if count of encrypted OSDs equals count of OSDs] ******************** 2025-08-29 18:02:57.367529 | orchestrator | Friday 29 August 2025 18:02:48 +0000 (0:00:00.491) 0:00:16.675 ********* 2025-08-29 18:02:57.367537 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367545 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367553 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367561 | orchestrator | 2025-08-29 18:02:57.367569 | orchestrator | TASK [Fail if count of unencrypted OSDs does not match] ************************ 2025-08-29 18:02:57.367577 | orchestrator | Friday 29 August 2025 18:02:49 +0000 (0:00:00.317) 0:00:16.993 ********* 2025-08-29 18:02:57.367585 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:57.367593 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:57.367601 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:57.367609 | orchestrator | 2025-08-29 18:02:57.367617 | orchestrator | TASK [Pass if count of unencrypted OSDs equals count of OSDs] ****************** 2025-08-29 18:02:57.367625 | orchestrator | Friday 29 August 2025 18:02:49 +0000 (0:00:00.281) 0:00:17.274 ********* 2025-08-29 18:02:57.367633 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:57.367641 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:57.367649 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:57.367656 | orchestrator | 2025-08-29 18:02:57.367664 | orchestrator | TASK [Prepare test data] ******************************************************* 2025-08-29 18:02:57.367673 | orchestrator | Friday 29 August 2025 18:02:49 +0000 (0:00:00.274) 0:00:17.549 ********* 2025-08-29 18:02:57.367682 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367691 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367700 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367709 | orchestrator | 2025-08-29 18:02:57.367719 | orchestrator | TASK [Get CRUSH node data of each OSD host and root node childs] *************** 2025-08-29 18:02:57.367728 | orchestrator | Friday 29 August 2025 18:02:50 +0000 (0:00:00.722) 0:00:18.271 ********* 2025-08-29 18:02:57.367737 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367746 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367755 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367764 | orchestrator | 2025-08-29 18:02:57.367773 | orchestrator | TASK [Calculate sub test expression results] *********************************** 2025-08-29 18:02:57.367783 | orchestrator | Friday 29 August 2025 18:02:51 +0000 (0:00:00.498) 0:00:18.769 ********* 2025-08-29 18:02:57.367792 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367801 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367810 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367819 | orchestrator | 2025-08-29 18:02:57.367834 | orchestrator | TASK [Fail test if any sub test failed] **************************************** 2025-08-29 18:02:57.367843 | orchestrator | Friday 29 August 2025 18:02:51 +0000 (0:00:00.285) 0:00:19.054 ********* 2025-08-29 18:02:57.367853 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:57.367863 | orchestrator | skipping: [testbed-node-4] 2025-08-29 18:02:57.367872 | orchestrator | skipping: [testbed-node-5] 2025-08-29 18:02:57.367882 | orchestrator | 2025-08-29 18:02:57.367891 | orchestrator | TASK [Pass test if no sub test failed] ***************************************** 2025-08-29 18:02:57.367900 | orchestrator | Friday 29 August 2025 18:02:51 +0000 (0:00:00.314) 0:00:19.369 ********* 2025-08-29 18:02:57.367909 | orchestrator | ok: [testbed-node-3] 2025-08-29 18:02:57.367918 | orchestrator | ok: [testbed-node-4] 2025-08-29 18:02:57.367927 | orchestrator | ok: [testbed-node-5] 2025-08-29 18:02:57.367936 | orchestrator | 2025-08-29 18:02:57.367945 | orchestrator | TASK [Set validation result to passed if no test failed] *********************** 2025-08-29 18:02:57.367954 | orchestrator | Friday 29 August 2025 18:02:52 +0000 (0:00:00.567) 0:00:19.936 ********* 2025-08-29 18:02:57.367963 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:57.367973 | orchestrator | 2025-08-29 18:02:57.367982 | orchestrator | TASK [Set validation result to failed if a test failed] ************************ 2025-08-29 18:02:57.367991 | orchestrator | Friday 29 August 2025 18:02:52 +0000 (0:00:00.255) 0:00:20.191 ********* 2025-08-29 18:02:57.368001 | orchestrator | skipping: [testbed-node-3] 2025-08-29 18:02:57.368010 | orchestrator | 2025-08-29 18:02:57.368031 | orchestrator | TASK [Aggregate test results step one] ***************************************** 2025-08-29 18:02:57.368042 | orchestrator | Friday 29 August 2025 18:02:52 +0000 (0:00:00.246) 0:00:20.438 ********* 2025-08-29 18:02:57.368051 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:57.368060 | orchestrator | 2025-08-29 18:02:57.368069 | orchestrator | TASK [Aggregate test results step two] ***************************************** 2025-08-29 18:02:57.368079 | orchestrator | Friday 29 August 2025 18:02:54 +0000 (0:00:01.591) 0:00:22.030 ********* 2025-08-29 18:02:57.368088 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:57.368097 | orchestrator | 2025-08-29 18:02:57.368106 | orchestrator | TASK [Aggregate test results step three] *************************************** 2025-08-29 18:02:57.368114 | orchestrator | Friday 29 August 2025 18:02:54 +0000 (0:00:00.242) 0:00:22.272 ********* 2025-08-29 18:02:57.368122 | orchestrator | changed: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:57.368130 | orchestrator | 2025-08-29 18:02:57.368138 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:57.368146 | orchestrator | Friday 29 August 2025 18:02:54 +0000 (0:00:00.249) 0:00:22.522 ********* 2025-08-29 18:02:57.368154 | orchestrator | 2025-08-29 18:02:57.368162 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:57.368169 | orchestrator | Friday 29 August 2025 18:02:54 +0000 (0:00:00.101) 0:00:22.623 ********* 2025-08-29 18:02:57.368177 | orchestrator | 2025-08-29 18:02:57.368185 | orchestrator | TASK [Flush handlers] ********************************************************** 2025-08-29 18:02:57.368193 | orchestrator | Friday 29 August 2025 18:02:55 +0000 (0:00:00.067) 0:00:22.691 ********* 2025-08-29 18:02:57.368201 | orchestrator | 2025-08-29 18:02:57.368209 | orchestrator | RUNNING HANDLER [Write report file] ******************************************** 2025-08-29 18:02:57.368216 | orchestrator | Friday 29 August 2025 18:02:55 +0000 (0:00:00.073) 0:00:22.764 ********* 2025-08-29 18:02:57.368224 | orchestrator | changed: [testbed-node-3 -> testbed-manager(192.168.16.5)] 2025-08-29 18:02:57.368232 | orchestrator | 2025-08-29 18:02:57.368240 | orchestrator | TASK [Print report file information] ******************************************* 2025-08-29 18:02:57.368248 | orchestrator | Friday 29 August 2025 18:02:56 +0000 (0:00:01.521) 0:00:24.286 ********* 2025-08-29 18:02:57.368255 | orchestrator | ok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => { 2025-08-29 18:02:57.368275 | orchestrator |  "msg": [ 2025-08-29 18:02:57.368284 | orchestrator |  "Validator run completed.", 2025-08-29 18:02:57.368299 | orchestrator |  "You can find the report file here:", 2025-08-29 18:02:57.368307 | orchestrator |  "/opt/reports/validator/ceph-osds-validator-2025-08-29T18:02:33+00:00-report.json", 2025-08-29 18:02:57.368316 | orchestrator |  "on the following host:", 2025-08-29 18:02:57.368324 | orchestrator |  "testbed-manager" 2025-08-29 18:02:57.368332 | orchestrator |  ] 2025-08-29 18:02:57.368340 | orchestrator | } 2025-08-29 18:02:57.368348 | orchestrator | 2025-08-29 18:02:57.368356 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 18:02:57.368365 | orchestrator | testbed-node-3 : ok=35  changed=4  unreachable=0 failed=0 skipped=17  rescued=0 ignored=0 2025-08-29 18:02:57.368374 | orchestrator | testbed-node-4 : ok=18  changed=1  unreachable=0 failed=0 skipped=9  rescued=0 ignored=0 2025-08-29 18:02:57.368382 | orchestrator | testbed-node-5 : ok=18  changed=1  unreachable=0 failed=0 skipped=9  rescued=0 ignored=0 2025-08-29 18:02:57.368390 | orchestrator | 2025-08-29 18:02:57.368398 | orchestrator | 2025-08-29 18:02:57.368406 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 18:02:57.368414 | orchestrator | Friday 29 August 2025 18:02:57 +0000 (0:00:00.744) 0:00:25.031 ********* 2025-08-29 18:02:57.368421 | orchestrator | =============================================================================== 2025-08-29 18:02:57.368429 | orchestrator | List ceph LVM volumes and collect data ---------------------------------- 2.54s 2025-08-29 18:02:57.368437 | orchestrator | Aggregate test results step one ----------------------------------------- 1.59s 2025-08-29 18:02:57.368445 | orchestrator | Get ceph osd tree ------------------------------------------------------- 1.58s 2025-08-29 18:02:57.368452 | orchestrator | Write report file ------------------------------------------------------- 1.52s 2025-08-29 18:02:57.368494 | orchestrator | Create report output directory ------------------------------------------ 1.01s 2025-08-29 18:02:57.368503 | orchestrator | Print report file information ------------------------------------------- 0.74s 2025-08-29 18:02:57.368514 | orchestrator | Prepare test data ------------------------------------------------------- 0.72s 2025-08-29 18:02:57.368522 | orchestrator | Get timestamp for report file ------------------------------------------- 0.63s 2025-08-29 18:02:57.368530 | orchestrator | Pass test if no sub test failed ----------------------------------------- 0.57s 2025-08-29 18:02:57.368538 | orchestrator | Get list of ceph-osd containers on host --------------------------------- 0.57s 2025-08-29 18:02:57.368546 | orchestrator | Calculate total number of OSDs in cluster ------------------------------- 0.55s 2025-08-29 18:02:57.368554 | orchestrator | Get CRUSH node data of each OSD host and root node childs --------------- 0.50s 2025-08-29 18:02:57.368562 | orchestrator | Prepare test data ------------------------------------------------------- 0.50s 2025-08-29 18:02:57.368569 | orchestrator | Set test result to passed if count matches ------------------------------ 0.49s 2025-08-29 18:02:57.368578 | orchestrator | Fail if count of encrypted OSDs does not match -------------------------- 0.49s 2025-08-29 18:02:57.368586 | orchestrator | Get unencrypted and encrypted OSDs -------------------------------------- 0.48s 2025-08-29 18:02:57.368599 | orchestrator | Set test result to failed if an OSD is not running ---------------------- 0.46s 2025-08-29 18:02:57.545461 | orchestrator | Prepare test data ------------------------------------------------------- 0.45s 2025-08-29 18:02:57.545546 | orchestrator | Flush handlers ---------------------------------------------------------- 0.39s 2025-08-29 18:02:57.545562 | orchestrator | Prepare test data ------------------------------------------------------- 0.34s 2025-08-29 18:02:57.752201 | orchestrator | + sh -c /opt/configuration/scripts/check/200-infrastructure.sh 2025-08-29 18:02:57.756814 | orchestrator | + set -e 2025-08-29 18:02:57.756854 | orchestrator | + source /opt/manager-vars.sh 2025-08-29 18:02:57.756868 | orchestrator | ++ export NUMBER_OF_NODES=6 2025-08-29 18:02:57.756880 | orchestrator | ++ NUMBER_OF_NODES=6 2025-08-29 18:02:57.756891 | orchestrator | ++ export CEPH_VERSION=reef 2025-08-29 18:02:57.756925 | orchestrator | ++ CEPH_VERSION=reef 2025-08-29 18:02:57.756937 | orchestrator | ++ export CONFIGURATION_VERSION=main 2025-08-29 18:02:57.756949 | orchestrator | ++ CONFIGURATION_VERSION=main 2025-08-29 18:02:57.756960 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 18:02:57.756971 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 18:02:57.756982 | orchestrator | ++ export OPENSTACK_VERSION=2024.2 2025-08-29 18:02:57.756993 | orchestrator | ++ OPENSTACK_VERSION=2024.2 2025-08-29 18:02:57.757004 | orchestrator | ++ export ARA=false 2025-08-29 18:02:57.757015 | orchestrator | ++ ARA=false 2025-08-29 18:02:57.757025 | orchestrator | ++ export DEPLOY_MODE=manager 2025-08-29 18:02:57.757036 | orchestrator | ++ DEPLOY_MODE=manager 2025-08-29 18:02:57.757047 | orchestrator | ++ export TEMPEST=false 2025-08-29 18:02:57.757058 | orchestrator | ++ TEMPEST=false 2025-08-29 18:02:57.757069 | orchestrator | ++ export IS_ZUUL=true 2025-08-29 18:02:57.757080 | orchestrator | ++ IS_ZUUL=true 2025-08-29 18:02:57.757091 | orchestrator | ++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 18:02:57.757102 | orchestrator | ++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.190 2025-08-29 18:02:57.757113 | orchestrator | ++ export EXTERNAL_API=false 2025-08-29 18:02:57.757124 | orchestrator | ++ EXTERNAL_API=false 2025-08-29 18:02:57.757135 | orchestrator | ++ export IMAGE_USER=ubuntu 2025-08-29 18:02:57.757145 | orchestrator | ++ IMAGE_USER=ubuntu 2025-08-29 18:02:57.757156 | orchestrator | ++ export IMAGE_NODE_USER=ubuntu 2025-08-29 18:02:57.757167 | orchestrator | ++ IMAGE_NODE_USER=ubuntu 2025-08-29 18:02:57.757178 | orchestrator | ++ export CEPH_STACK=ceph-ansible 2025-08-29 18:02:57.757189 | orchestrator | ++ CEPH_STACK=ceph-ansible 2025-08-29 18:02:57.757199 | orchestrator | + [[ -e /etc/redhat-release ]] 2025-08-29 18:02:57.757210 | orchestrator | + source /etc/os-release 2025-08-29 18:02:57.757221 | orchestrator | ++ PRETTY_NAME='Ubuntu 24.04.3 LTS' 2025-08-29 18:02:57.757232 | orchestrator | ++ NAME=Ubuntu 2025-08-29 18:02:57.757243 | orchestrator | ++ VERSION_ID=24.04 2025-08-29 18:02:57.757253 | orchestrator | ++ VERSION='24.04.3 LTS (Noble Numbat)' 2025-08-29 18:02:57.757294 | orchestrator | ++ VERSION_CODENAME=noble 2025-08-29 18:02:57.757307 | orchestrator | ++ ID=ubuntu 2025-08-29 18:02:57.757317 | orchestrator | ++ ID_LIKE=debian 2025-08-29 18:02:57.757328 | orchestrator | ++ HOME_URL=https://www.ubuntu.com/ 2025-08-29 18:02:57.757340 | orchestrator | ++ SUPPORT_URL=https://help.ubuntu.com/ 2025-08-29 18:02:57.757350 | orchestrator | ++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/ 2025-08-29 18:02:57.757362 | orchestrator | ++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy 2025-08-29 18:02:57.757373 | orchestrator | ++ UBUNTU_CODENAME=noble 2025-08-29 18:02:57.757385 | orchestrator | ++ LOGO=ubuntu-logo 2025-08-29 18:02:57.757396 | orchestrator | + [[ ubuntu == \u\b\u\n\t\u ]] 2025-08-29 18:02:57.757407 | orchestrator | + packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client' 2025-08-29 18:02:57.757419 | orchestrator | + dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client 2025-08-29 18:02:57.789508 | orchestrator | + sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client 2025-08-29 18:03:21.135208 | orchestrator | 2025-08-29 18:03:21.135381 | orchestrator | # Status of Elasticsearch 2025-08-29 18:03:21.135408 | orchestrator | 2025-08-29 18:03:21.135427 | orchestrator | + pushd /opt/configuration/contrib 2025-08-29 18:03:21.135448 | orchestrator | + echo 2025-08-29 18:03:21.135468 | orchestrator | + echo '# Status of Elasticsearch' 2025-08-29 18:03:21.135486 | orchestrator | + echo 2025-08-29 18:03:21.135505 | orchestrator | + bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s 2025-08-29 18:03:21.339901 | orchestrator | OK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0 | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0 2025-08-29 18:03:21.340015 | orchestrator | 2025-08-29 18:03:21.340069 | orchestrator | # Status of MariaDB 2025-08-29 18:03:21.340089 | orchestrator | 2025-08-29 18:03:21.340105 | orchestrator | + echo 2025-08-29 18:03:21.340123 | orchestrator | + echo '# Status of MariaDB' 2025-08-29 18:03:21.340140 | orchestrator | + echo 2025-08-29 18:03:21.340157 | orchestrator | + MARIADB_USER=root_shard_0 2025-08-29 18:03:21.340178 | orchestrator | + bash nagios-plugins/check_galera_cluster -u root_shard_0 -p password -H api-int.testbed.osism.xyz -c 1 2025-08-29 18:03:21.404744 | orchestrator | Reading package lists... 2025-08-29 18:03:21.729966 | orchestrator | Building dependency tree... 2025-08-29 18:03:21.730442 | orchestrator | Reading state information... 2025-08-29 18:03:22.127241 | orchestrator | bc is already the newest version (1.07.1-3ubuntu4). 2025-08-29 18:03:22.127360 | orchestrator | bc set to manually installed. 2025-08-29 18:03:22.127375 | orchestrator | 0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded. 2025-08-29 18:03:22.711552 | orchestrator | OK: number of NODES = 3 (wsrep_cluster_size) 2025-08-29 18:03:22.713226 | orchestrator | 2025-08-29 18:03:22.713321 | orchestrator | # Status of Prometheus 2025-08-29 18:03:22.713336 | orchestrator | 2025-08-29 18:03:22.713349 | orchestrator | + echo 2025-08-29 18:03:22.713361 | orchestrator | + echo '# Status of Prometheus' 2025-08-29 18:03:22.713372 | orchestrator | + echo 2025-08-29 18:03:22.713383 | orchestrator | + curl -s https://api-int.testbed.osism.xyz:9091/-/healthy 2025-08-29 18:03:22.772852 | orchestrator | Unauthorized 2025-08-29 18:03:22.776479 | orchestrator | + curl -s https://api-int.testbed.osism.xyz:9091/-/ready 2025-08-29 18:03:22.831117 | orchestrator | Unauthorized 2025-08-29 18:03:22.834956 | orchestrator | 2025-08-29 18:03:22.835011 | orchestrator | # Status of RabbitMQ 2025-08-29 18:03:22.835026 | orchestrator | 2025-08-29 18:03:22.835038 | orchestrator | + echo 2025-08-29 18:03:22.835049 | orchestrator | + echo '# Status of RabbitMQ' 2025-08-29 18:03:22.835060 | orchestrator | + echo 2025-08-29 18:03:22.835090 | orchestrator | + perl nagios-plugins/check_rabbitmq_cluster --ssl 1 -H api-int.testbed.osism.xyz -u openstack -p password 2025-08-29 18:03:23.294602 | orchestrator | RABBITMQ_CLUSTER OK - nb_running_node OK (3) nb_running_disc_node OK (3) nb_running_ram_node OK (0) 2025-08-29 18:03:23.302551 | orchestrator | 2025-08-29 18:03:23.302648 | orchestrator | # Status of Redis 2025-08-29 18:03:23.302673 | orchestrator | 2025-08-29 18:03:23.302693 | orchestrator | + echo 2025-08-29 18:03:23.302712 | orchestrator | + echo '# Status of Redis' 2025-08-29 18:03:23.302732 | orchestrator | + echo 2025-08-29 18:03:23.302752 | orchestrator | + /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\r\nPING\r\nINFO replication\r\nQUIT\r\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j 2025-08-29 18:03:23.309240 | orchestrator | TCP OK - 0.002 second response time on 192.168.16.10 port 6379|time=0.001774s;;;0.000000;10.000000 2025-08-29 18:03:23.309671 | orchestrator | 2025-08-29 18:03:23.309702 | orchestrator | # Create backup of MariaDB database 2025-08-29 18:03:23.309716 | orchestrator | 2025-08-29 18:03:23.309727 | orchestrator | + popd 2025-08-29 18:03:23.309739 | orchestrator | + echo 2025-08-29 18:03:23.309750 | orchestrator | + echo '# Create backup of MariaDB database' 2025-08-29 18:03:23.309761 | orchestrator | + echo 2025-08-29 18:03:23.309772 | orchestrator | + osism apply mariadb_backup -e mariadb_backup_type=full 2025-08-29 18:03:25.227031 | orchestrator | 2025-08-29 18:03:25 | INFO  | Task 447e9bab-d851-4468-93dd-90e0dcdaf033 (mariadb_backup) was prepared for execution. 2025-08-29 18:03:25.227138 | orchestrator | 2025-08-29 18:03:25 | INFO  | It takes a moment until task 447e9bab-d851-4468-93dd-90e0dcdaf033 (mariadb_backup) has been started and output is visible here. 2025-08-29 18:03:52.461246 | orchestrator | 2025-08-29 18:03:52.461430 | orchestrator | PLAY [Group hosts based on configuration] ************************************** 2025-08-29 18:03:52.461450 | orchestrator | 2025-08-29 18:03:52.461462 | orchestrator | TASK [Group hosts based on Kolla action] *************************************** 2025-08-29 18:03:52.461474 | orchestrator | Friday 29 August 2025 18:03:29 +0000 (0:00:00.192) 0:00:00.192 ********* 2025-08-29 18:03:52.461486 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:03:52.461498 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:03:52.461509 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:03:52.461520 | orchestrator | 2025-08-29 18:03:52.461531 | orchestrator | TASK [Group hosts based on enabled services] *********************************** 2025-08-29 18:03:52.461542 | orchestrator | Friday 29 August 2025 18:03:29 +0000 (0:00:00.292) 0:00:00.484 ********* 2025-08-29 18:03:52.461553 | orchestrator | ok: [testbed-node-0] => (item=enable_mariadb_True) 2025-08-29 18:03:52.461565 | orchestrator | ok: [testbed-node-1] => (item=enable_mariadb_True) 2025-08-29 18:03:52.461576 | orchestrator | ok: [testbed-node-2] => (item=enable_mariadb_True) 2025-08-29 18:03:52.461587 | orchestrator | 2025-08-29 18:03:52.461598 | orchestrator | PLAY [Apply role mariadb] ****************************************************** 2025-08-29 18:03:52.461608 | orchestrator | 2025-08-29 18:03:52.461645 | orchestrator | TASK [mariadb : Group MariaDB hosts based on shards] *************************** 2025-08-29 18:03:52.461657 | orchestrator | Friday 29 August 2025 18:03:29 +0000 (0:00:00.527) 0:00:01.012 ********* 2025-08-29 18:03:52.461668 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-0) 2025-08-29 18:03:52.461680 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-1) 2025-08-29 18:03:52.461691 | orchestrator | ok: [testbed-node-0] => (item=testbed-node-2) 2025-08-29 18:03:52.461701 | orchestrator | 2025-08-29 18:03:52.461712 | orchestrator | TASK [mariadb : include_tasks] ************************************************* 2025-08-29 18:03:52.461723 | orchestrator | Friday 29 August 2025 18:03:30 +0000 (0:00:00.382) 0:00:01.394 ********* 2025-08-29 18:03:52.461735 | orchestrator | included: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2 2025-08-29 18:03:52.461746 | orchestrator | 2025-08-29 18:03:52.461757 | orchestrator | TASK [mariadb : Get MariaDB container facts] *********************************** 2025-08-29 18:03:52.461771 | orchestrator | Friday 29 August 2025 18:03:30 +0000 (0:00:00.485) 0:00:01.880 ********* 2025-08-29 18:03:52.461783 | orchestrator | ok: [testbed-node-1] 2025-08-29 18:03:52.461796 | orchestrator | ok: [testbed-node-0] 2025-08-29 18:03:52.461808 | orchestrator | ok: [testbed-node-2] 2025-08-29 18:03:52.461821 | orchestrator | 2025-08-29 18:03:52.461834 | orchestrator | TASK [mariadb : Taking full database backup via Mariabackup] ******************* 2025-08-29 18:03:52.461846 | orchestrator | Friday 29 August 2025 18:03:33 +0000 (0:00:02.721) 0:00:04.601 ********* 2025-08-29 18:03:52.461859 | orchestrator | [WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart 2025-08-29 18:03:52.461871 | orchestrator | [WARNING]: Could not match supplied host pattern, ignoring: mariadb_start 2025-08-29 18:03:52.461884 | orchestrator | [WARNING]: Could not match supplied host pattern, ignoring: 2025-08-29 18:03:52.461896 | orchestrator | mariadb_bootstrap_restart 2025-08-29 18:03:52.461910 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:03:52.461922 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:03:52.461934 | orchestrator | changed: [testbed-node-0] 2025-08-29 18:03:52.461947 | orchestrator | 2025-08-29 18:03:52.461959 | orchestrator | PLAY [Restart mariadb services] ************************************************ 2025-08-29 18:03:52.461971 | orchestrator | skipping: no hosts matched 2025-08-29 18:03:52.461983 | orchestrator | 2025-08-29 18:03:52.461994 | orchestrator | PLAY [Start mariadb services] ************************************************** 2025-08-29 18:03:52.462005 | orchestrator | skipping: no hosts matched 2025-08-29 18:03:52.462015 | orchestrator | 2025-08-29 18:03:52.462086 | orchestrator | PLAY [Restart bootstrap mariadb service] *************************************** 2025-08-29 18:03:52.462097 | orchestrator | skipping: no hosts matched 2025-08-29 18:03:52.462109 | orchestrator | 2025-08-29 18:03:52.462124 | orchestrator | PLAY [Apply mariadb post-configuration] **************************************** 2025-08-29 18:03:52.462143 | orchestrator | 2025-08-29 18:03:52.462161 | orchestrator | TASK [Include mariadb post-deploy.yml] ***************************************** 2025-08-29 18:03:52.462181 | orchestrator | Friday 29 August 2025 18:03:51 +0000 (0:00:18.046) 0:00:22.647 ********* 2025-08-29 18:03:52.462199 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:03:52.462219 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:03:52.462238 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:03:52.462256 | orchestrator | 2025-08-29 18:03:52.462304 | orchestrator | TASK [Include mariadb post-upgrade.yml] **************************************** 2025-08-29 18:03:52.462342 | orchestrator | Friday 29 August 2025 18:03:51 +0000 (0:00:00.304) 0:00:22.952 ********* 2025-08-29 18:03:52.462361 | orchestrator | skipping: [testbed-node-0] 2025-08-29 18:03:52.462373 | orchestrator | skipping: [testbed-node-1] 2025-08-29 18:03:52.462383 | orchestrator | skipping: [testbed-node-2] 2025-08-29 18:03:52.462394 | orchestrator | 2025-08-29 18:03:52.462405 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 18:03:52.462417 | orchestrator | testbed-node-0 : ok=6  changed=1  unreachable=0 failed=0 skipped=2  rescued=0 ignored=0 2025-08-29 18:03:52.462439 | orchestrator | testbed-node-1 : ok=4  changed=0 unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 18:03:52.462450 | orchestrator | testbed-node-2 : ok=4  changed=0 unreachable=0 failed=0 skipped=3  rescued=0 ignored=0 2025-08-29 18:03:52.462461 | orchestrator | 2025-08-29 18:03:52.462472 | orchestrator | 2025-08-29 18:03:52.462482 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 18:03:52.462493 | orchestrator | Friday 29 August 2025 18:03:52 +0000 (0:00:00.223) 0:00:23.175 ********* 2025-08-29 18:03:52.462504 | orchestrator | =============================================================================== 2025-08-29 18:03:52.462514 | orchestrator | mariadb : Taking full database backup via Mariabackup ------------------ 18.05s 2025-08-29 18:03:52.462544 | orchestrator | mariadb : Get MariaDB container facts ----------------------------------- 2.72s 2025-08-29 18:03:52.462556 | orchestrator | Group hosts based on enabled services ----------------------------------- 0.53s 2025-08-29 18:03:52.462566 | orchestrator | mariadb : include_tasks ------------------------------------------------- 0.49s 2025-08-29 18:03:52.462577 | orchestrator | mariadb : Group MariaDB hosts based on shards --------------------------- 0.38s 2025-08-29 18:03:52.462588 | orchestrator | Include mariadb post-deploy.yml ----------------------------------------- 0.30s 2025-08-29 18:03:52.462598 | orchestrator | Group hosts based on Kolla action --------------------------------------- 0.29s 2025-08-29 18:03:52.462609 | orchestrator | Include mariadb post-upgrade.yml ---------------------------------------- 0.22s 2025-08-29 18:03:52.752116 | orchestrator | + sh -c /opt/configuration/scripts/check/300-openstack.sh 2025-08-29 18:03:52.760483 | orchestrator | + set -e 2025-08-29 18:03:52.760538 | orchestrator | + source /opt/configuration/scripts/include.sh 2025-08-29 18:03:52.760964 | orchestrator | ++ export INTERACTIVE=false 2025-08-29 18:03:52.761051 | orchestrator | ++ INTERACTIVE=false 2025-08-29 18:03:52.761152 | orchestrator | ++ export OSISM_APPLY_RETRY=1 2025-08-29 18:03:52.761863 | orchestrator | ++ OSISM_APPLY_RETRY=1 2025-08-29 18:03:52.761934 | orchestrator | + source /opt/configuration/scripts/manager-version.sh 2025-08-29 18:03:52.762602 | orchestrator | +++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml 2025-08-29 18:03:52.768777 | orchestrator | 2025-08-29 18:03:52.768833 | orchestrator | # OpenStack endpoints 2025-08-29 18:03:52.768848 | orchestrator | 2025-08-29 18:03:52.768861 | orchestrator | ++ export MANAGER_VERSION=9.2.0 2025-08-29 18:03:52.768873 | orchestrator | ++ MANAGER_VERSION=9.2.0 2025-08-29 18:03:52.768884 | orchestrator | + export OS_CLOUD=admin 2025-08-29 18:03:52.768896 | orchestrator | + OS_CLOUD=admin 2025-08-29 18:03:52.768908 | orchestrator | + echo 2025-08-29 18:03:52.768919 | orchestrator | + echo '# OpenStack endpoints' 2025-08-29 18:03:52.768929 | orchestrator | + echo 2025-08-29 18:03:52.768940 | orchestrator | + openstack endpoint list 2025-08-29 18:03:56.164170 | orchestrator | +----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+ 2025-08-29 18:03:56.164341 | orchestrator | | ID | Region | Service Name | Service Type | Enabled | Interface | URL | 2025-08-29 18:03:56.164360 | orchestrator | +----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+ 2025-08-29 18:03:56.164372 | orchestrator | | 075221b60fc344829814e5dcbabf1e64 | RegionOne | glance | image | True | public | https://api.testbed.osism.xyz:9292 | 2025-08-29 18:03:56.164383 | orchestrator | | 09283d8dfdcd440f997bd2d1bc06022d | RegionOne | cinderv3 | volumev3 | True | internal | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s | 2025-08-29 18:03:56.164394 | orchestrator | | 1a9ea3977652421ea8be133e6da178f2 | RegionOne | keystone | identity | True | public | https://api.testbed.osism.xyz:5000 | 2025-08-29 18:03:56.164405 | orchestrator | | 1dd8f46a84ff44da8975bcc376285972 | RegionOne | designate | dns | True | public | https://api.testbed.osism.xyz:9001 | 2025-08-29 18:03:56.164444 | orchestrator | | 2b3d7380b56741c69f8fcfb7b64d6505 | RegionOne | designate | dns | True | internal | https://api-int.testbed.osism.xyz:9001 | 2025-08-29 18:03:56.164456 | orchestrator | | 4027b7f65ba64d979de260f7f90285d0 | RegionOne | barbican | key-manager | True | public | https://api.testbed.osism.xyz:9311 | 2025-08-29 18:03:56.164472 | orchestrator | | 43be7cb3e5cf4868ac754b658b6eb87d | RegionOne | swift | object-store | True | public | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s | 2025-08-29 18:03:56.164491 | orchestrator | | 4eb314e91a254193a235aee84e580b88 | RegionOne | octavia | load-balancer | True | internal | https://api-int.testbed.osism.xyz:9876 | 2025-08-29 18:03:56.164510 | orchestrator | | 506874ae096d4a11a630b637d052495f | RegionOne | neutron | network | True | public | https://api.testbed.osism.xyz:9696 | 2025-08-29 18:03:56.164527 | orchestrator | | 5ee80c99d5064aa1928494f3cf68e5e2 | RegionOne | nova | compute | True | public | https://api.testbed.osism.xyz:8774/v2.1 | 2025-08-29 18:03:56.164547 | orchestrator | | 74195282d53e49d6a16884c53f367db9 | RegionOne | swift | object-store | True | internal | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s | 2025-08-29 18:03:56.164567 | orchestrator | | 76087a555bea48a58282a1d4e57bd716 | RegionOne | magnum | container-infra | True | public | https://api.testbed.osism.xyz:9511/v1 | 2025-08-29 18:03:56.164587 | orchestrator | | 996362ab7c014ce699ea52ba402577f8 | RegionOne | cinderv3 | volumev3 | True | public | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s | 2025-08-29 18:03:56.164600 | orchestrator | | b1a5e56b15be49d5ae2fd7379d717ed5 | RegionOne | octavia | load-balancer | True | public | https://api.testbed.osism.xyz:9876 | 2025-08-29 18:03:56.164611 | orchestrator | | bacf86a6914344aa994e7533b1531fd6 | RegionOne | nova | compute | True | internal | https://api-int.testbed.osism.xyz:8774/v2.1 | 2025-08-29 18:03:56.164622 | orchestrator | | bb0ce91f59c64982be22c4eec4df2d21 | RegionOne | placement | placement | True | public | https://api.testbed.osism.xyz:8780 | 2025-08-29 18:03:56.164633 | orchestrator | | bd78dddb75eb4095a0ab24bd5ae9ab45 | RegionOne | keystone | identity | True | internal | https://api-int.testbed.osism.xyz:5000 | 2025-08-29 18:03:56.164644 | orchestrator | | dc14b45be10b49ce96c7b96efe8f5b99 | RegionOne | neutron | network | True | internal | https://api-int.testbed.osism.xyz:9696 | 2025-08-29 18:03:56.164655 | orchestrator | | e566bdcf06fc48d388431aaedfd23639 | RegionOne | barbican | key-manager | True | internal | https://api-int.testbed.osism.xyz:9311 | 2025-08-29 18:03:56.164666 | orchestrator | | e66f331bdde14c30a4b5b4b9dfb6c88a | RegionOne | placement | placement | True | internal | https://api-int.testbed.osism.xyz:8780 | 2025-08-29 18:03:56.164696 | orchestrator | | eff62a026107446e9483262c70e92ed9 | RegionOne | magnum | container-infra | True | internal | https://api-int.testbed.osism.xyz:9511/v1 | 2025-08-29 18:03:56.164709 | orchestrator | | fe50e451a4a14c63a1f9dd3412b204b3 | RegionOne | glance | image | True | internal | https://api-int.testbed.osism.xyz:9292 | 2025-08-29 18:03:56.164721 | orchestrator | +----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+ 2025-08-29 18:03:56.408974 | orchestrator | 2025-08-29 18:03:56.409082 | orchestrator | # Cinder 2025-08-29 18:03:56.409097 | orchestrator | 2025-08-29 18:03:56.409133 | orchestrator | + echo 2025-08-29 18:03:56.409158 | orchestrator | + echo '# Cinder' 2025-08-29 18:03:56.409179 | orchestrator | + echo 2025-08-29 18:03:56.409196 | orchestrator | + openstack volume service list 2025-08-29 18:03:59.088886 | orchestrator | +------------------+----------------------------+----------+---------+-------+----------------------------+ 2025-08-29 18:03:59.088969 | orchestrator | | Binary | Host | Zone | Status | State | Updated At | 2025-08-29 18:03:59.088978 | orchestrator | +------------------+----------------------------+----------+---------+-------+----------------------------+ 2025-08-29 18:03:59.088986 | orchestrator | | cinder-scheduler | testbed-node-0 | internal | enabled | up | 2025-08-29T18:03:49.000000 | 2025-08-29 18:03:59.088993 | orchestrator | | cinder-scheduler | testbed-node-1 | internal | enabled | up | 2025-08-29T18:03:49.000000 | 2025-08-29 18:03:59.089000 | orchestrator | | cinder-scheduler | testbed-node-2 | internal | enabled | up | 2025-08-29T18:03:49.000000 | 2025-08-29 18:03:59.089007 | orchestrator | | cinder-volume | testbed-node-5@rbd-volumes | nova | enabled | up | 2025-08-29T18:03:49.000000 | 2025-08-29 18:03:59.089014 | orchestrator | | cinder-volume | testbed-node-4@rbd-volumes | nova | enabled | up | 2025-08-29T18:03:53.000000 | 2025-08-29 18:03:59.089021 | orchestrator | | cinder-volume | testbed-node-3@rbd-volumes | nova | enabled | up | 2025-08-29T18:03:55.000000 | 2025-08-29 18:03:59.089028 | orchestrator | | cinder-backup | testbed-node-5 | nova | enabled | up | 2025-08-29T18:03:57.000000 | 2025-08-29 18:03:59.089034 | orchestrator | | cinder-backup | testbed-node-4 | nova | enabled | up | 2025-08-29T18:03:58.000000 | 2025-08-29 18:03:59.089041 | orchestrator | | cinder-backup | testbed-node-3 | nova | enabled | up | 2025-08-29T18:03:58.000000 | 2025-08-29 18:03:59.089048 | orchestrator | +------------------+----------------------------+----------+---------+-------+----------------------------+ 2025-08-29 18:03:59.353866 | orchestrator | 2025-08-29 18:03:59.353956 | orchestrator | # Neutron 2025-08-29 18:03:59.353970 | orchestrator | 2025-08-29 18:03:59.353981 | orchestrator | + echo 2025-08-29 18:03:59.353993 | orchestrator | + echo '# Neutron' 2025-08-29 18:03:59.354077 | orchestrator | + echo 2025-08-29 18:03:59.354092 | orchestrator | + openstack network agent list 2025-08-29 18:04:02.639235 | orchestrator | +--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+ 2025-08-29 18:04:02.639364 | orchestrator | | ID | Agent Type | Host | Availability Zone | Alive | State | Binary | 2025-08-29 18:04:02.639379 | orchestrator | +--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+ 2025-08-29 18:04:02.639391 | orchestrator | | testbed-node-3 | OVN Controller agent | testbed-node-3 | | :-) | UP | ovn-controller | 2025-08-29 18:04:02.639402 | orchestrator | | testbed-node-5 | OVN Controller agent | testbed-node-5 | | :-) | UP | ovn-controller | 2025-08-29 18:04:02.639413 | orchestrator | | testbed-node-1 | OVN Controller Gateway agent | testbed-node-1 | nova | :-) | UP | ovn-controller | 2025-08-29 18:04:02.639423 | orchestrator | | testbed-node-4 | OVN Controller agent | testbed-node-4 | | :-) | UP | ovn-controller | 2025-08-29 18:04:02.639434 | orchestrator | | testbed-node-2 | OVN Controller Gateway agent | testbed-node-2 | nova | :-) | UP | ovn-controller | 2025-08-29 18:04:02.639445 | orchestrator | | testbed-node-0 | OVN Controller Gateway agent | testbed-node-0 | nova | :-) | UP | ovn-controller | 2025-08-29 18:04:02.639456 | orchestrator | | e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent | testbed-node-3 | | :-) | UP | neutron-ovn-metadata-agent | 2025-08-29 18:04:02.639494 | orchestrator | | 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent | testbed-node-5 | | :-) | UP | neutron-ovn-metadata-agent | 2025-08-29 18:04:02.639506 | orchestrator | | 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent | testbed-node-4 | | :-) | UP | neutron-ovn-metadata-agent | 2025-08-29 18:04:02.639517 | orchestrator | +--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+ 2025-08-29 18:04:02.916635 | orchestrator | + openstack network service provider list 2025-08-29 18:04:05.672612 | orchestrator | +---------------+------+---------+ 2025-08-29 18:04:05.672743 | orchestrator | | Service Type | Name | Default | 2025-08-29 18:04:05.672767 | orchestrator | +---------------+------+---------+ 2025-08-29 18:04:05.672788 | orchestrator | | L3_ROUTER_NAT | ovn | True | 2025-08-29 18:04:05.672808 | orchestrator | +---------------+------+---------+ 2025-08-29 18:04:05.939703 | orchestrator | 2025-08-29 18:04:05.939783 | orchestrator | # Nova 2025-08-29 18:04:05.939792 | orchestrator | 2025-08-29 18:04:05.939799 | orchestrator | + echo 2025-08-29 18:04:05.939806 | orchestrator | + echo '# Nova' 2025-08-29 18:04:05.939813 | orchestrator | + echo 2025-08-29 18:04:05.939820 | orchestrator | + openstack compute service list 2025-08-29 18:04:09.122089 | orchestrator | +--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+ 2025-08-29 18:04:09.122160 | orchestrator | | ID | Binary | Host | Zone | Status | State | Updated At | 2025-08-29 18:04:09.122167 | orchestrator | +--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+ 2025-08-29 18:04:09.122172 | orchestrator | | 04848393-0c5d-4b0e-b55b-733780d88fa3 | nova-scheduler | testbed-node-0 | internal | enabled | up | 2025-08-29T18:04:05.000000 | 2025-08-29 18:04:09.122176 | orchestrator | | fce5904b-7ff0-4311-a9c4-f82ba6ca8cce | nova-scheduler | testbed-node-2 | internal | enabled | up | 2025-08-29T18:04:01.000000 | 2025-08-29 18:04:09.122181 | orchestrator | | 3127f088-0919-4483-b026-a1468a7cbb6a | nova-scheduler | testbed-node-1 | internal | enabled | up | 2025-08-29T18:04:01.000000 | 2025-08-29 18:04:09.122185 | orchestrator | | 48005187-f821-4715-b154-e92eaa350f40 | nova-conductor | testbed-node-0 | internal | enabled | up | 2025-08-29T18:04:01.000000 | 2025-08-29 18:04:09.122190 | orchestrator | | 467a635e-bba9-41bd-b19c-057cf45f5c07 | nova-conductor | testbed-node-1 | internal | enabled | up | 2025-08-29T18:04:03.000000 | 2025-08-29 18:04:09.122194 | orchestrator | | 608dfb4c-8bbf-4881-bf94-7aa9625cb016 | nova-conductor | testbed-node-2 | internal | enabled | up | 2025-08-29T18:04:03.000000 | 2025-08-29 18:04:09.122198 | orchestrator | | 8c2a2579-8c40-41b6-903e-b696d03a8078 | nova-compute | testbed-node-3 | nova | enabled | up | 2025-08-29T18:04:06.000000 | 2025-08-29 18:04:09.122202 | orchestrator | | 5f8430e8-9f80-4d8a-981a-4b9588e28093 | nova-compute | testbed-node-5 | nova | enabled | up | 2025-08-29T18:04:06.000000 | 2025-08-29 18:04:09.122206 | orchestrator | | 17ab6101-83a7-4f95-aecc-cd943fc5c42f | nova-compute | testbed-node-4 | nova | enabled | up | 2025-08-29T18:04:07.000000 | 2025-08-29 18:04:09.122211 | orchestrator | +--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+ 2025-08-29 18:04:09.395260 | orchestrator | + openstack hypervisor list 2025-08-29 18:04:14.193733 | orchestrator | +--------------------------------------+---------------------+-----------------+---------------+-------+ 2025-08-29 18:04:14.193836 | orchestrator | | ID | Hypervisor Hostname | Hypervisor Type | Host IP | State | 2025-08-29 18:04:14.193850 | orchestrator | +--------------------------------------+---------------------+-----------------+---------------+-------+ 2025-08-29 18:04:14.193862 | orchestrator | | 332c13f3-1cb7-40d8-b788-e9a015e7d413 | testbed-node-3 | QEMU | 192.168.16.13 | up | 2025-08-29 18:04:14.193873 | orchestrator | | e14a82c8-7d94-4ca0-9feb-13708813dfc1 | testbed-node-5 | QEMU | 192.168.16.15 | up | 2025-08-29 18:04:14.193884 | orchestrator | | 8290ba93-0a30-4179-83f1-2677f1027f76 | testbed-node-4 | QEMU | 192.168.16.14 | up | 2025-08-29 18:04:14.193920 | orchestrator | +--------------------------------------+---------------------+-----------------+---------------+-------+ 2025-08-29 18:04:14.507041 | orchestrator | 2025-08-29 18:04:14.507119 | orchestrator | # Run OpenStack test play 2025-08-29 18:04:14.507129 | orchestrator | 2025-08-29 18:04:14.507136 | orchestrator | + echo 2025-08-29 18:04:14.507143 | orchestrator | + echo '# Run OpenStack test play' 2025-08-29 18:04:14.507150 | orchestrator | + echo 2025-08-29 18:04:14.507157 | orchestrator | + osism apply --environment openstack test 2025-08-29 18:04:16.393388 | orchestrator | 2025-08-29 18:04:16 | INFO  | Trying to run play test in environment openstack 2025-08-29 18:04:26.498125 | orchestrator | 2025-08-29 18:04:26 | INFO  | Task 6f36e238-a551-4496-9931-6051ed836217 (test) was prepared for execution. 2025-08-29 18:04:26.498200 | orchestrator | 2025-08-29 18:04:26 | INFO  | It takes a moment until task 6f36e238-a551-4496-9931-6051ed836217 (test) has been started and output is visible here. 2025-08-29 18:10:20.813127 | orchestrator | 2025-08-29 18:10:20.813220 | orchestrator | PLAY [Create test project] ***************************************************** 2025-08-29 18:10:20.813234 | orchestrator | 2025-08-29 18:10:20.813246 | orchestrator | TASK [Create test domain] ****************************************************** 2025-08-29 18:10:20.813258 | orchestrator | Friday 29 August 2025 18:04:30 +0000 (0:00:00.079) 0:00:00.079 ********* 2025-08-29 18:10:20.813269 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813282 | orchestrator | 2025-08-29 18:10:20.813294 | orchestrator | TASK [Create test-admin user] ************************************************** 2025-08-29 18:10:20.813305 | orchestrator | Friday 29 August 2025 18:04:33 +0000 (0:00:03.437) 0:00:03.516 ********* 2025-08-29 18:10:20.813315 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813326 | orchestrator | 2025-08-29 18:10:20.813337 | orchestrator | TASK [Add manager role to user test-admin] ************************************* 2025-08-29 18:10:20.813348 | orchestrator | Friday 29 August 2025 18:04:37 +0000 (0:00:03.929) 0:00:07.445 ********* 2025-08-29 18:10:20.813358 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813419 | orchestrator | 2025-08-29 18:10:20.813430 | orchestrator | TASK [Create test project] ***************************************************** 2025-08-29 18:10:20.813441 | orchestrator | Friday 29 August 2025 18:04:43 +0000 (0:00:05.667) 0:00:13.112 ********* 2025-08-29 18:10:20.813452 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813462 | orchestrator | 2025-08-29 18:10:20.813473 | orchestrator | TASK [Create test user] ******************************************************** 2025-08-29 18:10:20.813484 | orchestrator | Friday 29 August 2025 18:04:47 +0000 (0:00:03.661) 0:00:16.774 ********* 2025-08-29 18:10:20.813495 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813506 | orchestrator | 2025-08-29 18:10:20.813517 | orchestrator | TASK [Add member roles to user test] ******************************************* 2025-08-29 18:10:20.813527 | orchestrator | Friday 29 August 2025 18:04:51 +0000 (0:00:04.023) 0:00:20.797 ********* 2025-08-29 18:10:20.813539 | orchestrator | changed: [localhost] => (item=load-balancer_member) 2025-08-29 18:10:20.813551 | orchestrator | changed: [localhost] => (item=member) 2025-08-29 18:10:20.813562 | orchestrator | changed: [localhost] => (item=creator) 2025-08-29 18:10:20.813573 | orchestrator | 2025-08-29 18:10:20.813584 | orchestrator | TASK [Create test server group] ************************************************ 2025-08-29 18:10:20.813595 | orchestrator | Friday 29 August 2025 18:05:02 +0000 (0:00:11.175) 0:00:31.972 ********* 2025-08-29 18:10:20.813606 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813617 | orchestrator | 2025-08-29 18:10:20.813628 | orchestrator | TASK [Create ssh security group] *********************************************** 2025-08-29 18:10:20.813639 | orchestrator | Friday 29 August 2025 18:05:06 +0000 (0:00:03.884) 0:00:35.857 ********* 2025-08-29 18:10:20.813650 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813660 | orchestrator | 2025-08-29 18:10:20.813671 | orchestrator | TASK [Add rule to ssh security group] ****************************************** 2025-08-29 18:10:20.813684 | orchestrator | Friday 29 August 2025 18:05:11 +0000 (0:00:05.711) 0:00:41.569 ********* 2025-08-29 18:10:20.813714 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813727 | orchestrator | 2025-08-29 18:10:20.813739 | orchestrator | TASK [Create icmp security group] ********************************************** 2025-08-29 18:10:20.813751 | orchestrator | Friday 29 August 2025 18:05:15 +0000 (0:00:04.053) 0:00:45.623 ********* 2025-08-29 18:10:20.813763 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813775 | orchestrator | 2025-08-29 18:10:20.813787 | orchestrator | TASK [Add rule to icmp security group] ***************************************** 2025-08-29 18:10:20.813800 | orchestrator | Friday 29 August 2025 18:05:19 +0000 (0:00:03.920) 0:00:49.544 ********* 2025-08-29 18:10:20.813812 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813824 | orchestrator | 2025-08-29 18:10:20.813836 | orchestrator | TASK [Create test keypair] ***************************************************** 2025-08-29 18:10:20.813848 | orchestrator | Friday 29 August 2025 18:05:23 +0000 (0:00:04.106) 0:00:53.650 ********* 2025-08-29 18:10:20.813861 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813873 | orchestrator | 2025-08-29 18:10:20.813884 | orchestrator | TASK [Create test network topology] ******************************************** 2025-08-29 18:10:20.813897 | orchestrator | Friday 29 August 2025 18:05:27 +0000 (0:00:03.831) 0:00:57.481 ********* 2025-08-29 18:10:20.813909 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.813920 | orchestrator | 2025-08-29 18:10:20.813932 | orchestrator | TASK [Create test instances] *************************************************** 2025-08-29 18:10:20.813959 | orchestrator | Friday 29 August 2025 18:05:42 +0000 (0:00:14.430) 0:01:11.912 ********* 2025-08-29 18:10:20.813972 | orchestrator | changed: [localhost] => (item=test) 2025-08-29 18:10:20.813984 | orchestrator | changed: [localhost] => (item=test-1) 2025-08-29 18:10:20.813997 | orchestrator | changed: [localhost] => (item=test-2) 2025-08-29 18:10:20.814009 | orchestrator | 2025-08-29 18:10:20.814075 | orchestrator | STILL ALIVE [task 'Create test instances' is running] ************************** 2025-08-29 18:10:20.814087 | orchestrator | changed: [localhost] => (item=test-3) 2025-08-29 18:10:20.814099 | orchestrator | 2025-08-29 18:10:20.814111 | orchestrator | STILL ALIVE [task 'Create test instances' is running] ************************** 2025-08-29 18:10:20.814123 | orchestrator | changed: [localhost] => (item=test-4) 2025-08-29 18:10:20.814134 | orchestrator | 2025-08-29 18:10:20.814146 | orchestrator | TASK [Add metadata to instances] *********************************************** 2025-08-29 18:10:20.814158 | orchestrator | Friday 29 August 2025 18:08:58 +0000 (0:03:16.750) 0:04:28.662 ********* 2025-08-29 18:10:20.814170 | orchestrator | changed: [localhost] => (item=test) 2025-08-29 18:10:20.814181 | orchestrator | changed: [localhost] => (item=test-1) 2025-08-29 18:10:20.814193 | orchestrator | changed: [localhost] => (item=test-2) 2025-08-29 18:10:20.814205 | orchestrator | changed: [localhost] => (item=test-3) 2025-08-29 18:10:20.814217 | orchestrator | changed: [localhost] => (item=test-4) 2025-08-29 18:10:20.814228 | orchestrator | 2025-08-29 18:10:20.814240 | orchestrator | TASK [Add tag to instances] **************************************************** 2025-08-29 18:10:20.814252 | orchestrator | Friday 29 August 2025 18:09:22 +0000 (0:00:23.183) 0:04:51.846 ********* 2025-08-29 18:10:20.814263 | orchestrator | changed: [localhost] => (item=test) 2025-08-29 18:10:20.814275 | orchestrator | changed: [localhost] => (item=test-1) 2025-08-29 18:10:20.814287 | orchestrator | changed: [localhost] => (item=test-2) 2025-08-29 18:10:20.814299 | orchestrator | changed: [localhost] => (item=test-3) 2025-08-29 18:10:20.814386 | orchestrator | changed: [localhost] => (item=test-4) 2025-08-29 18:10:20.814401 | orchestrator | 2025-08-29 18:10:20.814413 | orchestrator | TASK [Create test volume] ****************************************************** 2025-08-29 18:10:20.814424 | orchestrator | Friday 29 August 2025 18:09:54 +0000 (0:00:32.711) 0:05:24.558 ********* 2025-08-29 18:10:20.814435 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.814446 | orchestrator | 2025-08-29 18:10:20.814457 | orchestrator | TASK [Attach test volume] ****************************************************** 2025-08-29 18:10:20.814468 | orchestrator | Friday 29 August 2025 18:10:01 +0000 (0:00:06.808) 0:05:31.367 ********* 2025-08-29 18:10:20.814478 | orchestrator | changed: [localhost] 2025-08-29 18:10:20.814499 | orchestrator | 2025-08-29 18:10:20.814510 | orchestrator | TASK [Create floating ip address] ********************************************** 2025-08-29 18:10:20.814521 | orchestrator | Friday 29 August 2025 18:10:15 +0000 (0:00:13.636) 0:05:45.003 ********* 2025-08-29 18:10:20.814532 | orchestrator | ok: [localhost] 2025-08-29 18:10:20.814544 | orchestrator | 2025-08-29 18:10:20.814555 | orchestrator | TASK [Print floating ip address] *********************************************** 2025-08-29 18:10:20.814565 | orchestrator | Friday 29 August 2025 18:10:20 +0000 (0:00:05.169) 0:05:50.173 ********* 2025-08-29 18:10:20.814576 | orchestrator | ok: [localhost] => { 2025-08-29 18:10:20.814587 | orchestrator |  "msg": "192.168.112.105" 2025-08-29 18:10:20.814598 | orchestrator | } 2025-08-29 18:10:20.814609 | orchestrator | 2025-08-29 18:10:20.814620 | orchestrator | PLAY RECAP ********************************************************************* 2025-08-29 18:10:20.814631 | orchestrator | localhost : ok=20  changed=18  unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 2025-08-29 18:10:20.814643 | orchestrator | 2025-08-29 18:10:20.814654 | orchestrator | 2025-08-29 18:10:20.814665 | orchestrator | TASKS RECAP ******************************************************************** 2025-08-29 18:10:20.814676 | orchestrator | Friday 29 August 2025 18:10:20 +0000 (0:00:00.043) 0:05:50.217 ********* 2025-08-29 18:10:20.814686 | orchestrator | =============================================================================== 2025-08-29 18:10:20.814703 | orchestrator | Create test instances ------------------------------------------------- 196.75s 2025-08-29 18:10:20.814714 | orchestrator | Add tag to instances --------------------------------------------------- 32.71s 2025-08-29 18:10:20.814725 | orchestrator | Add metadata to instances ---------------------------------------------- 23.18s 2025-08-29 18:10:20.814736 | orchestrator | Create test network topology ------------------------------------------- 14.43s 2025-08-29 18:10:20.814747 | orchestrator | Attach test volume ----------------------------------------------------- 13.64s 2025-08-29 18:10:20.814757 | orchestrator | Add member roles to user test ------------------------------------------ 11.18s 2025-08-29 18:10:20.814768 | orchestrator | Create test volume ------------------------------------------------------ 6.81s 2025-08-29 18:10:20.814779 | orchestrator | Create ssh security group ----------------------------------------------- 5.71s 2025-08-29 18:10:20.814790 | orchestrator | Add manager role to user test-admin ------------------------------------- 5.67s 2025-08-29 18:10:20.814800 | orchestrator | Create floating ip address ---------------------------------------------- 5.17s 2025-08-29 18:10:20.814811 | orchestrator | Add rule to icmp security group ----------------------------------------- 4.11s 2025-08-29 18:10:20.814822 | orchestrator | Add rule to ssh security group ------------------------------------------ 4.05s 2025-08-29 18:10:20.814833 | orchestrator | Create test user -------------------------------------------------------- 4.02s 2025-08-29 18:10:20.814843 | orchestrator | Create test-admin user -------------------------------------------------- 3.93s 2025-08-29 18:10:20.814854 | orchestrator | Create icmp security group ---------------------------------------------- 3.92s 2025-08-29 18:10:20.814865 | orchestrator | Create test server group ------------------------------------------------ 3.88s 2025-08-29 18:10:20.814876 | orchestrator | Create test keypair ----------------------------------------------------- 3.83s 2025-08-29 18:10:20.814886 | orchestrator | Create test project ----------------------------------------------------- 3.66s 2025-08-29 18:10:20.814904 | orchestrator | Create test domain ------------------------------------------------------ 3.44s 2025-08-29 18:10:20.814915 | orchestrator | Print floating ip address ----------------------------------------------- 0.04s 2025-08-29 18:10:21.134258 | orchestrator | + server_list 2025-08-29 18:10:21.134414 | orchestrator | + openstack --os-cloud test server list 2025-08-29 18:10:24.913604 | orchestrator | +--------------------------------------+--------+--------+----------------------------------------------------+--------------+------------+ 2025-08-29 18:10:24.913711 | orchestrator | | ID | Name | Status | Networks | Image | Flavor | 2025-08-29 18:10:24.913752 | orchestrator | +--------------------------------------+--------+--------+----------------------------------------------------+--------------+------------+ 2025-08-29 18:10:24.913764 | orchestrator | | 15281c56-38f8-415f-b4b9-d6b73e6d202c | test-4 | ACTIVE | auto_allocated_network=10.42.0.4, 192.168.112.129 | Cirros 0.6.2 | SCS-1L-1-5 | 2025-08-29 18:10:24.913774 | orchestrator | | 18c0095c-0b02-4e37-9792-bfda0462b301 | test-3 | ACTIVE | auto_allocated_network=10.42.0.40, 192.168.112.131 | Cirros 0.6.2 | SCS-1L-1-5 | 2025-08-29 18:10:24.913784 | orchestrator | | 14de51fa-1e80-43a0-a564-03a5684e3ff5 | test-2 | ACTIVE | auto_allocated_network=10.42.0.10, 192.168.112.127 | Cirros 0.6.2 | SCS-1L-1-5 | 2025-08-29 18:10:24.913794 | orchestrator | | 77e25b97-3cfb-447d-8691-a9d29c039668 | test-1 | ACTIVE | auto_allocated_network=10.42.0.51, 192.168.112.155 | Cirros 0.6.2 | SCS-1L-1-5 | 2025-08-29 18:10:24.913803 | orchestrator | | ab3d28ec-287e-44e9-b231-b6b8d4c67656 | test | ACTIVE | auto_allocated_network=10.42.0.38, 192.168.112.105 | Cirros 0.6.2 | SCS-1L-1-5 | 2025-08-29 18:10:24.913813 | orchestrator | +--------------------------------------+--------+--------+----------------------------------------------------+--------------+------------+ 2025-08-29 18:10:25.145520 | orchestrator | + openstack --os-cloud test server show test 2025-08-29 18:10:28.556766 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:28.556871 | orchestrator | | Field | Value | 2025-08-29 18:10:28.556887 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:28.556899 | orchestrator | | OS-DCF:diskConfig | MANUAL | 2025-08-29 18:10:28.556910 | orchestrator | | OS-EXT-AZ:availability_zone | nova | 2025-08-29 18:10:28.556921 | orchestrator | | OS-EXT-SRV-ATTR:host | None | 2025-08-29 18:10:28.556932 | orchestrator | | OS-EXT-SRV-ATTR:hostname | test | 2025-08-29 18:10:28.556944 | orchestrator | | OS-EXT-SRV-ATTR:hypervisor_hostname | None | 2025-08-29 18:10:28.556973 | orchestrator | | OS-EXT-SRV-ATTR:instance_name | None | 2025-08-29 18:10:28.556985 | orchestrator | | OS-EXT-SRV-ATTR:kernel_id | None | 2025-08-29 18:10:28.556996 | orchestrator | | OS-EXT-SRV-ATTR:launch_index | None | 2025-08-29 18:10:28.557023 | orchestrator | | OS-EXT-SRV-ATTR:ramdisk_id | None | 2025-08-29 18:10:28.557035 | orchestrator | | OS-EXT-SRV-ATTR:reservation_id | None | 2025-08-29 18:10:28.557056 | orchestrator | | OS-EXT-SRV-ATTR:root_device_name | None | 2025-08-29 18:10:28.557067 | orchestrator | | OS-EXT-SRV-ATTR:user_data | None | 2025-08-29 18:10:28.557078 | orchestrator | | OS-EXT-STS:power_state | Running | 2025-08-29 18:10:28.557089 | orchestrator | | OS-EXT-STS:task_state | None | 2025-08-29 18:10:28.557100 | orchestrator | | OS-EXT-STS:vm_state | active | 2025-08-29 18:10:28.557111 | orchestrator | | OS-SRV-USG:launched_at | 2025-08-29T18:06:14.000000 | 2025-08-29 18:10:28.557133 | orchestrator | | OS-SRV-USG:terminated_at | None | 2025-08-29 18:10:28.557145 | orchestrator | | accessIPv4 | | 2025-08-29 18:10:28.557156 | orchestrator | | accessIPv6 | | 2025-08-29 18:10:28.557167 | orchestrator | | addresses | auto_allocated_network=10.42.0.38, 192.168.112.105 | 2025-08-29 18:10:28.557184 | orchestrator | | config_drive | | 2025-08-29 18:10:28.557196 | orchestrator | | created | 2025-08-29T18:05:51Z | 2025-08-29 18:10:28.557207 | orchestrator | | description | None | 2025-08-29 18:10:28.557218 | orchestrator | | flavor | description=, disk='5', ephemeral='0', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:name-v1='SCS-1L:1:5', extra_specs.scs:name-v2='SCS-1L-1-5', id='SCS-1L-1-5', is_disabled=, is_public='True', location=, name='SCS-1L-1-5', original_name='SCS-1L-1-5', ram='1024', rxtx_factor=, swap='0', vcpus='1' | 2025-08-29 18:10:28.557229 | orchestrator | | hostId | a76233c43b77531d8d0e99e890bbed6fff004c9e7baed6bd5fdc93ba | 2025-08-29 18:10:28.557240 | orchestrator | | host_status | None | 2025-08-29 18:10:28.557257 | orchestrator | | id | ab3d28ec-287e-44e9-b231-b6b8d4c67656 | 2025-08-29 18:10:28.557272 | orchestrator | | image | Cirros 0.6.2 (66774fca-bbf1-4aae-b845-d3ff1965cec0) | 2025-08-29 18:10:28.557285 | orchestrator | | key_name | test | 2025-08-29 18:10:28.557299 | orchestrator | | locked | False | 2025-08-29 18:10:28.557312 | orchestrator | | locked_reason | None | 2025-08-29 18:10:28.557325 | orchestrator | | name | test | 2025-08-29 18:10:28.557344 | orchestrator | | pinned_availability_zone | None | 2025-08-29 18:10:28.557358 | orchestrator | | progress | 0 | 2025-08-29 18:10:28.557392 | orchestrator | | project_id | 9fa5d1851eea43cabe64163b6b118056 | 2025-08-29 18:10:28.557428 | orchestrator | | properties | hostname='test' | 2025-08-29 18:10:28.557441 | orchestrator | | security_groups | name='icmp' | 2025-08-29 18:10:28.557460 | orchestrator | | | name='ssh' | 2025-08-29 18:10:28.557472 | orchestrator | | server_groups | None | 2025-08-29 18:10:28.557489 | orchestrator | | status | ACTIVE | 2025-08-29 18:10:28.557503 | orchestrator | | tags | test | 2025-08-29 18:10:28.557514 | orchestrator | | trusted_image_certificates | None | 2025-08-29 18:10:28.557525 | orchestrator | | updated | 2025-08-29T18:09:03Z | 2025-08-29 18:10:28.557542 | orchestrator | | user_id | 793c6c2954dc431baa8bf6aff49cc227 | 2025-08-29 18:10:28.557554 | orchestrator | | volumes_attached | delete_on_termination='False', id='c3c82a44-2c9e-4a5c-8cfe-20c6ed16fba1' | 2025-08-29 18:10:28.562560 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:28.867559 | orchestrator | + openstack --os-cloud test server show test-1 2025-08-29 18:10:32.056709 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:32.056816 | orchestrator | | Field | Value | 2025-08-29 18:10:32.056856 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:32.056868 | orchestrator | | OS-DCF:diskConfig | MANUAL | 2025-08-29 18:10:32.056880 | orchestrator | | OS-EXT-AZ:availability_zone | nova | 2025-08-29 18:10:32.056905 | orchestrator | | OS-EXT-SRV-ATTR:host | None | 2025-08-29 18:10:32.056917 | orchestrator | | OS-EXT-SRV-ATTR:hostname | test-1 | 2025-08-29 18:10:32.056928 | orchestrator | | OS-EXT-SRV-ATTR:hypervisor_hostname | None | 2025-08-29 18:10:32.056939 | orchestrator | | OS-EXT-SRV-ATTR:instance_name | None | 2025-08-29 18:10:32.056949 | orchestrator | | OS-EXT-SRV-ATTR:kernel_id | None | 2025-08-29 18:10:32.056960 | orchestrator | | OS-EXT-SRV-ATTR:launch_index | None | 2025-08-29 18:10:32.056990 | orchestrator | | OS-EXT-SRV-ATTR:ramdisk_id | None | 2025-08-29 18:10:32.057002 | orchestrator | | OS-EXT-SRV-ATTR:reservation_id | None | 2025-08-29 18:10:32.057020 | orchestrator | | OS-EXT-SRV-ATTR:root_device_name | None | 2025-08-29 18:10:32.057032 | orchestrator | | OS-EXT-SRV-ATTR:user_data | None | 2025-08-29 18:10:32.057043 | orchestrator | | OS-EXT-STS:power_state | Running | 2025-08-29 18:10:32.057054 | orchestrator | | OS-EXT-STS:task_state | None | 2025-08-29 18:10:32.057065 | orchestrator | | OS-EXT-STS:vm_state | active | 2025-08-29 18:10:32.057076 | orchestrator | | OS-SRV-USG:launched_at | 2025-08-29T18:06:56.000000 | 2025-08-29 18:10:32.057087 | orchestrator | | OS-SRV-USG:terminated_at | None | 2025-08-29 18:10:32.057098 | orchestrator | | accessIPv4 | | 2025-08-29 18:10:32.057109 | orchestrator | | accessIPv6 | | 2025-08-29 18:10:32.057121 | orchestrator | | addresses | auto_allocated_network=10.42.0.51, 192.168.112.155 | 2025-08-29 18:10:32.057146 | orchestrator | | config_drive | | 2025-08-29 18:10:32.057164 | orchestrator | | created | 2025-08-29T18:06:35Z | 2025-08-29 18:10:32.057176 | orchestrator | | description | None | 2025-08-29 18:10:32.057187 | orchestrator | | flavor | description=, disk='5', ephemeral='0', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:name-v1='SCS-1L:1:5', extra_specs.scs:name-v2='SCS-1L-1-5', id='SCS-1L-1-5', is_disabled=, is_public='True', location=, name='SCS-1L-1-5', original_name='SCS-1L-1-5', ram='1024', rxtx_factor=, swap='0', vcpus='1' | 2025-08-29 18:10:32.057198 | orchestrator | | hostId | f4159d139dfe70bcae408829b5969366073ab64da9745d549150a058 | 2025-08-29 18:10:32.057221 | orchestrator | | host_status | None | 2025-08-29 18:10:32.057233 | orchestrator | | id | 77e25b97-3cfb-447d-8691-a9d29c039668 | 2025-08-29 18:10:32.057246 | orchestrator | | image | Cirros 0.6.2 (66774fca-bbf1-4aae-b845-d3ff1965cec0) | 2025-08-29 18:10:32.057258 | orchestrator | | key_name | test | 2025-08-29 18:10:32.057271 | orchestrator | | locked | False | 2025-08-29 18:10:32.057283 | orchestrator | | locked_reason | None | 2025-08-29 18:10:32.057303 | orchestrator | | name | test-1 | 2025-08-29 18:10:32.057323 | orchestrator | | pinned_availability_zone | None | 2025-08-29 18:10:32.057337 | orchestrator | | progress | 0 | 2025-08-29 18:10:32.057349 | orchestrator | | project_id | 9fa5d1851eea43cabe64163b6b118056 | 2025-08-29 18:10:32.057361 | orchestrator | | properties | hostname='test-1' | 2025-08-29 18:10:32.057400 | orchestrator | | security_groups | name='icmp' | 2025-08-29 18:10:32.057418 | orchestrator | | | name='ssh' | 2025-08-29 18:10:32.057432 | orchestrator | | server_groups | None | 2025-08-29 18:10:32.057445 | orchestrator | | status | ACTIVE | 2025-08-29 18:10:32.057458 | orchestrator | | tags | test | 2025-08-29 18:10:32.057470 | orchestrator | | trusted_image_certificates | None | 2025-08-29 18:10:32.057490 | orchestrator | | updated | 2025-08-29T18:09:08Z | 2025-08-29 18:10:32.057509 | orchestrator | | user_id | 793c6c2954dc431baa8bf6aff49cc227 | 2025-08-29 18:10:32.057522 | orchestrator | | volumes_attached | | 2025-08-29 18:10:32.063667 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:32.355808 | orchestrator | + openstack --os-cloud test server show test-2 2025-08-29 18:10:35.334299 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:35.334446 | orchestrator | | Field | Value | 2025-08-29 18:10:35.334481 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:35.334494 | orchestrator | | OS-DCF:diskConfig | MANUAL | 2025-08-29 18:10:35.334505 | orchestrator | | OS-EXT-AZ:availability_zone | nova | 2025-08-29 18:10:35.334516 | orchestrator | | OS-EXT-SRV-ATTR:host | None | 2025-08-29 18:10:35.334527 | orchestrator | | OS-EXT-SRV-ATTR:hostname | test-2 | 2025-08-29 18:10:35.334564 | orchestrator | | OS-EXT-SRV-ATTR:hypervisor_hostname | None | 2025-08-29 18:10:35.334602 | orchestrator | | OS-EXT-SRV-ATTR:instance_name | None | 2025-08-29 18:10:35.334613 | orchestrator | | OS-EXT-SRV-ATTR:kernel_id | None | 2025-08-29 18:10:35.334624 | orchestrator | | OS-EXT-SRV-ATTR:launch_index | None | 2025-08-29 18:10:35.334654 | orchestrator | | OS-EXT-SRV-ATTR:ramdisk_id | None | 2025-08-29 18:10:35.334667 | orchestrator | | OS-EXT-SRV-ATTR:reservation_id | None | 2025-08-29 18:10:35.334678 | orchestrator | | OS-EXT-SRV-ATTR:root_device_name | None | 2025-08-29 18:10:35.334695 | orchestrator | | OS-EXT-SRV-ATTR:user_data | None | 2025-08-29 18:10:35.334706 | orchestrator | | OS-EXT-STS:power_state | Running | 2025-08-29 18:10:35.334718 | orchestrator | | OS-EXT-STS:task_state | None | 2025-08-29 18:10:35.334739 | orchestrator | | OS-EXT-STS:vm_state | active | 2025-08-29 18:10:35.334750 | orchestrator | | OS-SRV-USG:launched_at | 2025-08-29T18:07:37.000000 | 2025-08-29 18:10:35.334761 | orchestrator | | OS-SRV-USG:terminated_at | None | 2025-08-29 18:10:35.334772 | orchestrator | | accessIPv4 | | 2025-08-29 18:10:35.334783 | orchestrator | | accessIPv6 | | 2025-08-29 18:10:35.334794 | orchestrator | | addresses | auto_allocated_network=10.42.0.10, 192.168.112.127 | 2025-08-29 18:10:35.334812 | orchestrator | | config_drive | | 2025-08-29 18:10:35.334826 | orchestrator | | created | 2025-08-29T18:07:15Z | 2025-08-29 18:10:35.334843 | orchestrator | | description | None | 2025-08-29 18:10:35.334856 | orchestrator | | flavor | description=, disk='5', ephemeral='0', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:name-v1='SCS-1L:1:5', extra_specs.scs:name-v2='SCS-1L-1-5', id='SCS-1L-1-5', is_disabled=, is_public='True', location=, name='SCS-1L-1-5', original_name='SCS-1L-1-5', ram='1024', rxtx_factor=, swap='0', vcpus='1' | 2025-08-29 18:10:35.334868 | orchestrator | | hostId | 4775762875e9a752da94d0a0ff0ba3f243b108d2ec7df72eda86d4dc | 2025-08-29 18:10:35.334887 | orchestrator | | host_status | None | 2025-08-29 18:10:35.334899 | orchestrator | | id | 14de51fa-1e80-43a0-a564-03a5684e3ff5 | 2025-08-29 18:10:35.334912 | orchestrator | | image | Cirros 0.6.2 (66774fca-bbf1-4aae-b845-d3ff1965cec0) | 2025-08-29 18:10:35.334924 | orchestrator | | key_name | test | 2025-08-29 18:10:35.334936 | orchestrator | | locked | False | 2025-08-29 18:10:35.334948 | orchestrator | | locked_reason | None | 2025-08-29 18:10:35.334961 | orchestrator | | name | test-2 | 2025-08-29 18:10:35.334980 | orchestrator | | pinned_availability_zone | None | 2025-08-29 18:10:35.334993 | orchestrator | | progress | 0 | 2025-08-29 18:10:35.335005 | orchestrator | | project_id | 9fa5d1851eea43cabe64163b6b118056 | 2025-08-29 18:10:35.335017 | orchestrator | | properties | hostname='test-2' | 2025-08-29 18:10:35.335036 | orchestrator | | security_groups | name='icmp' | 2025-08-29 18:10:35.335049 | orchestrator | | | name='ssh' | 2025-08-29 18:10:35.335061 | orchestrator | | server_groups | None | 2025-08-29 18:10:35.335081 | orchestrator | | status | ACTIVE | 2025-08-29 18:10:35.335094 | orchestrator | | tags | test | 2025-08-29 18:10:35.335105 | orchestrator | | trusted_image_certificates | None | 2025-08-29 18:10:35.335116 | orchestrator | | updated | 2025-08-29T18:09:12Z | 2025-08-29 18:10:35.335132 | orchestrator | | user_id | 793c6c2954dc431baa8bf6aff49cc227 | 2025-08-29 18:10:35.335144 | orchestrator | | volumes_attached | | 2025-08-29 18:10:35.340223 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:35.585942 | orchestrator | + openstack --os-cloud test server show test-3 2025-08-29 18:10:38.930749 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:38.930855 | orchestrator | | Field | Value | 2025-08-29 18:10:38.930871 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:38.930884 | orchestrator | | OS-DCF:diskConfig | MANUAL | 2025-08-29 18:10:38.930895 | orchestrator | | OS-EXT-AZ:availability_zone | nova | 2025-08-29 18:10:38.930906 | orchestrator | | OS-EXT-SRV-ATTR:host | None | 2025-08-29 18:10:38.930917 | orchestrator | | OS-EXT-SRV-ATTR:hostname | test-3 | 2025-08-29 18:10:38.930928 | orchestrator | | OS-EXT-SRV-ATTR:hypervisor_hostname | None | 2025-08-29 18:10:38.930939 | orchestrator | | OS-EXT-SRV-ATTR:instance_name | None | 2025-08-29 18:10:38.930950 | orchestrator | | OS-EXT-SRV-ATTR:kernel_id | None | 2025-08-29 18:10:38.930961 | orchestrator | | OS-EXT-SRV-ATTR:launch_index | None | 2025-08-29 18:10:38.931023 | orchestrator | | OS-EXT-SRV-ATTR:ramdisk_id | None | 2025-08-29 18:10:38.931037 | orchestrator | | OS-EXT-SRV-ATTR:reservation_id | None | 2025-08-29 18:10:38.931048 | orchestrator | | OS-EXT-SRV-ATTR:root_device_name | None | 2025-08-29 18:10:38.931059 | orchestrator | | OS-EXT-SRV-ATTR:user_data | None | 2025-08-29 18:10:38.931070 | orchestrator | | OS-EXT-STS:power_state | Running | 2025-08-29 18:10:38.931081 | orchestrator | | OS-EXT-STS:task_state | None | 2025-08-29 18:10:38.931092 | orchestrator | | OS-EXT-STS:vm_state | active | 2025-08-29 18:10:38.931103 | orchestrator | | OS-SRV-USG:launched_at | 2025-08-29T18:08:14.000000 | 2025-08-29 18:10:38.931115 | orchestrator | | OS-SRV-USG:terminated_at | None | 2025-08-29 18:10:38.931126 | orchestrator | | accessIPv4 | | 2025-08-29 18:10:38.931137 | orchestrator | | accessIPv6 | | 2025-08-29 18:10:38.931156 | orchestrator | | addresses | auto_allocated_network=10.42.0.40, 192.168.112.131 | 2025-08-29 18:10:38.931179 | orchestrator | | config_drive | | 2025-08-29 18:10:38.931191 | orchestrator | | created | 2025-08-29T18:07:59Z | 2025-08-29 18:10:38.931202 | orchestrator | | description | None | 2025-08-29 18:10:38.931213 | orchestrator | | flavor | description=, disk='5', ephemeral='0', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:name-v1='SCS-1L:1:5', extra_specs.scs:name-v2='SCS-1L-1-5', id='SCS-1L-1-5', is_disabled=, is_public='True', location=, name='SCS-1L-1-5', original_name='SCS-1L-1-5', ram='1024', rxtx_factor=, swap='0', vcpus='1' | 2025-08-29 18:10:38.931224 | orchestrator | | hostId | f4159d139dfe70bcae408829b5969366073ab64da9745d549150a058 | 2025-08-29 18:10:38.931235 | orchestrator | | host_status | None | 2025-08-29 18:10:38.931246 | orchestrator | | id | 18c0095c-0b02-4e37-9792-bfda0462b301 | 2025-08-29 18:10:38.931257 | orchestrator | | image | Cirros 0.6.2 (66774fca-bbf1-4aae-b845-d3ff1965cec0) | 2025-08-29 18:10:38.931268 | orchestrator | | key_name | test | 2025-08-29 18:10:38.931279 | orchestrator | | locked | False | 2025-08-29 18:10:38.931304 | orchestrator | | locked_reason | None | 2025-08-29 18:10:38.931315 | orchestrator | | name | test-3 | 2025-08-29 18:10:38.931337 | orchestrator | | pinned_availability_zone | None | 2025-08-29 18:10:38.931349 | orchestrator | | progress | 0 | 2025-08-29 18:10:38.931360 | orchestrator | | project_id | 9fa5d1851eea43cabe64163b6b118056 | 2025-08-29 18:10:38.931372 | orchestrator | | properties | hostname='test-3' | 2025-08-29 18:10:38.931411 | orchestrator | | security_groups | name='icmp' | 2025-08-29 18:10:38.931422 | orchestrator | | | name='ssh' | 2025-08-29 18:10:38.931433 | orchestrator | | server_groups | None | 2025-08-29 18:10:38.931444 | orchestrator | | status | ACTIVE | 2025-08-29 18:10:38.931455 | orchestrator | | tags | test | 2025-08-29 18:10:38.931475 | orchestrator | | trusted_image_certificates | None | 2025-08-29 18:10:38.931486 | orchestrator | | updated | 2025-08-29T18:09:17Z | 2025-08-29 18:10:38.931507 | orchestrator | | user_id | 793c6c2954dc431baa8bf6aff49cc227 | 2025-08-29 18:10:38.931519 | orchestrator | | volumes_attached | | 2025-08-29 18:10:38.934786 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:39.202296 | orchestrator | + openstack --os-cloud test server show test-4 2025-08-29 18:10:42.348673 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:42.348771 | orchestrator | | Field | Value | 2025-08-29 18:10:42.348785 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:42.348797 | orchestrator | | OS-DCF:diskConfig | MANUAL | 2025-08-29 18:10:42.348809 | orchestrator | | OS-EXT-AZ:availability_zone | nova | 2025-08-29 18:10:42.348845 | orchestrator | | OS-EXT-SRV-ATTR:host | None | 2025-08-29 18:10:42.348857 | orchestrator | | OS-EXT-SRV-ATTR:hostname | test-4 | 2025-08-29 18:10:42.348868 | orchestrator | | OS-EXT-SRV-ATTR:hypervisor_hostname | None | 2025-08-29 18:10:42.348879 | orchestrator | | OS-EXT-SRV-ATTR:instance_name | None | 2025-08-29 18:10:42.348890 | orchestrator | | OS-EXT-SRV-ATTR:kernel_id | None | 2025-08-29 18:10:42.348902 | orchestrator | | OS-EXT-SRV-ATTR:launch_index | None | 2025-08-29 18:10:42.348929 | orchestrator | | OS-EXT-SRV-ATTR:ramdisk_id | None | 2025-08-29 18:10:42.348941 | orchestrator | | OS-EXT-SRV-ATTR:reservation_id | None | 2025-08-29 18:10:42.348952 | orchestrator | | OS-EXT-SRV-ATTR:root_device_name | None | 2025-08-29 18:10:42.348963 | orchestrator | | OS-EXT-SRV-ATTR:user_data | None | 2025-08-29 18:10:42.348974 | orchestrator | | OS-EXT-STS:power_state | Running | 2025-08-29 18:10:42.348994 | orchestrator | | OS-EXT-STS:task_state | None | 2025-08-29 18:10:42.349005 | orchestrator | | OS-EXT-STS:vm_state | active | 2025-08-29 18:10:42.349016 | orchestrator | | OS-SRV-USG:launched_at | 2025-08-29T18:08:47.000000 | 2025-08-29 18:10:42.349027 | orchestrator | | OS-SRV-USG:terminated_at | None | 2025-08-29 18:10:42.349056 | orchestrator | | accessIPv4 | | 2025-08-29 18:10:42.349073 | orchestrator | | accessIPv6 | | 2025-08-29 18:10:42.349084 | orchestrator | | addresses | auto_allocated_network=10.42.0.4, 192.168.112.129 | 2025-08-29 18:10:42.349101 | orchestrator | | config_drive | | 2025-08-29 18:10:42.349113 | orchestrator | | created | 2025-08-29T18:08:32Z | 2025-08-29 18:10:42.349124 | orchestrator | | description | None | 2025-08-29 18:10:42.349136 | orchestrator | | flavor | description=, disk='5', ephemeral='0', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:name-v1='SCS-1L:1:5', extra_specs.scs:name-v2='SCS-1L-1-5', id='SCS-1L-1-5', is_disabled=, is_public='True', location=, name='SCS-1L-1-5', original_name='SCS-1L-1-5', ram='1024', rxtx_factor=, swap='0', vcpus='1' | 2025-08-29 18:10:42.349153 | orchestrator | | hostId | 4775762875e9a752da94d0a0ff0ba3f243b108d2ec7df72eda86d4dc | 2025-08-29 18:10:42.349165 | orchestrator | | host_status | None | 2025-08-29 18:10:42.349176 | orchestrator | | id | 15281c56-38f8-415f-b4b9-d6b73e6d202c | 2025-08-29 18:10:42.349187 | orchestrator | | image | Cirros 0.6.2 (66774fca-bbf1-4aae-b845-d3ff1965cec0) | 2025-08-29 18:10:42.349198 | orchestrator | | key_name | test | 2025-08-29 18:10:42.349214 | orchestrator | | locked | False | 2025-08-29 18:10:42.349227 | orchestrator | | locked_reason | None | 2025-08-29 18:10:42.349240 | orchestrator | | name | test-4 | 2025-08-29 18:10:42.349259 | orchestrator | | pinned_availability_zone | None | 2025-08-29 18:10:42.349272 | orchestrator | | progress | 0 | 2025-08-29 18:10:42.349291 | orchestrator | | project_id | 9fa5d1851eea43cabe64163b6b118056 | 2025-08-29 18:10:42.349304 | orchestrator | | properties | hostname='test-4' | 2025-08-29 18:10:42.349318 | orchestrator | | security_groups | name='icmp' | 2025-08-29 18:10:42.349330 | orchestrator | | | name='ssh' | 2025-08-29 18:10:42.349343 | orchestrator | | server_groups | None | 2025-08-29 18:10:42.349355 | orchestrator | | status | ACTIVE | 2025-08-29 18:10:42.349367 | orchestrator | | tags | test | 2025-08-29 18:10:42.349410 | orchestrator | | trusted_image_certificates | None | 2025-08-29 18:10:42.349423 | orchestrator | | updated | 2025-08-29T18:09:21Z | 2025-08-29 18:10:42.349441 | orchestrator | | user_id | 793c6c2954dc431baa8bf6aff49cc227 | 2025-08-29 18:10:42.349455 | orchestrator | | volumes_attached | | 2025-08-29 18:10:42.353035 | orchestrator | +-------------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+ 2025-08-29 18:10:42.627519 | orchestrator | + server_ping 2025-08-29 18:10:42.628994 | orchestrator | ++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address' 2025-08-29 18:10:42.629108 | orchestrator | ++ tr -d '\r' 2025-08-29 18:10:45.560358 | orchestrator | + for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c "Floating IP Address" | tr -d '\r') 2025-08-29 18:10:45.561243 | orchestrator | + ping -c3 192.168.112.155 2025-08-29 18:10:45.575881 | orchestrator | PING 192.168.112.155 (192.168.112.155) 56(84) bytes of data. 2025-08-29 18:10:45.575922 | orchestrator | 64 bytes from 192.168.112.155: icmp_seq=1 ttl=63 time=8.72 ms 2025-08-29 18:10:46.571835 | orchestrator | 64 bytes from 192.168.112.155: icmp_seq=2 ttl=63 time=2.50 ms 2025-08-29 18:10:47.573796 | orchestrator | 64 bytes from 192.168.112.155: icmp_seq=3 ttl=63 time=1.97 ms 2025-08-29 18:10:47.573899 | orchestrator | 2025-08-29 18:10:47.573916 | orchestrator | --- 192.168.112.155 ping statistics --- 2025-08-29 18:10:47.573929 | orchestrator | 3 packets transmitted, 3 received, 0% packet loss, time 2004ms 2025-08-29 18:10:47.573940 | orchestrator | rtt min/avg/max/mdev = 1.971/4.396/8.716/3.062 ms 2025-08-29 18:10:47.573952 | orchestrator | + for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c "Floating IP Address" | tr -d '\r') 2025-08-29 18:10:47.573964 | orchestrator | + ping -c3 192.168.112.127 2025-08-29 18:10:47.585657 | orchestrator | PING 192.168.112.127 (192.168.112.127) 56(84) bytes of data. 2025-08-29 18:10:47.585686 | orchestrator | 64 bytes from 192.168.112.127: icmp_seq=1 ttl=63 time=7.41 ms 2025-08-29 18:10:48.582839 | orchestrator | 64 bytes from 192.168.112.127: icmp_seq=2 ttl=63 time=2.51 ms 2025-08-29 18:10:49.585006 | orchestrator | 64 bytes from 192.168.112.127: icmp_seq=3 ttl=63 time=2.33 ms 2025-08-29 18:10:49.585100 | orchestrator | 2025-08-29 18:10:49.585114 | orchestrator | --- 192.168.112.127 ping statistics --- 2025-08-29 18:10:49.585126 | orchestrator | 3 packets transmitted, 3 received, 0% packet loss, time 2003ms 2025-08-29 18:10:49.585136 | orchestrator | rtt min/avg/max/mdev = 2.327/4.082/7.412/2.355 ms 2025-08-29 18:10:49.585798 | orchestrator | + for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c "Floating IP Address" | tr -d '\r') 2025-08-29 18:10:49.585888 | orchestrator | + ping -c3 192.168.112.129 2025-08-29 18:10:49.594855 | orchestrator | PING 192.168.112.129 (192.168.112.129) 56(84) bytes of data. 2025-08-29 18:10:49.594883 | orchestrator | 64 bytes from 192.168.112.129: icmp_seq=1 ttl=63 time=5.58 ms 2025-08-29 18:10:50.594258 | orchestrator | 64 bytes from 192.168.112.129: icmp_seq=2 ttl=63 time=2.70 ms 2025-08-29 18:10:51.595669 | orchestrator | 64 bytes from 192.168.112.129: icmp_seq=3 ttl=63 time=2.06 ms 2025-08-29 18:10:51.595770 | orchestrator | 2025-08-29 18:10:51.595785 | orchestrator | --- 192.168.112.129 ping statistics --- 2025-08-29 18:10:51.595797 | orchestrator | 3 packets transmitted, 3 received, 0% packet loss, time 2004ms 2025-08-29 18:10:51.595809 | orchestrator | rtt min/avg/max/mdev = 2.064/3.447/5.581/1.530 ms 2025-08-29 18:10:51.596595 | orchestrator | + for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c "Floating IP Address" | tr -d '\r') 2025-08-29 18:10:51.596619 | orchestrator | + ping -c3 192.168.112.131 2025-08-29 18:10:51.608559 | orchestrator | PING 192.168.112.131 (192.168.112.131) 56(84) bytes of data. 2025-08-29 18:10:51.608585 | orchestrator | 64 bytes from 192.168.112.131: icmp_seq=1 ttl=63 time=7.44 ms 2025-08-29 18:10:52.605765 | orchestrator | 64 bytes from 192.168.112.131: icmp_seq=2 ttl=63 time=2.62 ms 2025-08-29 18:10:53.607773 | orchestrator | 64 bytes from 192.168.112.131: icmp_seq=3 ttl=63 time=2.28 ms 2025-08-29 18:10:53.607875 | orchestrator | 2025-08-29 18:10:53.607890 | orchestrator | --- 192.168.112.131 ping statistics --- 2025-08-29 18:10:53.607903 | orchestrator | 3 packets transmitted, 3 received, 0% packet loss, time 2003ms 2025-08-29 18:10:53.607914 | orchestrator | rtt min/avg/max/mdev = 2.281/4.111/7.437/2.355 ms 2025-08-29 18:10:53.607926 | orchestrator | + for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c "Floating IP Address" | tr -d '\r') 2025-08-29 18:10:53.607979 | orchestrator | + ping -c3 192.168.112.105 2025-08-29 18:10:53.621437 | orchestrator | PING 192.168.112.105 (192.168.112.105) 56(84) bytes of data. 2025-08-29 18:10:53.621472 | orchestrator | 64 bytes from 192.168.112.105: icmp_seq=1 ttl=63 time=8.86 ms 2025-08-29 18:10:54.616989 | orchestrator | 64 bytes from 192.168.112.105: icmp_seq=2 ttl=63 time=3.08 ms 2025-08-29 18:10:55.616814 | orchestrator | 64 bytes from 192.168.112.105: icmp_seq=3 ttl=63 time=1.83 ms 2025-08-29 18:10:55.617846 | orchestrator | 2025-08-29 18:10:55.617882 | orchestrator | --- 192.168.112.105 ping statistics --- 2025-08-29 18:10:55.617896 | orchestrator | 3 packets transmitted, 3 received, 0% packet loss, time 2002ms 2025-08-29 18:10:55.617907 | orchestrator | rtt min/avg/max/mdev = 1.834/4.591/8.861/3.061 ms 2025-08-29 18:10:55.617931 | orchestrator | + [[ 9.2.0 == \l\a\t\e\s\t ]] 2025-08-29 18:10:55.712626 | orchestrator | ok: Runtime: 0:10:26.077361 2025-08-29 18:10:55.754884 | 2025-08-29 18:10:55.755018 | TASK [Run tempest] 2025-08-29 18:10:56.289128 | orchestrator | skipping: Conditional result was False 2025-08-29 18:10:56.307169 | 2025-08-29 18:10:56.307319 | TASK [Check prometheus alert status] 2025-08-29 18:10:56.842689 | orchestrator | skipping: Conditional result was False 2025-08-29 18:10:56.845866 | 2025-08-29 18:10:56.846075 | PLAY RECAP 2025-08-29 18:10:56.846231 | orchestrator | ok: 24 changed: 11 unreachable: 0 failed: 0 skipped: 5 rescued: 0 ignored: 0 2025-08-29 18:10:56.846300 | 2025-08-29 18:10:57.078487 | RUN END RESULT_NORMAL: [untrusted : github.com/osism/testbed/playbooks/deploy.yml@main] 2025-08-29 18:10:57.079617 | POST-RUN START: [untrusted : github.com/osism/testbed/playbooks/post.yml@main] 2025-08-29 18:10:57.840687 | 2025-08-29 18:10:57.840850 | PLAY [Post output play] 2025-08-29 18:10:57.856851 | 2025-08-29 18:10:57.856994 | LOOP [stage-output : Register sources] 2025-08-29 18:10:57.929742 | 2025-08-29 18:10:57.930102 | TASK [stage-output : Check sudo] 2025-08-29 18:10:58.789246 | orchestrator | sudo: a password is required 2025-08-29 18:10:58.971579 | orchestrator | ok: Runtime: 0:00:00.015211 2025-08-29 18:10:58.986261 | 2025-08-29 18:10:58.986488 | LOOP [stage-output : Set source and destination for files and folders] 2025-08-29 18:10:59.028856 | 2025-08-29 18:10:59.029175 | TASK [stage-output : Build a list of source, dest dictionaries] 2025-08-29 18:10:59.092858 | orchestrator | ok 2025-08-29 18:10:59.102221 | 2025-08-29 18:10:59.102378 | LOOP [stage-output : Ensure target folders exist] 2025-08-29 18:10:59.547461 | orchestrator | ok: "docs" 2025-08-29 18:10:59.547810 | 2025-08-29 18:10:59.775987 | orchestrator | ok: "artifacts" 2025-08-29 18:11:00.016997 | orchestrator | ok: "logs" 2025-08-29 18:11:00.035497 | 2025-08-29 18:11:00.035660 | LOOP [stage-output : Copy files and folders to staging folder] 2025-08-29 18:11:00.075696 | 2025-08-29 18:11:00.075987 | TASK [stage-output : Make all log files readable] 2025-08-29 18:11:00.347715 | orchestrator | ok 2025-08-29 18:11:00.357009 | 2025-08-29 18:11:00.357149 | TASK [stage-output : Rename log files that match extensions_to_txt] 2025-08-29 18:11:00.402887 | orchestrator | skipping: Conditional result was False 2025-08-29 18:11:00.419787 | 2025-08-29 18:11:00.419960 | TASK [stage-output : Discover log files for compression] 2025-08-29 18:11:00.445726 | orchestrator | skipping: Conditional result was False 2025-08-29 18:11:00.464175 | 2025-08-29 18:11:00.464365 | LOOP [stage-output : Archive everything from logs] 2025-08-29 18:11:00.513171 | 2025-08-29 18:11:00.513350 | PLAY [Post cleanup play] 2025-08-29 18:11:00.521637 | 2025-08-29 18:11:00.521751 | TASK [Set cloud fact (Zuul deployment)] 2025-08-29 18:11:00.581784 | orchestrator | ok 2025-08-29 18:11:00.594051 | 2025-08-29 18:11:00.594194 | TASK [Set cloud fact (local deployment)] 2025-08-29 18:11:00.629352 | orchestrator | skipping: Conditional result was False 2025-08-29 18:11:00.643197 | 2025-08-29 18:11:00.643367 | TASK [Clean the cloud environment] 2025-08-29 18:11:01.200766 | orchestrator | 2025-08-29 18:11:01 - clean up servers 2025-08-29 18:11:01.924928 | orchestrator | 2025-08-29 18:11:01 - testbed-manager 2025-08-29 18:11:02.002707 | orchestrator | 2025-08-29 18:11:02 - testbed-node-1 2025-08-29 18:11:02.090190 | orchestrator | 2025-08-29 18:11:02 - testbed-node-4 2025-08-29 18:11:02.178675 | orchestrator | 2025-08-29 18:11:02 - testbed-node-2 2025-08-29 18:11:02.268634 | orchestrator | 2025-08-29 18:11:02 - testbed-node-0 2025-08-29 18:11:02.357893 | orchestrator | 2025-08-29 18:11:02 - testbed-node-5 2025-08-29 18:11:02.440460 | orchestrator | 2025-08-29 18:11:02 - testbed-node-3 2025-08-29 18:11:02.536167 | orchestrator | 2025-08-29 18:11:02 - clean up keypairs 2025-08-29 18:11:02.550376 | orchestrator | 2025-08-29 18:11:02 - testbed 2025-08-29 18:11:02.573018 | orchestrator | 2025-08-29 18:11:02 - wait for servers to be gone 2025-08-29 18:11:11.299884 | orchestrator | 2025-08-29 18:11:11 - clean up ports 2025-08-29 18:11:11.489673 | orchestrator | 2025-08-29 18:11:11 - 439b83ae-b76c-43e8-bba2-fd94170abfb8 2025-08-29 18:11:11.957182 | orchestrator | 2025-08-29 18:11:11 - 4dac935d-a84d-4be3-830c-60b89606d097 2025-08-29 18:11:12.243819 | orchestrator | 2025-08-29 18:11:12 - 7de1b8d7-5a09-45f0-967e-92f385f39e56 2025-08-29 18:11:12.447678 | orchestrator | 2025-08-29 18:11:12 - c3682e91-0bc9-4afc-a372-129810224734 2025-08-29 18:11:12.667137 | orchestrator | 2025-08-29 18:11:12 - c6afcf8e-1df2-4630-afe1-b0a00d8cd915 2025-08-29 18:11:12.861679 | orchestrator | 2025-08-29 18:11:12 - ca38cb25-bf61-4013-8f49-8127f925ca15 2025-08-29 18:11:13.109208 | orchestrator | 2025-08-29 18:11:13 - ff163896-8f2c-47d0-88f7-f5289aa5741f 2025-08-29 18:11:13.331964 | orchestrator | 2025-08-29 18:11:13 - clean up volumes 2025-08-29 18:11:13.454316 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-0-node-base 2025-08-29 18:11:13.493480 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-1-node-base 2025-08-29 18:11:13.531834 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-4-node-base 2025-08-29 18:11:13.574493 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-2-node-base 2025-08-29 18:11:13.619038 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-5-node-base 2025-08-29 18:11:13.657965 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-3-node-base 2025-08-29 18:11:13.703670 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-manager-base 2025-08-29 18:11:13.751102 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-2-node-5 2025-08-29 18:11:13.795459 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-0-node-3 2025-08-29 18:11:13.834639 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-6-node-3 2025-08-29 18:11:13.878132 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-1-node-4 2025-08-29 18:11:13.919352 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-3-node-3 2025-08-29 18:11:13.958988 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-7-node-4 2025-08-29 18:11:13.998071 | orchestrator | 2025-08-29 18:11:13 - testbed-volume-4-node-4 2025-08-29 18:11:14.043114 | orchestrator | 2025-08-29 18:11:14 - testbed-volume-8-node-5 2025-08-29 18:11:14.095955 | orchestrator | 2025-08-29 18:11:14 - testbed-volume-5-node-5 2025-08-29 18:11:14.140484 | orchestrator | 2025-08-29 18:11:14 - disconnect routers 2025-08-29 18:11:14.209204 | orchestrator | 2025-08-29 18:11:14 - testbed 2025-08-29 18:11:15.244650 | orchestrator | 2025-08-29 18:11:15 - clean up subnets 2025-08-29 18:11:15.285177 | orchestrator | 2025-08-29 18:11:15 - subnet-testbed-management 2025-08-29 18:11:15.436122 | orchestrator | 2025-08-29 18:11:15 - clean up networks 2025-08-29 18:11:15.581585 | orchestrator | 2025-08-29 18:11:15 - net-testbed-management 2025-08-29 18:11:15.869301 | orchestrator | 2025-08-29 18:11:15 - clean up security groups 2025-08-29 18:11:15.905562 | orchestrator | 2025-08-29 18:11:15 - testbed-node 2025-08-29 18:11:16.035515 | orchestrator | 2025-08-29 18:11:16 - testbed-management 2025-08-29 18:11:16.167197 | orchestrator | 2025-08-29 18:11:16 - clean up floating ips 2025-08-29 18:11:16.202240 | orchestrator | 2025-08-29 18:11:16 - 81.163.192.190 2025-08-29 18:11:16.589838 | orchestrator | 2025-08-29 18:11:16 - clean up routers 2025-08-29 18:11:17.075589 | orchestrator | 2025-08-29 18:11:17 - testbed 2025-08-29 18:11:18.196670 | orchestrator | ok: Runtime: 0:00:16.940171 2025-08-29 18:11:18.200909 | 2025-08-29 18:11:18.201080 | PLAY RECAP 2025-08-29 18:11:18.201212 | orchestrator | ok: 6 changed: 2 unreachable: 0 failed: 0 skipped: 7 rescued: 0 ignored: 0 2025-08-29 18:11:18.201275 | 2025-08-29 18:11:18.353336 | POST-RUN END RESULT_NORMAL: [untrusted : github.com/osism/testbed/playbooks/post.yml@main] 2025-08-29 18:11:18.355814 | POST-RUN START: [untrusted : github.com/osism/testbed/playbooks/cleanup.yml@main] 2025-08-29 18:11:19.114673 | 2025-08-29 18:11:19.114913 | PLAY [Cleanup play] 2025-08-29 18:11:19.133967 | 2025-08-29 18:11:19.134145 | TASK [Set cloud fact (Zuul deployment)] 2025-08-29 18:11:19.196380 | orchestrator | ok 2025-08-29 18:11:19.203542 | 2025-08-29 18:11:19.203688 | TASK [Set cloud fact (local deployment)] 2025-08-29 18:11:19.239370 | orchestrator | skipping: Conditional result was False 2025-08-29 18:11:19.258133 | 2025-08-29 18:11:19.258297 | TASK [Clean the cloud environment] 2025-08-29 18:11:20.405838 | orchestrator | 2025-08-29 18:11:20 - clean up servers 2025-08-29 18:11:20.887187 | orchestrator | 2025-08-29 18:11:20 - clean up keypairs 2025-08-29 18:11:20.906768 | orchestrator | 2025-08-29 18:11:20 - wait for servers to be gone 2025-08-29 18:11:20.951210 | orchestrator | 2025-08-29 18:11:20 - clean up ports 2025-08-29 18:11:21.020521 | orchestrator | 2025-08-29 18:11:21 - clean up volumes 2025-08-29 18:11:21.077880 | orchestrator | 2025-08-29 18:11:21 - disconnect routers 2025-08-29 18:11:21.100097 | orchestrator | 2025-08-29 18:11:21 - clean up subnets 2025-08-29 18:11:21.117620 | orchestrator | 2025-08-29 18:11:21 - clean up networks 2025-08-29 18:11:21.270329 | orchestrator | 2025-08-29 18:11:21 - clean up security groups 2025-08-29 18:11:21.303813 | orchestrator | 2025-08-29 18:11:21 - clean up floating ips 2025-08-29 18:11:21.331005 | orchestrator | 2025-08-29 18:11:21 - clean up routers 2025-08-29 18:11:21.801545 | orchestrator | ok: Runtime: 0:00:01.303944 2025-08-29 18:11:21.805286 | 2025-08-29 18:11:21.805516 | PLAY RECAP 2025-08-29 18:11:21.805648 | orchestrator | ok: 2 changed: 1 unreachable: 0 failed: 0 skipped: 1 rescued: 0 ignored: 0 2025-08-29 18:11:21.805709 | 2025-08-29 18:11:21.943212 | POST-RUN END RESULT_NORMAL: [untrusted : github.com/osism/testbed/playbooks/cleanup.yml@main] 2025-08-29 18:11:21.945881 | POST-RUN START: [trusted : github.com/osism/zuul-config/playbooks/base/post-fetch.yaml@main] 2025-08-29 18:11:22.687468 | 2025-08-29 18:11:22.687657 | PLAY [Base post-fetch] 2025-08-29 18:11:22.703570 | 2025-08-29 18:11:22.703890 | TASK [fetch-output : Set log path for multiple nodes] 2025-08-29 18:11:22.760132 | orchestrator | skipping: Conditional result was False 2025-08-29 18:11:22.771344 | 2025-08-29 18:11:22.771539 | TASK [fetch-output : Set log path for single node] 2025-08-29 18:11:22.810002 | orchestrator | ok 2025-08-29 18:11:22.818289 | 2025-08-29 18:11:22.818440 | LOOP [fetch-output : Ensure local output dirs] 2025-08-29 18:11:23.302006 | orchestrator -> localhost | ok: "/var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/work/logs" 2025-08-29 18:11:23.576093 | orchestrator -> localhost | changed: "/var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/work/artifacts" 2025-08-29 18:11:23.848835 | orchestrator -> localhost | changed: "/var/lib/zuul/builds/931a11d02915441d8037f098b05d6016/work/docs" 2025-08-29 18:11:23.874724 | 2025-08-29 18:11:23.874955 | LOOP [fetch-output : Collect logs, artifacts and docs] 2025-08-29 18:11:24.802928 | orchestrator | changed: .d..t...... ./ 2025-08-29 18:11:24.803372 | orchestrator | changed: All items complete 2025-08-29 18:11:24.803461 | 2025-08-29 18:11:25.540481 | orchestrator | changed: .d..t...... ./ 2025-08-29 18:11:26.245107 | orchestrator | changed: .d..t...... ./ 2025-08-29 18:11:26.270327 | 2025-08-29 18:11:26.270547 | LOOP [merge-output-to-logs : Move artifacts and docs to logs dir] 2025-08-29 18:11:26.308659 | orchestrator | skipping: Conditional result was False 2025-08-29 18:11:26.312566 | orchestrator | skipping: Conditional result was False 2025-08-29 18:11:26.328949 | 2025-08-29 18:11:26.329057 | PLAY RECAP 2025-08-29 18:11:26.329133 | orchestrator | ok: 3 changed: 2 unreachable: 0 failed: 0 skipped: 2 rescued: 0 ignored: 0 2025-08-29 18:11:26.329172 | 2025-08-29 18:11:26.464844 | POST-RUN END RESULT_NORMAL: [trusted : github.com/osism/zuul-config/playbooks/base/post-fetch.yaml@main] 2025-08-29 18:11:26.467435 | POST-RUN START: [trusted : github.com/osism/zuul-config/playbooks/base/post.yaml@main] 2025-08-29 18:11:27.254736 | 2025-08-29 18:11:27.254947 | PLAY [Base post] 2025-08-29 18:11:27.269573 | 2025-08-29 18:11:27.269709 | TASK [remove-build-sshkey : Remove the build SSH key from all nodes] 2025-08-29 18:11:28.188002 | orchestrator | changed 2025-08-29 18:11:28.196534 | 2025-08-29 18:11:28.196668 | PLAY RECAP 2025-08-29 18:11:28.196740 | orchestrator | ok: 1 changed: 1 unreachable: 0 failed: 0 skipped: 0 rescued: 0 ignored: 0 2025-08-29 18:11:28.196810 | 2025-08-29 18:11:28.327516 | POST-RUN END RESULT_NORMAL: [trusted : github.com/osism/zuul-config/playbooks/base/post.yaml@main] 2025-08-29 18:11:28.328608 | POST-RUN START: [trusted : github.com/osism/zuul-config/playbooks/base/post-logs.yaml@main] 2025-08-29 18:11:29.095351 | 2025-08-29 18:11:29.095593 | PLAY [Base post-logs] 2025-08-29 18:11:29.107160 | 2025-08-29 18:11:29.107329 | TASK [generate-zuul-manifest : Generate Zuul manifest] 2025-08-29 18:11:29.573076 | localhost | changed 2025-08-29 18:11:29.592002 | 2025-08-29 18:11:29.592226 | TASK [generate-zuul-manifest : Return Zuul manifest URL to Zuul] 2025-08-29 18:11:29.630504 | localhost | ok 2025-08-29 18:11:29.636018 | 2025-08-29 18:11:29.636182 | TASK [Set zuul-log-path fact] 2025-08-29 18:11:29.653833 | localhost | ok 2025-08-29 18:11:29.667037 | 2025-08-29 18:11:29.667167 | TASK [set-zuul-log-path-fact : Set log path for a build] 2025-08-29 18:11:29.703130 | localhost | ok 2025-08-29 18:11:29.706962 | 2025-08-29 18:11:29.707078 | TASK [upload-logs : Create log directories] 2025-08-29 18:11:30.226458 | localhost | changed 2025-08-29 18:11:30.231507 | 2025-08-29 18:11:30.231678 | TASK [upload-logs : Ensure logs are readable before uploading] 2025-08-29 18:11:30.733616 | localhost -> localhost | ok: Runtime: 0:00:00.007286 2025-08-29 18:11:30.738073 | 2025-08-29 18:11:30.738197 | TASK [upload-logs : Upload logs to log server] 2025-08-29 18:11:31.312450 | localhost | Output suppressed because no_log was given 2025-08-29 18:11:31.315614 | 2025-08-29 18:11:31.315777 | LOOP [upload-logs : Compress console log and json output] 2025-08-29 18:11:31.372268 | localhost | skipping: Conditional result was False 2025-08-29 18:11:31.376923 | localhost | skipping: Conditional result was False 2025-08-29 18:11:31.390020 | 2025-08-29 18:11:31.390264 | LOOP [upload-logs : Upload compressed console log and json output] 2025-08-29 18:11:31.441490 | localhost | skipping: Conditional result was False 2025-08-29 18:11:31.442172 | 2025-08-29 18:11:31.446777 | localhost | skipping: Conditional result was False 2025-08-29 18:11:31.461066 | 2025-08-29 18:11:31.461246 | LOOP [upload-logs : Upload console log and json output]